-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
Update performance of video models #256
Changes from all commits
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -74,7 +74,7 @@ | |
dict(type='ToTensor', keys=['imgs']) | ||
] | ||
data = dict( | ||
videos_per_gpu=8, | ||
videos_per_gpu=24, | ||
workers_per_gpu=4, | ||
train=dict( | ||
type=dataset_type, | ||
|
@@ -93,7 +93,7 @@ | |
pipeline=test_pipeline)) | ||
# optimizer | ||
optimizer = dict( | ||
type='SGD', lr=0.1, momentum=0.9, | ||
type='SGD', lr=0.3, momentum=0.9, | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. double check that this lr is for 8 gpus There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Yes, that is for 8 gpu (since we use batch size 24 on each GPU |
||
weight_decay=0.0001) # this lr is used for 8 gpus | ||
optimizer_config = dict(grad_clip=dict(max_norm=40, norm_type=2)) | ||
# learning policy | ||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -35,12 +35,7 @@ | |
num_clips=3, | ||
start_index=0), | ||
dict(type='DecordDecode'), | ||
dict( | ||
type='MultiScaleCrop', | ||
input_size=224, | ||
scales=(1, 0.875, 0.75, 0.66), | ||
random_crop=False, | ||
max_wh_scale_gap=1), | ||
dict(type='RandomResizedCrop'), | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. does it mean that RandomResizedCrop is better than MultiScaleCrop in this case? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Sure, we already validated that RandomResizedCrop can outperform MultiScaleCrop. The contribution of that PR is to show that training with videos doesn't lead to any performance drop. |
||
dict(type='Resize', scale=(224, 224), keep_ratio=False), | ||
dict(type='Flip', flip_ratio=0.5), | ||
dict(type='Normalize', **img_norm_cfg), | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
v100?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nope, slowfast_4x16 doesn't consume lots of memory, we can fit 24 samples onto a 1080 Ti.