Skip to content

Commit

Permalink
Merge 7f9de81 into 14d250f
Browse files Browse the repository at this point in the history
  • Loading branch information
ZwwWayne authored Feb 25, 2021
2 parents 14d250f + 7f9de81 commit 4662568
Show file tree
Hide file tree
Showing 95 changed files with 100 additions and 95 deletions.
2 changes: 1 addition & 1 deletion configs/cascade_rcnn/cascade_rcnn_r50_fpn_20e_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './cascade_rcnn_r50_fpn_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 19])
runner = dict(max_epochs=20)
runner = dict(type='EpochBasedRunner', max_epochs=20)
Original file line number Diff line number Diff line change
Expand Up @@ -102,4 +102,4 @@
warmup_iters=500,
warmup_ratio=1.0 / 3,
step=[190])
runner = dict(max_epochs=210)
runner = dict(type='EpochBasedRunner', max_epochs=210)
3 changes: 2 additions & 1 deletion configs/cityscapes/faster_rcnn_r50_fpn_1x_cityscapes.py
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,8 @@
warmup_ratio=0.001,
# [7] yields higher performance than [6]
step=[7])
runner = dict(max_epochs=8) # actual epoch = 8 * 8 = 64
runner = dict(
type='EpochBasedRunner', max_epochs=8) # actual epoch = 8 * 8 = 64
log_config = dict(interval=100)
# For better, more stable performance initialize from COCO
load_from = 'https://download.openmmlab.com/mmdetection/v2.0/faster_rcnn/faster_rcnn_r50_fpn_1x_coco/faster_rcnn_r50_fpn_1x_coco_20200130-047c8118.pth' # noqa
3 changes: 2 additions & 1 deletion configs/cityscapes/mask_rcnn_r50_fpn_1x_cityscapes.py
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,8 @@
warmup_ratio=0.001,
# [7] yields higher performance than [6]
step=[7])
runner = dict(max_epochs=8) # actual epoch = 8 * 8 = 64
runner = dict(
type='EpochBasedRunner', max_epochs=8) # actual epoch = 8 * 8 = 64
log_config = dict(interval=100)
# For better, more stable performance initialize from COCO
load_from = 'https://download.openmmlab.com/mmdetection/v2.0/mask_rcnn/mask_rcnn_r50_fpn_1x_coco/mask_rcnn_r50_fpn_1x_coco_20200205-d4b0c5d6.pth' # noqa
Original file line number Diff line number Diff line change
Expand Up @@ -102,4 +102,4 @@
warmup_iters=500,
warmup_ratio=1.0 / 3,
step=[180])
runner = dict(max_epochs=210)
runner = dict(type='EpochBasedRunner', max_epochs=210)
Original file line number Diff line number Diff line change
Expand Up @@ -102,4 +102,4 @@
warmup_iters=500,
warmup_ratio=1.0 / 3,
step=[180])
runner = dict(max_epochs=210)
runner = dict(type='EpochBasedRunner', max_epochs=210)
Original file line number Diff line number Diff line change
Expand Up @@ -102,4 +102,4 @@
warmup_iters=500,
warmup_ratio=1.0 / 3,
step=[180])
runner = dict(max_epochs=210)
runner = dict(type='EpochBasedRunner', max_epochs=210)
2 changes: 1 addition & 1 deletion configs/deepfashion/mask_rcnn_r50_fpn_15e_deepfashion.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,4 +7,4 @@
roi_head=dict(
bbox_head=dict(num_classes=15), mask_head=dict(num_classes=15)))
# runtime settings
runner = dict(max_epochs=15)
runner = dict(type='EpochBasedRunner', max_epochs=15)
2 changes: 1 addition & 1 deletion configs/detr/detr_r50_8x2_150e_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -128,4 +128,4 @@
optimizer_config = dict(grad_clip=dict(max_norm=0.1, norm_type=2))
# learning policy
lr_config = dict(policy='step', step=[100])
runner = dict(max_epochs=150)
runner = dict(type='EpochBasedRunner', max_epochs=150)
2 changes: 1 addition & 1 deletion configs/fast_rcnn/fast_rcnn_r50_fpn_2x_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,4 +2,4 @@

# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './faster_rcnn_r50_caffe_dc5_mstrain_1x_coco.py'
# learning policy
lr_config = dict(step=[28, 34])
runner = dict(max_epochs=36)
runner = dict(type='EpochBasedRunner', max_epochs=36)
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './faster_rcnn_r50_caffe_fpn_mstrain_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 23])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './faster_rcnn_r50_caffe_fpn_mstrain_1x_coco.py'
# learning policy
lr_config = dict(step=[28, 34])
runner = dict(max_epochs=36)
runner = dict(type='EpochBasedRunner', max_epochs=36)
Original file line number Diff line number Diff line change
Expand Up @@ -41,4 +41,4 @@
test=dict(pipeline=test_pipeline))
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/fcos/fcos_r50_caffe_fpn_gn-head_1x_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -102,4 +102,4 @@
warmup_iters=500,
warmup_ratio=1.0 / 3,
step=[8, 11])
runner = dict(max_epochs=12)
runner = dict(type='EpochBasedRunner', max_epochs=12)
Original file line number Diff line number Diff line change
Expand Up @@ -36,4 +36,4 @@
test=dict(pipeline=test_pipeline))
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
Original file line number Diff line number Diff line change
Expand Up @@ -56,4 +56,4 @@
_delete_=True, grad_clip=dict(max_norm=35, norm_type=2))
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
Original file line number Diff line number Diff line change
Expand Up @@ -7,4 +7,4 @@
norm_cfg=dict(type='GN', num_groups=32, requires_grad=True)))
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
Original file line number Diff line number Diff line change
Expand Up @@ -24,4 +24,4 @@
data = dict(train=dict(pipeline=train_pipeline))
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,6 @@
norm_cfg=dict(type='GN', num_groups=32, requires_grad=True)))
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
optimizer_config = dict(
_delete_=True, grad_clip=dict(max_norm=35, norm_type=2))
Original file line number Diff line number Diff line change
Expand Up @@ -22,4 +22,4 @@
data = dict(train=dict(pipeline=train_pipeline))
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/foveabox/fovea_r50_fpn_4x4_2x_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './fovea_r50_fpn_4x4_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/gfl/gfl_r50_fpn_mstrain_2x_coco.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
_base_ = './gfl_r50_fpn_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
# multi-scale training
img_norm_cfg = dict(
mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True)
Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './mask_rcnn_r101_fpn_gn_ws-all_2x_coco.py'
# learning policy
lr_config = dict(step=[20, 23])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './mask_rcnn_r50_fpn_gn_ws-all_2x_coco.py'
# learning policy
lr_config = dict(step=[20, 23])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/gn+ws/mask_rcnn_r50_fpn_gn_ws-all_2x_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,4 +14,4 @@
mask_head=dict(conv_cfg=conv_cfg, norm_cfg=norm_cfg)))
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './mask_rcnn_x101_32x4d_fpn_gn_ws-all_2x_coco.py'
# learning policy
lr_config = dict(step=[20, 23])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './mask_rcnn_x50_32x4d_fpn_gn_ws-all_2x_coco.py'
# learning policy
lr_config = dict(step=[20, 23])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/gn/mask_rcnn_r101_fpn_gn-all_3x_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,4 +2,4 @@

# learning policy
lr_config = dict(step=[28, 34])
runner = dict(max_epochs=36)
runner = dict(type='EpochBasedRunner', max_epochs=36)
2 changes: 1 addition & 1 deletion configs/gn/mask_rcnn_r50_fpn_gn-all_2x_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -43,4 +43,4 @@
test=dict(pipeline=test_pipeline))
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/gn/mask_rcnn_r50_fpn_gn-all_3x_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,4 +2,4 @@

# learning policy
lr_config = dict(step=[28, 34])
runner = dict(max_epochs=36)
runner = dict(type='EpochBasedRunner', max_epochs=36)
2 changes: 1 addition & 1 deletion configs/gn/mask_rcnn_r50_fpn_gn-all_contrib_2x_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,4 +12,4 @@
mask_head=dict(norm_cfg=norm_cfg)))
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/gn/mask_rcnn_r50_fpn_gn-all_contrib_3x_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,4 +2,4 @@

# learning policy
lr_config = dict(step=[28, 34])
runner = dict(max_epochs=36)
runner = dict(type='EpochBasedRunner', max_epochs=36)
2 changes: 1 addition & 1 deletion configs/grid_rcnn/grid_rcnn_r50_fpn_gn-head_1x_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -8,4 +8,4 @@
step=[8, 11])
checkpoint_config = dict(interval=1)
# runtime settings
runner = dict(max_epochs=12)
runner = dict(type='EpochBasedRunner', max_epochs=12)
2 changes: 1 addition & 1 deletion configs/grid_rcnn/grid_rcnn_r50_fpn_gn-head_2x_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -132,4 +132,4 @@
warmup_iters=3665,
warmup_ratio=1.0 / 80,
step=[17, 23])
runner = dict(max_epochs=25)
runner = dict(type='EpochBasedRunner', max_epochs=25)
Original file line number Diff line number Diff line change
Expand Up @@ -20,4 +20,4 @@
warmup_iters=3665,
warmup_ratio=1.0 / 80,
step=[17, 23])
runner = dict(max_epochs=25)
runner = dict(type='EpochBasedRunner', max_epochs=25)
Original file line number Diff line number Diff line change
Expand Up @@ -164,4 +164,4 @@
])
# yapf:enable
# runtime settings
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/hrnet/cascade_mask_rcnn_hrnetv2p_w32_20e_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,4 +36,4 @@
out_channels=256))
# learning policy
lr_config = dict(step=[16, 19])
runner = dict(max_epochs=20)
runner = dict(type='EpochBasedRunner', max_epochs=20)
2 changes: 1 addition & 1 deletion configs/hrnet/cascade_rcnn_hrnetv2p_w32_20e_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,4 +36,4 @@
out_channels=256))
# learning policy
lr_config = dict(step=[16, 19])
runner = dict(max_epochs=20)
runner = dict(type='EpochBasedRunner', max_epochs=20)
2 changes: 1 addition & 1 deletion configs/hrnet/faster_rcnn_hrnetv2p_w18_2x_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,4 +2,4 @@

# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/hrnet/faster_rcnn_hrnetv2p_w32_2x_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './faster_rcnn_hrnetv2p_w32_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/hrnet/faster_rcnn_hrnetv2p_w40_2x_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './faster_rcnn_hrnetv2p_w40_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/hrnet/fcos_hrnetv2p_w18_gn-head_4x4_2x_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './fcos_hrnetv2p_w18_gn-head_4x4_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/hrnet/fcos_hrnetv2p_w32_gn-head_4x4_2x_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './fcos_hrnetv2p_w32_gn-head_4x4_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
Original file line number Diff line number Diff line change
Expand Up @@ -36,4 +36,4 @@
test=dict(pipeline=test_pipeline))
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/hrnet/htc_hrnetv2p_w40_28e_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './htc_hrnetv2p_w40_20e_coco.py'
# learning policy
lr_config = dict(step=[24, 27])
runner = dict(max_epochs=28)
runner = dict(type='EpochBasedRunner', max_epochs=28)
2 changes: 1 addition & 1 deletion configs/hrnet/htc_x101_64x4d_fpn_16x1_28e_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = '../htc/htc_x101_64x4d_fpn_16x1_20e_coco.py'
# learning policy
lr_config = dict(step=[24, 27])
runner = dict(max_epochs=28)
runner = dict(type='EpochBasedRunner', max_epochs=28)
2 changes: 1 addition & 1 deletion configs/hrnet/mask_rcnn_hrnetv2p_w18_2x_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './mask_rcnn_hrnetv2p_w18_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/hrnet/mask_rcnn_hrnetv2p_w32_2x_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './mask_rcnn_hrnetv2p_w32_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/hrnet/mask_rcnn_hrnetv2p_w40_2x_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './mask_rcnn_hrnetv2p_w40_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/htc/htc_r101_fpn_20e_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,4 +2,4 @@
model = dict(pretrained='torchvision://resnet101', backbone=dict(depth=101))
# learning policy
lr_config = dict(step=[16, 19])
runner = dict(max_epochs=20)
runner = dict(type='EpochBasedRunner', max_epochs=20)
2 changes: 1 addition & 1 deletion configs/htc/htc_r50_fpn_20e_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './htc_r50_fpn_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 19])
runner = dict(max_epochs=20)
runner = dict(type='EpochBasedRunner', max_epochs=20)
2 changes: 1 addition & 1 deletion configs/htc/htc_x101_32x4d_fpn_16x1_20e_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,4 +15,4 @@
data = dict(samples_per_gpu=1, workers_per_gpu=1)
# learning policy
lr_config = dict(step=[16, 19])
runner = dict(max_epochs=20)
runner = dict(type='EpochBasedRunner', max_epochs=20)
2 changes: 1 addition & 1 deletion configs/htc/htc_x101_64x4d_fpn_16x1_20e_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,4 +15,4 @@
data = dict(samples_per_gpu=1, workers_per_gpu=1)
# learning policy
lr_config = dict(step=[16, 19])
runner = dict(max_epochs=20)
runner = dict(type='EpochBasedRunner', max_epochs=20)
Original file line number Diff line number Diff line change
Expand Up @@ -39,4 +39,4 @@
samples_per_gpu=1, workers_per_gpu=1, train=dict(pipeline=train_pipeline))
# learning policy
lr_config = dict(step=[16, 19])
runner = dict(max_epochs=20)
runner = dict(type='EpochBasedRunner', max_epochs=20)
Original file line number Diff line number Diff line change
Expand Up @@ -25,4 +25,4 @@
data = dict(train=dict(pipeline=train_pipeline))
# learning policy
lr_config = dict(step=[32, 44])
runner = dict(max_epochs=48)
runner = dict(type='EpochBasedRunner', max_epochs=48)
2 changes: 1 addition & 1 deletion configs/instaboost/mask_rcnn_r50_fpn_instaboost_4x_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -25,4 +25,4 @@
data = dict(train=dict(pipeline=train_pipeline))
# learning policy
lr_config = dict(step=[32, 44])
runner = dict(max_epochs=48)
runner = dict(type='EpochBasedRunner', max_epochs=48)
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './mask_rcnn_r50_caffe_fpn_mstrain-poly_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 23])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './mask_rcnn_r50_caffe_fpn_mstrain-poly_1x_coco.py'
# learning policy
lr_config = dict(step=[28, 34])
runner = dict(max_epochs=36)
runner = dict(type='EpochBasedRunner', max_epochs=36)
Original file line number Diff line number Diff line change
Expand Up @@ -58,4 +58,4 @@
test=dict(pipeline=test_pipeline))

lr_config = dict(step=[28, 34])
runner = dict(max_epochs=36)
runner = dict(type='EpochBasedRunner', max_epochs=36)
2 changes: 1 addition & 1 deletion configs/ms_rcnn/ms_rcnn_r101_caffe_fpn_2x_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './ms_rcnn_r101_caffe_fpn_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/ms_rcnn/ms_rcnn_r50_caffe_fpn_2x_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './ms_rcnn_r50_caffe_fpn_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/ms_rcnn/ms_rcnn_x101_64x4d_fpn_2x_coco.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
_base_ = './ms_rcnn_x101_64x4d_fpn_1x_coco.py'
# learning policy
lr_config = dict(step=[16, 22])
runner = dict(max_epochs=24)
runner = dict(type='EpochBasedRunner', max_epochs=24)
2 changes: 1 addition & 1 deletion configs/nas_fpn/retinanet_r50_fpn_crop640_50e_coco.py
Original file line number Diff line number Diff line change
Expand Up @@ -77,4 +77,4 @@
warmup_ratio=0.1,
step=[30, 40])
# runtime settings
runner = dict(max_epochs=50)
runner = dict(type='EpochBasedRunner', max_epochs=50)
Loading

0 comments on commit 4662568

Please sign in to comment.