Skip to content

Commit 9ed2e1c

Browse files
author
zhengmiao
committed
Merge branch 'limengzhang/fix_optim_delete' into 'refactor_dev'
[Fix] Remove _delete_=True in optimizer See merge request openmmlab-enterprise/openmmlab-ce/mmsegmentation!51
2 parents e5952a9 + 2c4bcbd commit 9ed2e1c

25 files changed

+60
-156
lines changed

configs/beit/upernet_beit-base_8x2_640x640_160k_ade20k.py

Lines changed: 3 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -9,16 +9,11 @@
99
pretrained='pretrain/beit_base_patch16_224_pt22k_ft22k.pth',
1010
test_cfg=dict(mode='slide', crop_size=(640, 640), stride=(426, 426)))
1111

12-
optimizer = dict(
13-
_delete_=True,
14-
type='AdamW',
15-
lr=3e-5,
16-
betas=(0.9, 0.999),
17-
weight_decay=0.05)
18-
1912
optim_wrapper = dict(
13+
_delete_=True,
2014
type='OptimWrapper',
21-
optimizer=optimizer,
15+
optimizer=dict(
16+
type='AdamW', lr=3e-5, betas=(0.9, 0.999), weight_decay=0.05),
2217
constructor='LayerDecayOptimizerConstructor',
2318
paramwise_cfg=dict(num_layers=12, layer_decay_rate=0.9))
2419

configs/beit/upernet_beit-large_fp16_8x1_640x640_160k_ade20k.py

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -23,17 +23,11 @@
2323
auxiliary_head=dict(in_channels=1024, num_classes=150),
2424
test_cfg=dict(mode='slide', crop_size=(640, 640), stride=(426, 426)))
2525

26-
optimizer = dict(
27-
_delete_=True,
28-
type='AdamW',
29-
lr=2e-5,
30-
betas=(0.9, 0.999),
31-
weight_decay=0.05)
32-
3326
optim_wrapper = dict(
3427
_delete_=True,
3528
type='AmpOptimWrapper',
36-
optimizer=optimizer,
29+
optimizer=dict(
30+
type='AdamW', lr=2e-5, betas=(0.9, 0.999), weight_decay=0.05),
3731
constructor='LayerDecayOptimizerConstructor',
3832
paramwise_cfg=dict(num_layers=24, layer_decay_rate=0.95),
3933
accumulative_counts=2)
Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,6 @@
11
_base_ = './bisenetv2_fcn_4x4_1024x1024_160k_cityscapes.py'
2-
optimizer = dict(type='SGD', lr=0.05, momentum=0.9, weight_decay=0.0005)
32
optim_wrapper = dict(
43
_delete_=True,
54
type='AmpOptimWrapper',
6-
optimizer=optimizer,
5+
optimizer=dict(type='SGD', lr=0.05, momentum=0.9, weight_decay=0.0005),
76
loss_scale=512.)

configs/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k.py

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -11,17 +11,11 @@
1111
test_cfg=dict(mode='slide', crop_size=crop_size, stride=(341, 341)),
1212
)
1313

14-
optimizer = dict(
15-
_delete_=True,
16-
type='AdamW',
17-
lr=0.0001,
18-
betas=(0.9, 0.999),
19-
weight_decay=0.05)
20-
2114
optim_wrapper = dict(
2215
_delete_=True,
2316
type='AmpOptimWrapper',
24-
optimizer=optimizer,
17+
optimizer=dict(
18+
type='AdamW', lr=0.0001, betas=(0.9, 0.999), weight_decay=0.05),
2519
paramwise_cfg={
2620
'decay_rate': 0.9,
2721
'decay_type': 'stage_wise',

configs/convnext/upernet_convnext_base_fp16_640x640_160k_ade20k.py

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -26,17 +26,11 @@
2626
test_cfg=dict(mode='slide', crop_size=crop_size, stride=(426, 426)),
2727
)
2828

29-
optimizer = dict(
30-
_delete_=True,
31-
type='AdamW',
32-
lr=0.0001,
33-
betas=(0.9, 0.999),
34-
weight_decay=0.05)
35-
3629
optim_wrapper = dict(
3730
_delete_=True,
3831
type='AmpOptimWrapper',
39-
optimizer=optimizer,
32+
optimizer=dict(
33+
type='AdamW', lr=0.0001, betas=(0.9, 0.999), weight_decay=0.05),
4034
paramwise_cfg={
4135
'decay_rate': 0.9,
4236
'decay_type': 'stage_wise',

configs/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k.py

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -26,17 +26,11 @@
2626
test_cfg=dict(mode='slide', crop_size=crop_size, stride=(426, 426)),
2727
)
2828

29-
optimizer = dict(
30-
_delete_=True,
31-
type='AdamW',
32-
lr=0.0001,
33-
betas=(0.9, 0.999),
34-
weight_decay=0.05)
35-
3629
optim_wrapper = dict(
3730
_delete_=True,
3831
type='AmpOptimWrapper',
39-
optimizer=optimizer,
32+
optimizer=dict(
33+
type='AdamW', lr=0.0001, betas=(0.9, 0.999), weight_decay=0.05),
4034
paramwise_cfg={
4135
'decay_rate': 0.9,
4236
'decay_type': 'stage_wise',

configs/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k.py

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -25,17 +25,11 @@
2525
test_cfg=dict(mode='slide', crop_size=crop_size, stride=(341, 341)),
2626
)
2727

28-
optimizer = dict(
29-
_delete_=True,
30-
type='AdamW',
31-
lr=0.0001,
32-
betas=(0.9, 0.999),
33-
weight_decay=0.05)
34-
3528
optim_wrapper = dict(
3629
_delete_=True,
3730
type='AmpOptimWrapper',
38-
optimizer=optimizer,
31+
optimizer=dict(
32+
type='AdamW', lr=0.0001, betas=(0.9, 0.999), weight_decay=0.05),
3933
paramwise_cfg={
4034
'decay_rate': 0.9,
4135
'decay_type': 'stage_wise',

configs/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k.py

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -25,17 +25,11 @@
2525
test_cfg=dict(mode='slide', crop_size=crop_size, stride=(341, 341)),
2626
)
2727

28-
optimizer = dict(
29-
_delete_=True,
30-
type='AdamW',
31-
lr=0.0001,
32-
betas=(0.9, 0.999),
33-
weight_decay=0.05)
34-
3528
optim_wrapper = dict(
3629
_delete_=True,
3730
type='AmpOptimWrapper',
38-
optimizer=optimizer,
31+
optimizer=dict(
32+
type='AdamW', lr=0.0001, betas=(0.9, 0.999), weight_decay=0.05),
3933
paramwise_cfg={
4034
'decay_rate': 0.9,
4135
'decay_type': 'stage_wise',

configs/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k.py

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -26,17 +26,11 @@
2626
test_cfg=dict(mode='slide', crop_size=crop_size, stride=(426, 426)),
2727
)
2828

29-
optimizer = dict(
30-
_delete_=True,
31-
type='AdamW',
32-
lr=0.00008,
33-
betas=(0.9, 0.999),
34-
weight_decay=0.05)
35-
3629
optim_wrapper = dict(
3730
_delete_=True,
3831
type='AmpOptimWrapper',
39-
optimizer=optimizer,
32+
optimizer=dict(
33+
type='AdamW', lr=0.00008, betas=(0.9, 0.999), weight_decay=0.05),
4034
paramwise_cfg={
4135
'decay_rate': 0.9,
4236
'decay_type': 'stage_wise',

configs/dpt/dpt_vit-b16_512x512_160k_ade20k.py

Lines changed: 2 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -7,16 +7,11 @@
77
model = dict(data_preprocessor=data_preprocessor)
88
# AdamW optimizer, no weight decay for position embedding & layer norm
99
# in backbone
10-
optimizer = dict(
11-
_delete_=True,
12-
type='AdamW',
13-
lr=0.00006,
14-
betas=(0.9, 0.999),
15-
weight_decay=0.01)
1610

1711
optim_wrapper = dict(
1812
type='OptimWrapper',
19-
optimizer=optimizer,
13+
optimizer=dict(
14+
type='AdamW', lr=0.00006, betas=(0.9, 0.999), weight_decay=0.01),
2015
paramwise_cfg=dict(
2116
custom_keys={
2217
'pos_embed': dict(decay_mult=0.),
Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,6 @@
11
_base_ = './fcn_r101-d8_512x1024_80k_cityscapes.py'
2-
optimizer = dict(type='SGD', lr=0.01, momentum=0.9, weight_decay=0.0005)
32
optim_wrapper = dict(
43
_delete_=True,
54
type='AmpOptimWrapper',
6-
optimizer=optimizer,
5+
optimizer=dict(type='SGD', lr=0.01, momentum=0.9, weight_decay=0.0005),
76
loss_scale=512.)

configs/mae/upernet_mae-base_fp16_8x2_512x512_160k_ade20k.py

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -24,16 +24,10 @@
2424
auxiliary_head=dict(in_channels=768, num_classes=150),
2525
test_cfg=dict(mode='slide', crop_size=(512, 512), stride=(341, 341)))
2626

27-
optimizer = dict(
28-
_delete_=True,
29-
type='AdamW',
30-
lr=1e-4,
31-
betas=(0.9, 0.999),
32-
weight_decay=0.05)
33-
3427
optim_wrapper = dict(
3528
type='OptimWrapper',
36-
optimizer=optimizer,
29+
optimizer=dict(
30+
type='AdamW', lr=1e-4, betas=(0.9, 0.999), weight_decay=0.05),
3731
paramwise_cfg=dict(num_layers=12, layer_decay_rate=0.65),
3832
constructor='LayerDecayOptimizerConstructor')
3933

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,6 @@
11
_base_ = './pspnet_r101-d8_512x1024_80k_cityscapes.py'
2-
optimizer = dict(type='SGD', lr=0.01, momentum=0.9, weight_decay=0.0005)
32
optim_wrapper = dict(
43
_delete_=True,
54
type='AmpOptimWrapper',
6-
optimizer=optimizer,
5+
optimizer=dict(type='SGD', lr=0.01, momentum=0.9, weight_decay=0.0005),
76
loss_scale=512.)

configs/pspnet/pspnet_r50-d32_rsb-pretrain_512x1024_adamw_80k_cityscapes.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,11 +15,10 @@
1515
dilations=(1, 1, 2, 4),
1616
strides=(1, 2, 2, 2)))
1717

18-
optimizer = dict(_delete_=True, type='AdamW', lr=0.0005, weight_decay=0.05)
1918
optim_wrapper = dict(
2019
_delete_=True,
2120
type='OptimWrapper',
22-
optimizer=optimizer,
21+
optimizer=dict(type='AdamW', lr=0.0005, weight_decay=0.05),
2322
clip_grad=dict(max_norm=1, norm_type=2))
2423
# learning policy
2524
param_scheduler = [

configs/pspnet/pspnet_r50-d8_rsb-pretrain_512x1024_adamw_80k_cityscapes.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -13,11 +13,10 @@
1313
init_cfg=dict(
1414
type='Pretrained', prefix='backbone.', checkpoint=checkpoint)))
1515

16-
optimizer = dict(_delete_=True, type='AdamW', lr=0.0005, weight_decay=0.05)
1716
optim_wrapper = dict(
1817
_delete_=True,
1918
type='OptimWrapper',
20-
optimizer=optimizer,
19+
optimizer=dict(type='AdamW', lr=0.0005, weight_decay=0.05),
2120
clip_grad=dict(max_norm=1, norm_type=2))
2221
# learning policy
2322
param_scheduler = [

configs/segformer/segformer_mit-b0_512x512_160k_ade20k.py

Lines changed: 3 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -9,17 +9,11 @@
99
pretrained='pretrain/mit_b0.pth',
1010
decode_head=dict(num_classes=150))
1111

12-
# optimizer
13-
optimizer = dict(
14-
_delete_=True,
15-
type='AdamW',
16-
lr=0.00006,
17-
betas=(0.9, 0.999),
18-
weight_decay=0.01)
19-
2012
optim_wrapper = dict(
13+
_delete_=True,
2114
type='OptimWrapper',
22-
optimizer=optimizer,
15+
optimizer=dict(
16+
type='AdamW', lr=0.00006, betas=(0.9, 0.999), weight_decay=0.01),
2317
paramwise_cfg=dict(
2418
custom_keys={
2519
'pos_block': dict(decay_mult=0.),

configs/segformer/segformer_mit-b0_8x1_1024x1024_160k_cityscapes.py

Lines changed: 3 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -11,17 +11,11 @@
1111
init_cfg=dict(type='Pretrained', checkpoint='pretrain/mit_b0.pth')),
1212
test_cfg=dict(mode='slide', crop_size=(1024, 1024), stride=(768, 768)))
1313

14-
# optimizer
15-
optimizer = dict(
16-
_delete_=True,
17-
type='AdamW',
18-
lr=0.00006,
19-
betas=(0.9, 0.999),
20-
weight_decay=0.01)
21-
2214
optim_wrapper = dict(
15+
_delete_=True,
2316
type='OptimWrapper',
24-
optimizer=optimizer,
17+
optimizer=dict(
18+
type='AdamW', lr=0.00006, betas=(0.9, 0.999), weight_decay=0.01),
2519
paramwise_cfg=dict(
2620
custom_keys={
2721
'pos_block': dict(decay_mult=0.),

configs/swin/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K.py

Lines changed: 3 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -21,16 +21,11 @@
2121

2222
# AdamW optimizer, no weight decay for position embedding & layer norm
2323
# in backbone
24-
optimizer = dict(
25-
_delete_=True,
26-
type='AdamW',
27-
lr=0.00006,
28-
betas=(0.9, 0.999),
29-
weight_decay=0.01)
30-
3124
optim_wrapper = dict(
25+
_delete_=True,
3226
type='OptimWrapper',
33-
optimizer=optimizer,
27+
optimizer=dict(
28+
type='AdamW', lr=0.00006, betas=(0.9, 0.999), weight_decay=0.01),
3429
paramwise_cfg=dict(
3530
custom_keys={
3631
'absolute_pos_embed': dict(decay_mult=0.),

configs/twins/twins_pcpvt-s_fpn_fpnhead_8x4_512x512_80k_ade20k.py

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,4 +5,8 @@
55
crop_size = (512, 512)
66
data_preprocessor = dict(size=crop_size)
77
model = dict(data_preprocessor=data_preprocessor)
8-
optimizer = dict(_delete_=True, type='AdamW', lr=0.0001, weight_decay=0.0001)
8+
optim_wrapper = dict(
9+
_delete_=True,
10+
type='OptimWrapper',
11+
optimizer=dict(type='AdamW', lr=0.0001, weight_decay=0.0001),
12+
clip_grad=None)

configs/twins/twins_pcpvt-s_uperhead_8x4_512x512_160k_ade20k.py

Lines changed: 3 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -6,16 +6,12 @@
66
crop_size = (512, 512)
77
data_preprocessor = dict(size=crop_size)
88
model = dict(data_preprocessor=data_preprocessor)
9-
optimizer = dict(
10-
_delete_=True,
11-
type='AdamW',
12-
lr=0.00006,
13-
betas=(0.9, 0.999),
14-
weight_decay=0.01)
159

1610
optim_wrapper = dict(
11+
_delete_=True,
1712
type='OptimWrapper',
18-
optimizer=optimizer,
13+
optimizer=dict(
14+
type='AdamW', lr=0.00006, betas=(0.9, 0.999), weight_decay=0.01),
1915
paramwise_cfg=dict(custom_keys={
2016
'pos_block': dict(decay_mult=0.),
2117
'norm': dict(decay_mult=0.)

configs/twins/twins_svt-s_fpn_fpnhead_8x4_512x512_80k_ade20k.py

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,4 +21,8 @@
2121
decode_head=dict(num_classes=150),
2222
)
2323

24-
optimizer = dict(_delete_=True, type='AdamW', lr=0.0001, weight_decay=0.0001)
24+
optim_wrapper = dict(
25+
_delete_=True,
26+
type='OptimWrapper',
27+
optimizer=dict(type='AdamW', lr=0.0001, weight_decay=0.0001),
28+
clip_grad=None)

configs/twins/twins_svt-s_uperhead_8x2_512x512_160k_ade20k.py

Lines changed: 3 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -21,16 +21,11 @@
2121
decode_head=dict(in_channels=[64, 128, 256, 512]),
2222
auxiliary_head=dict(in_channels=256))
2323

24-
optimizer = dict(
25-
_delete_=True,
26-
type='AdamW',
27-
lr=0.00006,
28-
betas=(0.9, 0.999),
29-
weight_decay=0.01)
30-
3124
optim_wrapper = dict(
25+
_delete_=True,
3226
type='OptimWrapper',
33-
optimizer=optimizer,
27+
optimizer=dict(
28+
type='AdamW', lr=0.00006, betas=(0.9, 0.999), weight_decay=0.01),
3429
paramwise_cfg=dict(custom_keys={
3530
'pos_block': dict(decay_mult=0.),
3631
'norm': dict(decay_mult=0.)

configs/vit/upernet_vit-b16_ln_mln_512x512_160k_ade20k.py

Lines changed: 3 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -14,16 +14,11 @@
1414

1515
# AdamW optimizer, no weight decay for position embedding & layer norm
1616
# in backbone
17-
optimizer = dict(
18-
_delete_=True,
19-
type='AdamW',
20-
lr=0.00006,
21-
betas=(0.9, 0.999),
22-
weight_decay=0.01)
23-
2417
optim_wrapper = dict(
18+
_delete_=True,
2519
type='OptimWrapper',
26-
optimizer=optimizer,
20+
optimizer=dict(
21+
type='AdamW', lr=0.00006, betas=(0.9, 0.999), weight_decay=0.01),
2722
paramwise_cfg=dict(
2823
custom_keys={
2924
'pos_embed': dict(decay_mult=0.),

0 commit comments

Comments
 (0)