84
84
#config.core.jit_model_path = "<your-script-or-quantize-model-path>"
85
85
86
86
## -------------------- training ------------------
87
- config .core .epoch_num = 200
87
+ config .core .epoch_num = 100
88
88
config .core .save_num = 1
89
89
config .core .shuffle = True
90
90
config .core .batch_size = 16
105
105
weight = torch .from_numpy (config .data ['classWeights' ]).to (config .core .device )
106
106
config .core .criterion = torch .nn .CrossEntropyLoss (weight )
107
107
108
- # config.core.teacher = AttrDict()
109
- # config.core.teacher.net = EESPNet_Seg(config.core.cls_num)
110
- # config.core.teacher.criterion = torch.nn.CrossEntropyLoss(weight)
111
- # config.core.teacher.optimizer = torch.optim.Adam(config.core.teacher.net.parameters(), 5e-4, (0.9, 0.999), eps=1e-08, weight_decay=5e-4)
112
-
113
108
config .core .mean = config .data ['mean' ]
114
109
config .core .std = config .data ['std' ]
115
110
## -------------------- optimizer and scheduler ------------------
116
111
config .core .optimizer = torch .optim .Adam (config .core .net .parameters (), 3e-4 , (0.9 , 0.999 ), eps = 1e-08 , weight_decay = 5e-4 )
117
- # config.core.optimizer = torch.optim.SGD(config.core.net.parameters(), 5e-3, momentum=0.9)
118
-
119
112
lambda_lr = lambda epoch : round ((1 - epoch / config .core .epoch_num ) ** 0.9 , 8 )
120
113
config .core .scheduler = optim .lr_scheduler .LambdaLR (config .core .optimizer , lr_lambda = lambda_lr )
121
- # config.core.scheduler = optim.lr_scheduler.MultiStepLR(config.core.optimizer, milestones=[20,40,55,70,80,90,100,110,120,130,140,150,160], gamma=0.27030)
122
114
123
115
## -------------------- loader ------------------
124
116
config .core .num_workers = 3
182
174
batch_size = config .core .batch_size , shuffle = True , num_workers = config .core .num_workers , pin_memory = config .pin_memory )
183
175
scale4_train_loader .is_last_loader = False
184
176
# multi loader
185
- config .core .train_loader_list = [scale1_train_loader , scale2_train_loader , scale4_train_loader , scale3_train_loader , last_train_loader ]
177
+ config .core .train_loader_list = [scale1_train_loader , scale2_train_loader , scale4_train_loader , scale3_train_loader , last_train_loader ]
0 commit comments