|
|
@@ -182,7 +182,8 @@ def train():
|
|
|
optimizer, start_epoch = build_optimizer(model_cfg, model_without_ddp, model_cfg['lr0'], args.resume)
|
|
|
|
|
|
# Scheduler
|
|
|
- scheduler, lf = build_lr_scheduler(model_cfg, optimizer, args.max_epoch)
|
|
|
+ total_epochs = args.max_epoch + args.wp_epoch
|
|
|
+ scheduler, lf = build_lr_scheduler(model_cfg, optimizer, total_epochs)
|
|
|
scheduler.last_epoch = start_epoch - 1 # do not move
|
|
|
if args.resume:
|
|
|
scheduler.step()
|
|
|
@@ -197,7 +198,6 @@ def train():
|
|
|
# start training loop
|
|
|
best_map = -1.0
|
|
|
last_opt_step = -1
|
|
|
- total_epochs = args.max_epoch + args.wp_epoch
|
|
|
heavy_eval = False
|
|
|
optimizer.zero_grad()
|
|
|
|