Skip to content

[BUG] Training job cannot resume if LR scheduler is Plateau. #871

Open
@wuye9036

Description

@wuye9036

File "Trainer.py", line 233, in main
lr_scheduler.step(start_epoch)
File "/multimedia-nfs/wuye/libs/miniconda3/envs/py38/lib/python3.8/site-packages/timm/scheduler/plateau_lr.py", line 83, in step
self.lr_scheduler.step(metric, epoch) # step the base scheduler
File "/multimedia-nfs/wuye/libs/miniconda3/envs/py38/lib/python3.8/site-packages/torch/optim/lr_scheduler.py", line 624, in step
current = float(metrics)
TypeError: float() argument must be a string or a number, not 'NoneType'

Trainer.py is just renamed from official training script.

Metadata

Metadata

Assignees

Labels

bugSomething isn't working

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions