-
Notifications
You must be signed in to change notification settings - Fork 126
Open
Description
I wonder whether you forgot to modify like the line shown below in:
| warmup_lr = [base_lr * ((self.multiplier - 1.) * self.last_epoch / self.total_epoch + 1.) for base_lr in self.base_lrs] |
+ warmup_lr = self.get_lr()
- warmup_lr = [base_lr * ((self.multiplier - 1.) * self.last_epoch / self.total_epoch + 1.) for base_lr in self.base_lrs]
Here is the details:
- When I use
ReduceLROnPlateauas theafter_schedulerofGradualWarmupScheduler, the warm-up failed. The way I get the learning rate is:optim.param_groups[0]['lr']. Then I useget_lrto get the learning rate, I found it is correct. - I use
StepLRas theafter_scheduler, I found there was no exception and no error.
Therefor, I think the learning rate of the optimizer hadn't been warmed up correctly.
Metadata
Metadata
Assignees
Labels
No labels