mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-08-09 01:52:41 +00:00
Implement 2p version of learning rate schedule.
This commit is contained in:
parent
da50525ca5
commit
82d58629ea
@ -791,7 +791,7 @@ def run(rank, world_size, args):
|
|||||||
scheduler = torch.optim.lr_scheduler.LambdaLR(
|
scheduler = torch.optim.lr_scheduler.LambdaLR(
|
||||||
optimizer,
|
optimizer,
|
||||||
lambda step: (((step**2 + params.lr_steps**2) / params.lr_steps**2) ** -0.25 *
|
lambda step: (((step**2 + params.lr_steps**2) / params.lr_steps**2) ** -0.25 *
|
||||||
(((epoch**2 + params.lr_epochs**2) / params.lr_epochs**2) ** -0.25))
|
(((epoch**2 + params.lr_epochs**2) / params.lr_epochs**2) ** -0.25)))
|
||||||
|
|
||||||
|
|
||||||
if checkpoints and "optimizer" in checkpoints:
|
if checkpoints and "optimizer" in checkpoints:
|
||||||
|
Loading…
x
Reference in New Issue
Block a user