From 82d58629eaa54b64b32e68cb44d519bb58e530e6 Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Sun, 10 Apr 2022 13:50:31 +0800 Subject: [PATCH] Implement 2p version of learning rate schedule. --- egs/librispeech/ASR/pruned_transducer_stateless2/train.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless2/train.py b/egs/librispeech/ASR/pruned_transducer_stateless2/train.py index a8aaa4dde..73ba17a71 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless2/train.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless2/train.py @@ -791,7 +791,7 @@ def run(rank, world_size, args): scheduler = torch.optim.lr_scheduler.LambdaLR( optimizer, lambda step: (((step**2 + params.lr_steps**2) / params.lr_steps**2) ** -0.25 * - (((epoch**2 + params.lr_epochs**2) / params.lr_epochs**2) ** -0.25)) + (((epoch**2 + params.lr_epochs**2) / params.lr_epochs**2) ** -0.25))) if checkpoints and "optimizer" in checkpoints: