From e771472a30818f00a8a5503477940854c4699e70 Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Sat, 28 May 2022 11:30:45 +0800 Subject: [PATCH] Speed up learning rate schedule. --- egs/librispeech/ASR/pruned_transducer_stateless4/train.py | 8 ++------ 1 file changed, 2 insertions(+), 6 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless4/train.py b/egs/librispeech/ASR/pruned_transducer_stateless4/train.py index 2e2f97d98..ac7da8468 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless4/train.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless4/train.py @@ -168,7 +168,7 @@ def get_parser(): parser.add_argument( "--lr-batches", type=float, - default=5000, + default=3000, help="""Number of steps that affects how rapidly the learning rate decreases. We suggest not to change this.""", ) @@ -176,7 +176,7 @@ def get_parser(): parser.add_argument( "--lr-epochs", type=float, - default=6, + default=4, help="""Number of epochs that affects how rapidly the learning rate decreases. """, ) @@ -881,10 +881,6 @@ def run(rank, world_size, args): logging.info("Loading optimizer state dict") optimizer.load_state_dict(checkpoints["optimizer"]) - for p in optimizer.param_groups: - p['max_eff_lr'] = params.initial_lr # TEMP!! - - if ( checkpoints and "scheduler" in checkpoints