From 42c152f5cb180e2d2e8210b15a54df3533a73892 Mon Sep 17 00:00:00 2001 From: Triplecq Date: Sun, 14 Jan 2024 12:12:15 -0500 Subject: [PATCH] decrease learning-rate to solve the error: RuntimeError: grad_scale is too small, exiting: 5.820766091346741e-11 --- egs/reazonspeech/ASR/zipformer/train.py | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) diff --git a/egs/reazonspeech/ASR/zipformer/train.py b/egs/reazonspeech/ASR/zipformer/train.py index ddd089176..cfd7fe909 100755 --- a/egs/reazonspeech/ASR/zipformer/train.py +++ b/egs/reazonspeech/ASR/zipformer/train.py @@ -30,7 +30,6 @@ export CUDA_VISIBLE_DEVICES="0,1,2,3" --start-epoch 1 \ --use-fp16 1 \ --exp-dir zipformer/exp \ - --full-libri 1 \ --max-duration 1000 # For streaming model training: @@ -41,7 +40,6 @@ export CUDA_VISIBLE_DEVICES="0,1,2,3" --use-fp16 1 \ --exp-dir zipformer/exp \ --causal 1 \ - --full-libri 1 \ --max-duration 1000 It supports training with: @@ -330,7 +328,7 @@ def get_parser(): ) parser.add_argument( - "--base-lr", type=float, default=0.045, help="The base learning rate." + "--base-lr", type=float, default=0.035, help="The base learning rate." ) parser.add_argument(