From 0989aec74133ccc8aee9753699e7a20724910a06 Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Sun, 15 May 2022 14:33:08 +0800 Subject: [PATCH] Replace Eve optimizer with Abel. --- egs/librispeech/ASR/pruned_transducer_stateless4/train.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless4/train.py b/egs/librispeech/ASR/pruned_transducer_stateless4/train.py index 4ff69d521..a79f29c30 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless4/train.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless4/train.py @@ -66,7 +66,7 @@ from lhotse.cut import Cut from lhotse.dataset.sampling.base import CutSampler from lhotse.utils import fix_random_seed from model import Transducer -from optim import Eden, Eve +from optim import Eden, Eve, Abel from torch import Tensor from torch.cuda.amp import GradScaler from torch.nn.parallel import DistributedDataParallel as DDP @@ -871,7 +871,7 @@ def run(rank, world_size, args): logging.info("Using DDP") model = DDP(model, device_ids=[rank]) - optimizer = Eve(model.parameters(), lr=params.initial_lr) + optimizer = Abel(model.parameters(), lr=params.initial_lr) scheduler = Eden(optimizer, params.lr_batches, params.lr_epochs)