diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp index 6c99a21d1..80da68d90 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py index cbc3430ab..ccf4b5a84 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py @@ -1521,6 +1521,13 @@ def run_adapter(rank, world_size, args, wb=None): if 'adapters' in n: adapter_names.append(n) adapter_param.append(p) + + optimizer_adapter = ScaledAdam( + adapter_, + lr=params.peak_enc_lr, + clipping_scale=None, + parameters_names=[enc_names], + ) if params.multi_optim: logging.info("Using seperate optimizers over encoder, decoder ...")