diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp index 60b856dab..6b4aa53a1 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py index 735210c61..f07c5663b 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py @@ -1130,13 +1130,17 @@ def run(rank, world_size, args, wb=None): if params.multi_optim: logging.info("Using seperate optimizers over encoder, decoder ...") + enc_param = [] + enc_names = [] dec_names = [] dec_param = [] + for n, p in model.named_parameters(): name = n.split('.')[1] if name == 'encoder' and 'feature_extractor' not in n: + enc_names.append(n) enc_param.append(p) elif 'feature_extractor' not in n: dec_names.append(n)