diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp index f8a6afc8b..6c99a21d1 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py index 43fd7993b..cbc3430ab 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py @@ -1519,8 +1519,8 @@ def run_adapter(rank, world_size, args, wb=None): adapter_param = [] for n, p in model.named_parameters(): if 'adapters' in n: - logging.info(n) - exit() + adapter_names.append(n) + adapter_param.append(p) if params.multi_optim: logging.info("Using seperate optimizers over encoder, decoder ...")