diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp index 3d9205291..10dbffb81 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py index 206f1af9b..74d5cca45 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py @@ -1613,8 +1613,8 @@ def run_adapter(rank, world_size, args, wb=None): adapter_param = [] for i, lora in enumerate(lora_modules): for n, p in lora.lora.named_parameters(): - new_name = str(i) + n - adapter_names.append(n) + new_n = str(i) + n + adapter_names.append(new_n) adapter_param.append(p) for n, p in model.named_parameters():