diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp index a41135670..914514cac 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py index 2982a3294..d83874c08 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py @@ -1600,6 +1600,8 @@ def run_adapter(rank, world_size, args, wb=None): adapter_names = [] adapter_param = [] for n, p in model.named_parameters(): + print(n) + ''' if 'adapters' in n:# or 'joiner' in n or 'simple' in n or 'ctc' in n: adapter_names.append(n) adapter_param.append(p) @@ -1607,7 +1609,7 @@ def run_adapter(rank, world_size, args, wb=None): p.requires_grad = True else: p.requires_grad = False - + ''' optimizer_adapter = ScaledAdam( adapter_param, lr=params.adapter_lr,