diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.bitfit.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.bitfit.py.swp index ab8435a5b..6d02f31b4 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.bitfit.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.bitfit.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/bitfit.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/bitfit.py index 9e673d7f0..35be7f5b7 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/bitfit.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/bitfit.py @@ -1602,16 +1602,15 @@ def run_adapter(rank, world_size, args, wb=None): else: p.requires_grad = False ''' - - for n, p in model.named_parameters(): - p.requires_grad = False - optimizer_adapter = ScaledAdam( adapter_param, lr=params.adapter_lr, clipping_scale=5.0, parameters_names=[adapter_names], ) + + #for n, p in model.named_parameters(): + # p.requires_grad = False #prompt = torch.randn((100, 512), requires_grad=True) #optimizer_adapter = ScaledAdam(