diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.bitfit.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.bitfit.py.swp index 477a01d72..dbee27a9d 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.bitfit.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.bitfit.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/bitfit.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/bitfit.py index adc77f8f9..b9ec2eb62 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/bitfit.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/bitfit.py @@ -1584,7 +1584,6 @@ def run_adapter(rank, world_size, args, wb=None): logging.info("Using DDP") model = DDP(model, device_ids=[rank], find_unused_parameters=True) - ''' adapter_names = [] adapter_param = [] for n, p in model.named_parameters(): @@ -1605,15 +1604,14 @@ def run_adapter(rank, world_size, args, wb=None): clipping_scale=5.0, parameters_names=[adapter_names], ) - ''' #prompt = torch.randn((100, 512), requires_grad=True) - optimizer_adapter = ScaledAdam( - [model.prompt], - lr=params.adapter_lr, - clipping_scale=5.0, - parameters_names=['P'], - ) + #optimizer_adapter = ScaledAdam( + # [model.prompt], + # lr=params.adapter_lr, + # clipping_scale=5.0, + # parameters_names=['P'], + #) scheduler_adapter = Eden(optimizer_adapter, 10000, 7) #params.lr_batche, params.lr_epochs)