diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.optim.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.optim.py.swp index ed4f747e0..af20f364e 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.optim.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.optim.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.prompt_tuning.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.prompt_tuning.py.swp index 9586782e5..14bf7df31 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.prompt_tuning.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.prompt_tuning.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/prompt_tuning.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/prompt_tuning.py index 67f6abf81..4ad3dc93a 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/prompt_tuning.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/prompt_tuning.py @@ -1579,7 +1579,6 @@ def run_adapter(rank, world_size, args, wb=None): logging.info("Using DDP") model = DDP(model, device_ids=[rank], find_unused_parameters=True) - ''' adapter_names = [] adapter_param = [] for n, p in model.named_parameters(): @@ -1590,14 +1589,12 @@ def run_adapter(rank, world_size, args, wb=None): p.requires_grad = True else: p.requires_grad = False - ''' for n, p in model.named_parameters(): p.requires_grad = False prompt = torch.nn.Parameter(torch.randn(50, 512)).to(device) - ''' optimizer_adapter = ScaledAdam( adapter_param, lr=params.adapter_lr, @@ -1611,7 +1608,7 @@ def run_adapter(rank, world_size, args, wb=None): clipping_scale=5.0, parameters_names=['prompt'], ) - + ''' scheduler_adapter = Eden(optimizer_adapter, 10000, 7) #params.lr_batche, params.lr_epochs) optimizer, scheduler = optimizer_adapter, scheduler_adapter