From f4c33cf7ea2689b0e7debdd1d4c90a8568da3f61 Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Thu, 25 May 2023 21:06:16 +0900 Subject: [PATCH] from local --- .../.train_lora.py.swp | Bin 86016 -> 86016 bytes .../train_lora.py | 4 ++-- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp index be2cc259f6eb7be6e6846a0942b7311481ea5ec8..c768fb7efd3ed44bb4c4e070201f48ac152c44e8 100644 GIT binary patch delta 444 zcmWm9!7GDt9LMqRZ|#|9WA>;$r=jJf*laBeDTkfZw04=xQka`E2f3I{9JB+isfXKg za-oFg570PClQL5H16U4{cJTJCSAFW!=kP65v_eHQ=R&E;DgQ_?ESw?YO)O>lHp~lC zeGX}>HW82I45>(dy;CHP05rV0MPBiU6RaSK09|At;X~6Qa*Z0c zv55s_5OqxRktXtu8=T`9N2p*2YnVkGW9UZ@On6bZi|itWBn)_}rLfA`s?S}jrdz|h iGOU+z)hb_JxvfPXA$xBxH-mPo(5<5Xp*k?39*jRJZchaO delta 452 zcmWN|O(?^09LMqRZ}VriS*8}&iUX#^Hc!dFZ5Ia*Ewuwu@{ps4gOZCjr5#97^t(us z(v(PPXX7L-4pT1fc_B(=o?eH^oOZJ`ZTPX^lztf>X)!OxBC75Ogf?l-T%e)O@#me diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py index 60d25ee0e..33a1da0d2 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py @@ -1616,12 +1616,12 @@ def run_adapter(rank, world_size, args, wb=None): adapter_names = [] adapter_param = [] - ''' for i, lora in enumerate(lora_modules): for n, p in lora.lora.named_parameters(): new_n = str(i) + n adapter_names.append(new_n) adapter_param.append(p) + ''' for n, p in model.named_parameters(): if 'joiner' in n or 'simple' in n or 'ctc' in n: @@ -1630,7 +1630,7 @@ def run_adapter(rank, world_size, args, wb=None): p.requires_grad = True else: p.requires_grad = False - + ''' #for lora in lora_modules: # print(lora.lora.state_dict()) #print(adapter_names)