From c89bc6784906904b5da79c959769c4300d3c5624 Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Fri, 26 May 2023 12:14:34 +0900 Subject: [PATCH] from local --- .../.train_lora.py.swp | Bin 81920 -> 81920 bytes .../train_lora.py | 9 +-------- 2 files changed, 1 insertion(+), 8 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp index 16c26b857e9c3ae6a44e02c5eec707b5872b4a83..e04e5b291136c77cbfa03d608b1018fcc51409e8 100644 GIT binary patch delta 717 zcmZ|NPe_w-7{Kvo=1TV`=LCg?Z7^c~OE(oH4KXNau2&IOCuz8L(Mn}ZMH#$r1ztpj zUOWhT&`KeKuovp&!AmZPOhiz$+Cg(*9=cYCzRT>^1D_wy?+-i=JU>k{Ynr(@>c16K zJeR8dLO4WLz5TH(%2Kh*VRGsLhuf4iIvhrgJ!KcEI(z)asTa1i=X=@N@X6?IxoLO% zF8V5J{pI5M8~+k-f4sXlp}$pX-2b^dIW{WWwoU4Ij>vTGaiwIoD`LAqWCaU&hB!j- z;d8bJ>|z|f2%#EQr0LwiLp*?ntEhw%KW!p=c!d`>QDu>20)vR66&K*fd6@CtD)I%Z zc#fxdf_e0y1I_TE5Jwh~1AM?T7BGn(#Bc|#2v}4Kk=XDpU*s$H@g4?dFp4nR5QHBk zIEBMJkts}K0t4v8bp)_;Mr0Eo@dmRPLjny5s4P_|!jID;8Ejw?bC|&++{O(wq8@c{ z;g^|)v4(d@;UykJ!(D`M4VTdb6EciE1oNZ4uBhm~^$ Wf4)=My}o!?q;DWmZfr%$(!T*kv3K`lQDaEy|2{VYA6wJzp-1H@=qFmDr|&BZliK0;x%TD2b&s1ez3CSTbcfF z;L3Qx9xs^Ut!7gPtWol_>2sZ9R`FuO?1*mdR;Q*!s}ZxWTBC(Td|y1-l{Wpo$Cef6 zQjN$&?bb_Em8ytkU8IbMn8pp9fC&|^spS=3;xhVRB7~pRSjH^QV-N$_gCI6S+>f`o z8xluuk<4HOIb@K)Hf(Me`G++u;SC<)Ca&WeT=XD?T~P5$6ZwSqc!~#@MjkyliX%v9 zj)+dO85CK^N4&!Vo?sS545JGjh@%020wUk>0b|Hx1gCKb?a=X}Uc|?9+{Yv=oI;bs zphLs&I*~6}L6SUYTiID(?K;oSw)<;?ZQuR_XzG3Y diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py index 902a77438..326265a02 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py @@ -138,16 +138,9 @@ class LoRAHook(): lora_alpha=1.0, ) def hook_fn(self, module, input, output): - #print('-'*20) - #print(input[0].size(), output.size()) - #print('-'*20) - - logging.info(output) - logging.info('-'*20) lora_out = self.lora(input[0]) + print(lora_out) output += lora_out - logging.info(output) - logging.info('*'*20) def save_checkpoint(self, i, iter_, save_dir): if isinstance(self.lora, DDP):