diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp index 16c26b857..e04e5b291 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train_lora.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py index 902a77438..326265a02 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train_lora.py @@ -138,16 +138,9 @@ class LoRAHook(): lora_alpha=1.0, ) def hook_fn(self, module, input, output): - #print('-'*20) - #print(input[0].size(), output.size()) - #print('-'*20) - - logging.info(output) - logging.info('-'*20) lora_out = self.lora(input[0]) + print(lora_out) output += lora_out - logging.info(output) - logging.info('*'*20) def save_checkpoint(self, i, iter_, save_dir): if isinstance(self.lora, DDP):