diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp index 1c40d0ebc..f7adb4988 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py index f85962ef5..05f2a6422 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py @@ -493,7 +493,7 @@ class Data2VecAudioModel(BaseFairseqModel): print(features.size()) print(padding_mask.size()) print(padding_mask.nonzero()) - print(padding_mask.nonezero().size()) + print(padding_mask.nonzero().size()) exit() prompt = prompt.expand((features.size()[0], prompt.size()[0], prompt.size()[1])) features = torch.cat([prompt, features], dim=1)