diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp index 64f0f2468..649dc18b6 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py index 3b6162fc4..afe0a2d7f 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py @@ -493,9 +493,12 @@ class Data2VecAudioModel(BaseFairseqModel): print(features.size()) print(padding_mask.size()) print(padding_mask[0]) - conv_feat_all = None + conv_feat_all = torch.tensor([]) for i in range(padding_mask.size()[0]): - print(padding_mask[i].nonzero()) + nonzero = padding_mask[i].nonzero() + + conv_feat_all = + #print(padding_mask[i].nonzero()) print(padding_mask.nonzero()) print(padding_mask.nonzero().size())