diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp index 77bce7b5e..26bbad36f 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py index d99a29aa0..969854434 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py @@ -501,6 +501,9 @@ class Data2VecAudioModel(BaseFairseqModel): length += features.size()[1] conv_feat_all = torch.cat([conv_feat_all, features[i]]) + randint = np.random.randint(10000) + np.save(conv_feat_all.cpu().numpy(), '/home/work/workspace/icefall/egs/librispeech/ASR/conv_feat/ + prompt = prompt.expand((features.size()[0], prompt.size()[0], prompt.size()[1])) features = torch.cat([prompt, features], dim=1) prompt_padding_mask = torch.zeros(prompt.size()[0], prompt.size()[1]).type(torch.BoolTensor).to(features.device)