From 67ebc38e7d63fab979e3e79cd54d8a007a0b6b0b Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Tue, 11 Apr 2023 15:50:53 +0900 Subject: [PATCH] from local --- .../.data2vec_audio.py.swp | Bin 40960 -> 40960 bytes .../data2vec_audio.py | 1 + 2 files changed, 1 insertion(+) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp index dd04d7d1e5046984e11111e7a773fd2f39f4c86f..591c7ff5cc3f5fefd2ab594910f1a1eef2374771 100644 GIT binary patch delta 135 zcmZoTz|?SnNi@kI%+puFQqO<^2m}}y6q!v^Qa6fL3ov?Yo+wbP&uhfOz>voZk@T1> z=;AP)z=4iz-*(;N8vhiHm&AF~` L7&o`OS4IE;TzMOg diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py index aca3e34ac..32b06f290 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py @@ -497,6 +497,7 @@ class Data2VecAudioModel(BaseFairseqModel): features = torch.cat([prompt, features], dim=1) print(features.size()) prompt_padding_mask = torch.zeros(prompt.size()).type(torch.BoolTensor).to(features.device) + print(prompt_padding_mask.size()) padding_mask = torch.cat([prompt_padding_mask, padding_mask]) print(padding_mask.size())