diff --git a/egs/librispeech/ASR/.run_adapter.sh.swp b/egs/librispeech/ASR/.run_adapter.sh.swp index 9b870dde3..c2bb7abb5 100644 Binary files a/egs/librispeech/ASR/.run_adapter.sh.swp and b/egs/librispeech/ASR/.run_adapter.sh.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp index 56d70d35a..7853181ce 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py index 84e625615..ab569966f 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py @@ -280,8 +280,8 @@ class Data2VecAudioModel(BaseFairseqModel): torch.FloatTensor(cfg.encoder_embed_dim).uniform_() ) - self.encoder = TransformerEncoder(cfg) - #self.encoder = TransformerEncoderAdapter(cfg) + #self.encoder = TransformerEncoder(cfg) + self.encoder = TransformerEncoderAdapter(cfg) self.layer_norm = LayerNorm(self.extractor_embed) self.final_proj = nn.Linear(self.embed, self.embed)