From 7e7e3105e0cc0619fe16ec03f80c0126a479d47b Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Tue, 3 Jan 2023 17:05:36 +0900 Subject: [PATCH] from local --- .../.data2vec_audio.py.swp | Bin 45056 -> 45056 bytes .../data2vec_audio.py | 5 +++-- 2 files changed, 3 insertions(+), 2 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_audio.py.swp index 62292f92a5af0cd836e9102fc8811dba1adc5b10..a2f694473d00653a39307e56fd9d5637f6aec2f4 100644 GIT binary patch delta 449 zcmXBQ&ntrg9LMqRe%Nd?Jx}T(KkB(j%l1qn2joX-`7v3v3pL}UOwq1eJ1Vt#rlfE+ z2TRXMNO5uI;Nmbx%gP@x@y4fK^{LP8ldagZ6}z@-EJbGZp`al|5h=Y{$#wlwYgcS> zE37J%-n`*v0^VBR-zzn8g$Ae$w6iI}jLcwd3$Q3M@*vAHz5k)^b;l`Iw zq>dMqaE3JIFo*#tKFLVUEAoL`6tRdYwBX(&l1C0P*zwsYQpN@LkisTn7{?e?+_^=z z5yiJlq=p+@;}A)#x+X;?D7|QfhC2UL@s3BFql5y|h+zRC{C@fFv~13l!~7_=ne)f$ tpOLcH8a?Rtn7e^4b0yHx&}RgD#LNXEicYqIYl)qN9varehE+Y$)F08+PT>Fm delta 425 zcmXBQJxBrp7{>9}(=08|6U3~g<0LdB(J_L`C30x$LmG;#q_IWYLK^~yhTx30ggy*f zN^Z-B8k(Xas-?-cT$>oYw<9XrTo?=5OHk80pd^RKwr;X+jvS-?E}c=3qTkVg_B82IQCxy22RafA)5A%k(m&;u`2 zG&@D=xIhVuNMQm7T5c)Gk6YvcRjgtT{djhX9H5LW{7`V+A##NZHc>g%s9tHMmYu$fr@YT`NAtMQO6-l$YKRa{O$PZuxyTXyV)!UP0cVA eT{jbF-PYi4*kz7Hht2JMmDbIZXvzv!-nCze;z?)# diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py index aba350f42..a6cc1006d 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_audio.py @@ -42,8 +42,9 @@ class TransformerEncoderAdapter(TransformerEncoder): self.adapters = ResidualAdapterModule() for p in self.adapters.parameters(): - # p.data = nn.Parameter(torch.zeros(p.size()).to('cuda')) - p.data = nn.Parameter(torch.randn(p.size()).to('cuda')/20.) + p.data /= 10 + #p.data = nn.Parameter(torch.zeros(p.size()).to('cuda')) + #p.data = nn.Parameter(torch.randn(p.size()).to('cuda')/20.) def forward(self, x, padding_mask=None, layer=None, tgt_layer=None): x, layer_results = self.extract_features_with_adapter(