From 06dca57f4b9f01ef9919230b732acacf003bbabd Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Mon, 9 Jan 2023 19:27:00 +0900 Subject: [PATCH] from local --- .../ASR/incremental_transf/.conformer.py.swp | Bin 98304 -> 102400 bytes .../ASR/incremental_transf/conformer.py | 4 ++-- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/egs/librispeech/ASR/incremental_transf/.conformer.py.swp b/egs/librispeech/ASR/incremental_transf/.conformer.py.swp index ef2213610068d72fbab5c933df6affdc453976f4..b1b0db93df1984a5df316b2475c9009d45097ec8 100644 GIT binary patch delta 918 zcmX}qT}YEr9LMqh+2-1uy6s|Zf#nkgZD^*!E{rXRs1P(7gh7pbSyqrzI7bN87UV@l zH!WRBi1dP2QYyMkD3snP771Z^(M{TwDAT(nUEilx2R_d^I}guxp5JeYggG%_cC}PB z)YfY?l~uxUh`d>xj7$u>mh8%%G&*)Gf3n))c6#)?4By{LZjaVD2JIu7DO()aBvN5v zpNIV^r$c>>>mTRY7UosvALzvm_~C*PV`hn@;9>fKN@|}OD;~pG>=TG0mQ+m>ROA8{3XV89BgP9p{{vLw2IY1~H>8c~ULLex`!8{|UNaue;? r2@_tEUIM33ivpN1#~~+h2Vv`|GCj2DZ&R3jHe>Fs=dJh?cAI|#S2>&( delta 665 zcmXBROK1~O6vpxIHqA5-GcyHi66|8A3m3ks7SXOsMWRSWp(uhxY84AAwQVB!z+_`U z6iSIY#YL?KblWJj1tD5o2;xG~=%y&*vItYx*7y%z_;JtT+{@v7#aXLZv^vgoA3rqU z?c3fhjw_W`UJl=^cKmY8hH}W=X0qj-Zt}&5JMBz;@Sc0EK_X7$BlEvI>$kh+Yhiyx zN+OQ6HYvbkOrj4SY`nIlaU4Jo!dMJT7qARKXLBp${=c@gpR?#~jWh6Vev7 zQM_r9W^fk=kwGW^Fl!ynnP-e&H6~e}`)EBFFwVmD+vb+#7YoMrdsC*j+{kbABiD?( zFrP9D?=EE@7<84B^gAAfrTPHom2zubc z!B2Mk4vUzKtDs91hp`uF#IPEZzGEIGoW(Kh!X~7V!fKS^QO9T0P{DoNKo&djm(aeU zhFM&}Nu(o#Qi7sEa=%f*6#CHx7c~+bMHiMyZW^c1j}_ATgcA0knO+d8E^k_bJIiH)5{sE_Aab5rb diff --git a/egs/librispeech/ASR/incremental_transf/conformer.py b/egs/librispeech/ASR/incremental_transf/conformer.py index 1006fd0d8..98605c791 100644 --- a/egs/librispeech/ASR/incremental_transf/conformer.py +++ b/egs/librispeech/ASR/incremental_transf/conformer.py @@ -136,7 +136,7 @@ class Conformer(EncoderInterface): self._init_state: List[torch.Tensor] = [torch.empty(0)] def forward( - self, x: torch.Tensor, x_lens: torch.Tensor, warmup: float = 1.0 + self, x: torch.Tensor, x_lens: torch.Tensor, warmup: float = 1.0, layer_output = False, ) -> Tuple[torch.Tensor, torch.Tensor]: """ Args: @@ -198,7 +198,7 @@ class Conformer(EncoderInterface): x = x.permute(1, 0, 2) # (T, N, C) ->(N, T, C) layer_outputs = [x.permute(1, 0, 2) for x in layer_outputs] - return x, lengths, layer_outputs + return x, lengths @torch.jit.export def get_init_state(