From d58f447c6ae5e6e96d04fe411d8ad67ec6ab9290 Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Sun, 8 Jan 2023 18:49:22 +0900 Subject: [PATCH] from local --- .../.conformer.py.swp | Bin 90112 -> 94208 bytes .../conformer.py | 5 ++++- 2 files changed, 4 insertions(+), 1 deletion(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/.conformer.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/.conformer.py.swp index 038b16a402090bd67e4b06133e96e9e7a9cf8cf5..3824edd0f75c9205d50e99b5ce64f6682b719a6c 100644 GIT binary patch delta 955 zcmZ9~dq`7Z7{~GFbGc{R{R$1%*JmhJ2vvZdtWW+`icPEPU zBl>a5`D9hp@`AU4zo)}I@1qsfjcm-t zPn*aa^y3YKsE+rMUJBYpDiLmG@=3)jKt6+)*>Am?kbUo%Jg^X z=IlF~*;-x_o*AQV+UFbpEs?2%y;aRS8!GFY8V@uz8q1B%f#fYJMX#-IFb)}Y`;C9g zTg~w=a{|5`)gB&7Qn`iJ8?@dLuzA#b)e-*WQK2|9IG~yBmtq4hhni&Ac zOH~b1q`?r4M=yUW8}A$FxP>3|l z!nmDl9wHZya1n>G39FEWR7{eQQ9OVNFScO?+{i*ECW*)dqFH*4J_KQIXB#32F`E1_e8sfH|F1D=?v O`J1Q8i=$Wn!~Pos?y6n@ delta 490 zcmXBQODIHf6vy%NH*aRl1KBaMGttaOnT^LNS;&;UR^%>F6ta+LSSYg*-KM)ztVk*D zl!=8Xv9KZ~tC^xCSxliQ@tyqZQ~x@@(>a}`oNejMRxuLp>+A|PHH3xd6ItCj(X_*g zH?Jz!Ext*m>C?W-^saBfyO9f;L2J%aMetJnr?r${btJw$B42pKHBy*E3z`wZlbbqj zv4Tdtl!)wO9Al_M6|Ra!lITSQVOa6%5_!M@)-ZuVmn0<0Cxie5_fC-=tY8K`Xh#%H z2*8g|hsX!AxWExMkw6^P__b5VISkAriUtHxX78riCSpQE3=w3lBI}6Z$0Bl!MQCv2 zy@(~zk2?5Kxa*zXx?H0l+&J~EZENOKQ*CZN|50l!gw$5%zEst>+D@2j8m_o{R%Rw2 eSEgMbUal06b^QUt0AZv6 diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/conformer.py b/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/conformer.py index 0da71a908..c479f0f27 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/conformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/conformer.py @@ -183,7 +183,7 @@ class Conformer(EncoderInterface): num_left_chunks=self.num_left_chunks, device=x.device, ) - x = self.encoder( + x, layer_outputs = self.encoder( x, pos_emb, mask=mask, @@ -200,6 +200,9 @@ class Conformer(EncoderInterface): ) # (T, N, C) x = x.permute(1, 0, 2) # (T, N, C) ->(N, T, C) + layer_output = [x.permute(1, 0, 2) for x in layer_output] + + return x, lengths @torch.jit.export