From 8b4a2ecfe82b6051ad6cd07c4f14822307be8aab Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Sun, 8 Jan 2023 18:48:33 +0900 Subject: [PATCH] from local --- .../.conformer.py.swp | Bin 90112 -> 90112 bytes .../.train.py.swp | Bin 61440 -> 61440 bytes .../conformer.py | 3 ++- 3 files changed, 2 insertions(+), 1 deletion(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/.conformer.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/.conformer.py.swp index f8183062f34f80ab61b311076e7f1bc25267e7d0..b641a3e9a3559da8a3856978e49b5a49f3fbd49e 100644 GIT binary patch delta 286 zcmXBOF-rmg0LAflh;o^xQ5m&#kW(DT*;Wt)3Q^D+HYrAg5GGoSB;8hP+|bz8T-^}Z zVjmzBf}%+pf*TrcZL&Y~;0JGb!}I!auOH9V+`a8xYt3;*vQqQ9cMvW-37$>D44A4tzMYj|W$EaZ*A6$ILGtN0EFQSAtA=rMugCzY6mBw5(U9->Za(Zpo=!*UvO~H%;I3l5K(ZmyL_7{ z^be>35$u#*C}hyVNni2cfg5hPr0*nsCyb-*wasRGE)of(diVa|rG6I None: super(Conformer, self).__init__() @@ -124,7 +125,7 @@ class Conformer(EncoderInterface): self.encoder = ConformerEncoder(encoder_layer, num_encoder_layers) self._init_state: List[torch.Tensor] = [torch.empty(0)] - self.group_size = 6 + self.group_size = self.alpha = nn.Parameter(torch.rand(self.group_size)) self.sigmoid = nn.Sigmoid() self.layer_norm = nn.LayerNorm(512)