From f7778dfb7e1826be397511b9fae6ee7adb62d769 Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Sun, 8 Jan 2023 18:48:56 +0900 Subject: [PATCH] from local --- .../.conformer.py.swp | Bin 90112 -> 90112 bytes .../conformer.py | 2 +- 2 files changed, 1 insertion(+), 1 deletion(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/.conformer.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/.conformer.py.swp index 2b36e4fe448a361431e89f24bc27bdab0e454808..ca774cc22ada7166c1937930274955c110e6ac70 100644 GIT binary patch delta 67 zcmZoTz}j$tRV>LM%+puFQqO<^2m}}y$|mkgwvyT?_U=8a7Y74_++@xVhZ*HIpZyTX WFLM%+puFQqO<^2m}}y3MTGKej&M0?A?1-KMn>4rOBKh4l^okKKmh% YgC#j5zId|SPjzNfL!-?;Kewa-0N}zGfB*mh diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/conformer.py b/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/conformer.py index 782b0f882..218448f49 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/conformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_gtrans/conformer.py @@ -128,7 +128,7 @@ class Conformer(EncoderInterface): self.group_size = group_size self.alpha = nn.Parameter(torch.rand(self.group_size)) self.sigmoid = nn.Sigmoid() - self.layer_norm = nn.LayerNorm(512) + self.layer_norm = nn.LayerNorm(d_model) def forward( self, x: torch.Tensor, x_lens: torch.Tensor, warmup: float = 1.0