From 4fe91ce67ce1bf112bb914b9cc6ab4e27fb61ce1 Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Sat, 14 Jan 2023 17:19:34 +0800 Subject: [PATCH] Double hidden_channels in NonlinAttention from embed_dim//4 to embed_dim//2. --- egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py index 1595d0544..2da05f5b1 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py @@ -489,7 +489,7 @@ class ZipformerEncoderLayer(nn.Module): dropout) self.nonlin_attention = NonlinAttention(embed_dim, - hidden_channels=embed_dim // 4) + hidden_channels=embed_dim // 2) self.small_conv_module = SmallConvolutionModule(embed_dim)