From 30c6e5b92937d368b15fe49ee98aa9fce0c106c3 Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Sat, 10 Dec 2022 00:08:38 +0800 Subject: [PATCH] Make attention_squeeze use full dim. --- egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py index 3ea11189a..6476f0aa0 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py @@ -444,7 +444,7 @@ class ZipformerEncoderLayer(nn.Module): cnn_module_kernel) - self.attention_squeeze = AttentionSqueeze(embed_dim, embed_dim // 2) + self.attention_squeeze = AttentionSqueeze(embed_dim, embed_dim) self.norm_final = BasicNorm(embed_dim)