From 05c7cb5c83cb4862d7359acd7f162f00e373b5e2 Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Sun, 11 Dec 2022 18:51:01 +0800 Subject: [PATCH] Reduce attention_squeeze dim from 512 to 128. --- egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py index f7142bab3..62a38e0d5 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py @@ -446,7 +446,7 @@ class ZipformerEncoderLayer(nn.Module): cnn_module_kernel) - self.attention_squeeze = AttentionSqueeze(embed_dim, 512) + self.attention_squeeze = AttentionSqueeze(embed_dim, 128) self.norm_final = BasicNorm(embed_dim)