From 6eb9a0bc9bd62307a73fc52a52147f282f540b31 Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Tue, 20 Sep 2022 14:39:17 +0800 Subject: [PATCH] Halve max_var_per_eig to 0.05 --- egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py b/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py index 7d785a369..328cb4434 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py @@ -201,7 +201,7 @@ class ConformerEncoderLayer(nn.Module): d_model, channel_dim=-1, min_positive=0.45, max_positive=0.55, max_abs=6.0, - max_var_per_eig=0.1, + max_var_per_eig=0.05, ) self.dropout = nn.Dropout(dropout) @@ -469,7 +469,7 @@ class RelPositionMultiheadAttention(nn.Module): self.in_proj = nn.Linear(embed_dim, 3 * embed_dim, bias=True) self.in_balancer = ActivationBalancer(3 * embed_dim, channel_dim=-1, max_abs=5.0, - max_var_per_eig=0.1) + max_var_per_eig=0.05) self.proj_balancer = ActivationBalancer(embed_dim, channel_dim=-1, max_abs=10.0, min_positive=0.0, max_positive=1.0)