From 1d20c12bc02f544de828906956b4d29064673bc8 Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Thu, 22 Sep 2022 12:28:35 +0800 Subject: [PATCH] Increase max_var_per_eig to 0.2 --- egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py b/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py index 328cb4434..182b78eee 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py @@ -201,7 +201,7 @@ class ConformerEncoderLayer(nn.Module): d_model, channel_dim=-1, min_positive=0.45, max_positive=0.55, max_abs=6.0, - max_var_per_eig=0.05, + max_var_per_eig=0.2, ) self.dropout = nn.Dropout(dropout) @@ -469,7 +469,7 @@ class RelPositionMultiheadAttention(nn.Module): self.in_proj = nn.Linear(embed_dim, 3 * embed_dim, bias=True) self.in_balancer = ActivationBalancer(3 * embed_dim, channel_dim=-1, max_abs=5.0, - max_var_per_eig=0.05) + max_var_per_eig=0.2) self.proj_balancer = ActivationBalancer(embed_dim, channel_dim=-1, max_abs=10.0, min_positive=0.0, max_positive=1.0)