From a00ed7e976cf49cd47e80bffea5e140dbed229f4 Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Fri, 9 Dec 2022 20:05:50 +0800 Subject: [PATCH] Decrease min_abs of NonlinAttentionModule from 0.75 to 0.5; make its max_abs (not active) a constant. --- egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py index 735f08d98..37be53c2a 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py @@ -1489,8 +1489,8 @@ class NonlinAttentionModule(nn.Module): hidden_channels, channel_dim=-1, min_positive=ScheduledFloat((0.0, 0.25), (20000.0, 0.05)), max_positive=ScheduledFloat((0.0, 0.75), (20000.0, 0.95)), - min_abs=0.75, - max_abs=ScheduledFloat((0.0, 2.5), (8000.0, 5.0), default=1.0), + min_abs=0.5, + max_abs=5.0, ) self.tanh = nn.Tanh()