From 87df9f321504064cfc597abc3be07ac783a46256 Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Wed, 14 Dec 2022 21:37:32 +0800 Subject: [PATCH] Simplify schedules of output balancers for nonlin_attention_module and attention_squeeze. --- egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py index 8ff3855c7..5764b3f10 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py @@ -1406,7 +1406,7 @@ class AttentionSqueeze(nn.Module): self.out_balancer = ActivationBalancer( embed_dim, channel_dim=-1, min_positive=0.3, max_positive=0.7, - min_abs=ScheduledFloat((0.0, 0.001), (8000.0, 0.01), (20000.0, 0.005)), + min_abs=ScheduledFloat((0.0, 0.001), (8000.0, 0.05)), ) @@ -1541,7 +1541,7 @@ class NonlinAttentionModule(nn.Module): self.balancer2 = ActivationBalancer( channels, channel_dim=-1, min_positive=0.3, max_positive=0.7, - min_abs=ScheduledFloat((0.0, 0.001), (8000.0, 0.01), (20000.0, 0.005)), + min_abs=ScheduledFloat((0.0, 0.001), (8000.0, 0.05)), )