From d260b54177b3e30dc058e8867d23df511907ae3d Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Fri, 2 Dec 2022 15:55:48 +0800 Subject: [PATCH] Subtract, not add, 0.025. --- egs/librispeech/ASR/pruned_transducer_stateless7/scaling.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/scaling.py b/egs/librispeech/ASR/pruned_transducer_stateless7/scaling.py index 628714019..54c597ebc 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/scaling.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/scaling.py @@ -1119,7 +1119,7 @@ class DoubleSwishFunction(torch.autograd.Function): d_int = d_scaled.to(torch.uint8) ctx.save_for_backward(d_int) # on wolframalpha, do: (x * sigmoid(x-1) - 0.05 * x + 0.05 * min(0.15, max(-0.15, x)) + 0.025) from x=-3 to 2 - y = y + alpha * x + beta * x.clamp(min=-x_limit, max=x_limit) + 0.025 + y = y + alpha * x + beta * x.clamp(min=-x_limit, max=x_limit) - 0.025 if x.dtype == torch.float16 or torch.is_autocast_enabled(): y = y.to(torch.float16) return y