mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-12-11 06:55:27 +00:00
Remove out_balancer of NonlinAttentionModule
This commit is contained in:
parent
9138695dfe
commit
b88f12fe83
@ -1443,16 +1443,6 @@ class NonlinAttentionModule(nn.Module):
|
|||||||
initial_scale=0.05)
|
initial_scale=0.05)
|
||||||
|
|
||||||
|
|
||||||
# put quite strict limits on the min_positive and max_positive at the output,
|
|
||||||
# because we noticed that poorly-trained instances of NonlinAttentionModule seem
|
|
||||||
# to have a larger mean-offset at the output for some reason.
|
|
||||||
self.out_balancer = ActivationBalancer(
|
|
||||||
channels, channel_dim=-1,
|
|
||||||
min_positive=0.45, max_positive=0.55,
|
|
||||||
min_abs=0.005, max_abs=1.0,
|
|
||||||
min_prob=0.05,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def forward(self,
|
def forward(self,
|
||||||
x: Tensor,
|
x: Tensor,
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user