mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-12-11 06:55:27 +00:00
RemoveAttentionSqueeze
This commit is contained in:
parent
ec8804283c
commit
ed65330261
@ -497,7 +497,7 @@ class ZipformerEncoderLayer(nn.Module):
|
|||||||
cnn_module_kernel)
|
cnn_module_kernel)
|
||||||
|
|
||||||
|
|
||||||
self.attention_squeeze = AttentionSqueeze(embed_dim, embed_dim // 2)
|
#self.attention_squeeze = AttentionSqueeze(embed_dim, embed_dim // 2)
|
||||||
|
|
||||||
self.norm = BasicNorm(embed_dim)
|
self.norm = BasicNorm(embed_dim)
|
||||||
|
|
||||||
@ -642,10 +642,10 @@ class ZipformerEncoderLayer(nn.Module):
|
|||||||
|
|
||||||
src = src + self.feed_forward1(src)
|
src = src + self.feed_forward1(src)
|
||||||
|
|
||||||
# pooling module
|
## pooling module
|
||||||
if torch.jit.is_scripting() or use_self_attn:
|
#if torch.jit.is_scripting() or use_self_attn:
|
||||||
src = src + self.balancer_as(
|
# src = src + self.balancer_as(
|
||||||
self.attention_squeeze(src, selected_attn_weights[1:2]))
|
# self.attention_squeeze(src, selected_attn_weights[1:2]))
|
||||||
|
|
||||||
if torch.jit.is_scripting() or use_self_attn:
|
if torch.jit.is_scripting() or use_self_attn:
|
||||||
src = src + self.self_attn(
|
src = src + self.self_attn(
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user