From e9326a7d1649a0174e5b80618f6921018d1f519e Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Sat, 1 Oct 2022 13:13:10 +0800 Subject: [PATCH] Remove dropout from inside ConformerEncoderLayer, for adding to residuals --- .../ASR/pruned_transducer_stateless7/conformer.py | 13 +++++-------- 1 file changed, 5 insertions(+), 8 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py b/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py index 85049f915..283d89268 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py @@ -205,8 +205,6 @@ class ConformerEncoderLayer(nn.Module): max_var_per_eig=0.2, ) - self.dropout = nn.Dropout(dropout) - def forward( self, @@ -256,7 +254,7 @@ class ConformerEncoderLayer(nn.Module): alpha = 1.0 # macaron style feed forward module - src = src + self.dropout(self.feed_forward_macaron(src)) + src = src + self.feed_forward_macaron(src) # multi-headed self-attention module src_att, _, attn_scores_out = self.self_attn( @@ -266,17 +264,16 @@ class ConformerEncoderLayer(nn.Module): attn_mask=src_mask, key_padding_mask=src_key_padding_mask, ) - src = src + self.dropout(src_att) + src = src + src_att # convolution module - src = src + self.dropout( - self.conv_module(src, src_key_padding_mask=src_key_padding_mask) - ) + src = src + self.conv_module(src, src_key_padding_mask=src_key_padding_mask) + # feed forward module - src = src + self.dropout(self.feed_forward(src)) + src = src + self.feed_forward(src) src = self.norm_final(self.balancer(src))