From 072776b2a13afcd8d6dfca1f5cfa5833cc49bb37 Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Sat, 29 Oct 2022 20:11:39 +0800 Subject: [PATCH] Apply layer-skip dropout prob --- egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py index 706d9ffa2..95eed36e4 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py @@ -256,7 +256,9 @@ class Zipformer(EncoderInterface): for i, module in enumerate(self.encoders): ds = self.zipformer_downsampling_factors[i] if self.skip_layers[i] is not None: - x = self.skip_modules[i](outputs[self.skip_layers[i]], x) + layer_skip_dropout_prob = 0.05 + if (not self.training) or random.random() > layer_skip_dropout_prob: + x = self.skip_modules[i](outputs[self.skip_layers[i]], x) x = module(x, feature_mask=feature_masks[i], src_key_padding_mask=None if mask is None else mask[...,::ds])