From 300da1306d2bf2ef9526803b2caa66d1bd8ddd70 Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Sat, 8 Oct 2022 12:16:53 +0800 Subject: [PATCH] Add warmup schedule where dropout disappears from earlier layers first. --- .../ASR/pruned_transducer_stateless7/conformer.py | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py b/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py index 32c61e81d..41a847a1c 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py @@ -49,8 +49,7 @@ class Conformer(EncoderInterface): layer_dropout (float): layer-dropout rate. cnn_module_kernel (int): Kernel size of convolution module vgg_frontend (bool): whether to use vgg frontend. - warmup_batches (float): number of batches to warm up over (gradually skip - layer bypass) + warmup_batches (float): number of batches to warm up over """ def __init__( @@ -65,7 +64,7 @@ class Conformer(EncoderInterface): num_encoder_layers: Tuple[int] = (12, 12), dropout: float = 0.1, cnn_module_kernel: Tuple[int] = (31, 31), - warmup_batches: float = 3000, + warmup_batches: float = 6000.0, ) -> None: super(Conformer, self).__init__()