Add warmup schedule where dropout disappears from earlier layers first.

This commit is contained in:
Daniel Povey 2022-10-08 12:16:53 +08:00
parent 9c1a239931
commit 300da1306d

View File

@ -49,8 +49,7 @@ class Conformer(EncoderInterface):
layer_dropout (float): layer-dropout rate. layer_dropout (float): layer-dropout rate.
cnn_module_kernel (int): Kernel size of convolution module cnn_module_kernel (int): Kernel size of convolution module
vgg_frontend (bool): whether to use vgg frontend. vgg_frontend (bool): whether to use vgg frontend.
warmup_batches (float): number of batches to warm up over (gradually skip warmup_batches (float): number of batches to warm up over
layer bypass)
""" """
def __init__( def __init__(
@ -65,7 +64,7 @@ class Conformer(EncoderInterface):
num_encoder_layers: Tuple[int] = (12, 12), num_encoder_layers: Tuple[int] = (12, 12),
dropout: float = 0.1, dropout: float = 0.1,
cnn_module_kernel: Tuple[int] = (31, 31), cnn_module_kernel: Tuple[int] = (31, 31),
warmup_batches: float = 3000, warmup_batches: float = 6000.0,
) -> None: ) -> None:
super(Conformer, self).__init__() super(Conformer, self).__init__()