diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py index f7efcf458..f37530483 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py @@ -1559,7 +1559,7 @@ class FeedforwardModule(nn.Module): class NonlinAttentionModule(nn.Module): """This is like the ConvolutionModule, but refactored so that we use multiplication by attention weights (borrowed - from the attention module) in palce of actual convolution. + from the attention module) in place of actual convolution. Args: channels (int): The number of channels of conv layers.