From 36bff9b369b351f219a66a3f3902a4bcb52179ef Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Mon, 7 Nov 2022 12:33:12 +0800 Subject: [PATCH] Fix to comment --- egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py index f7efcf458..f37530483 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/zipformer.py @@ -1559,7 +1559,7 @@ class FeedforwardModule(nn.Module): class NonlinAttentionModule(nn.Module): """This is like the ConvolutionModule, but refactored so that we use multiplication by attention weights (borrowed - from the attention module) in palce of actual convolution. + from the attention module) in place of actual convolution. Args: channels (int): The number of channels of conv layers.