From 7cc43f7bf653be586059c1c438bbd7682c256557 Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Thu, 2 Feb 2023 13:51:55 +0900 Subject: [PATCH] from local --- .../ASR/conformer_ctc/.transformer.py.swp | Bin 65536 -> 65536 bytes egs/aishell/ASR/conformer_ctc/transformer.py | 2 +- 2 files changed, 1 insertion(+), 1 deletion(-) diff --git a/egs/aishell/ASR/conformer_ctc/.transformer.py.swp b/egs/aishell/ASR/conformer_ctc/.transformer.py.swp index 90c94ccb1d5b38083307e4cfab5b3cbbdd694bd7..1256d0536761ca733c5054512848ddb49b5f170d 100644 GIT binary patch delta 56 zcmZo@U}98}KhN67#lXNdne#$6BiH8A3v;;` Li#7}2eyaiix>ymY diff --git a/egs/aishell/ASR/conformer_ctc/transformer.py b/egs/aishell/ASR/conformer_ctc/transformer.py index 996bb3803..bcc4cf6f9 100644 --- a/egs/aishell/ASR/conformer_ctc/transformer.py +++ b/egs/aishell/ASR/conformer_ctc/transformer.py @@ -410,7 +410,7 @@ class TransformerEncoder(Module): self.mask_check = mask_check def forward(self, src: Tensor, mask: Optional[Tensor] = None, src_key_padding_mask: Optional[Tensor] = None) -> Tensor: - r"""Pass the input through the encoder layers in turn. + """Pass the input through the encoder layers in turn. Args: src: the sequence to the encoder (required).