From 909ec64842244bc7c93b6b2557b07bcbd6a930fa Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Thu, 2 Feb 2023 13:52:02 +0900 Subject: [PATCH] from local --- .../ASR/conformer_ctc/.transformer.py.swp | Bin 65536 -> 65536 bytes egs/aishell/ASR/conformer_ctc/transformer.py | 2 +- 2 files changed, 1 insertion(+), 1 deletion(-) diff --git a/egs/aishell/ASR/conformer_ctc/.transformer.py.swp b/egs/aishell/ASR/conformer_ctc/.transformer.py.swp index f74656bbae8a1b399d3fcbcd957d2200994471ac..1752eb638b567054f15a46dc74d5ac386b41b7dd 100644 GIT binary patch delta 71 zcmZo@U}%lps{jDz^B8CV delta 71 zcmZo@U}%lps{jBBPZ=`+ diff --git a/egs/aishell/ASR/conformer_ctc/transformer.py b/egs/aishell/ASR/conformer_ctc/transformer.py index bcc4cf6f9..f2990c514 100644 --- a/egs/aishell/ASR/conformer_ctc/transformer.py +++ b/egs/aishell/ASR/conformer_ctc/transformer.py @@ -409,7 +409,7 @@ class TransformerEncoder(Module): self.enable_nested_tensor = enable_nested_tensor self.mask_check = mask_check - def forward(self, src: Tensor, mask: Optional[Tensor] = None, src_key_padding_mask: Optional[Tensor] = None) -> Tensor: + def forward(self, src: Tensor, mask: Optional[Tensor] = None, src_key_padding_mask: Optional[Tensor] = None) -> Tensor: """Pass the input through the encoder layers in turn. Args: