From f2387fe523c6f89987f3723bfa967095e8de5127 Mon Sep 17 00:00:00 2001 From: Fangjun Kuang Date: Thu, 14 Oct 2021 20:09:38 +0800 Subject: [PATCH] Fix a bug introduced while supporting torch script. (#79) --- egs/librispeech/ASR/conformer_ctc/transformer.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/egs/librispeech/ASR/conformer_ctc/transformer.py b/egs/librispeech/ASR/conformer_ctc/transformer.py index a2e36a41e..3e6abb695 100644 --- a/egs/librispeech/ASR/conformer_ctc/transformer.py +++ b/egs/librispeech/ASR/conformer_ctc/transformer.py @@ -660,7 +660,7 @@ class PositionalEncoding(nn.Module): self.xscale = math.sqrt(self.d_model) self.dropout = nn.Dropout(p=dropout) # not doing: self.pe = None because of errors thrown by torchscript - self.pe = torch.zeros(0, self.d_model, dtype=torch.float32) + self.pe = torch.zeros(0, 0, dtype=torch.float32) def extend_pe(self, x: torch.Tensor) -> None: """Extend the time t in the positional encoding if required.