From 37dc3941163aacf85639778e569dff0e889e7ae9 Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Thu, 2 Feb 2023 13:58:45 +0900 Subject: [PATCH] from local --- .../ASR/conformer_ctc/.transformer.py.swp | Bin 69632 -> 69632 bytes egs/aishell/ASR/conformer_ctc/transformer.py | 4 +--- 2 files changed, 1 insertion(+), 3 deletions(-) diff --git a/egs/aishell/ASR/conformer_ctc/.transformer.py.swp b/egs/aishell/ASR/conformer_ctc/.transformer.py.swp index 08932d237afdd392433d8d43e70e54a469b06b8e..0ea38276fdebedb4251212972b3c0f4bcc15dfe5 100644 GIT binary patch delta 125 zcmZozz|ydQMKsAE%+puFQqO<^2m}}y4m#aVp0H8$H$S7~Wx_1_mJ>h$K)* z;j=%h1dz96GiO0MqsbB`28K`|W(8svApXS&lmg<#K>QDg-vjYFAYKT>V4$GDuz6~s XBqN(|eoARh>g0#(L^fY9(wqwb#wj3x diff --git a/egs/aishell/ASR/conformer_ctc/transformer.py b/egs/aishell/ASR/conformer_ctc/transformer.py index b82cc486f..d70abb6d2 100644 --- a/egs/aishell/ASR/conformer_ctc/transformer.py +++ b/egs/aishell/ASR/conformer_ctc/transformer.py @@ -23,8 +23,6 @@ import torch.nn as nn from label_smoothing import LabelSmoothingLoss from subsampling import Conv2dSubsampling, VggSubsampling from torch.nn.utils.rnn import pad_sequence -from torch.nn.modules import Module -from torch import Tensor # Note: TorchScript requires Dict/List/etc. to be fully typed. Supervisions = Dict[str, torch.Tensor] @@ -382,7 +380,7 @@ class Transformer(nn.Module): return nll -class TransformerEncoder(Module): +class TransformerEncoder(nn.TransformerEncoder): r"""TransformerEncoder is a stack of N encoder layers. Users can build the BERT(https://arxiv.org/abs/1810.04805) model with corresponding parameters.