From 0f3696ed914f6a1688ba48e4ec3b1eb7a27815af Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Thu, 2 Feb 2023 14:06:20 +0900 Subject: [PATCH] from local --- .../ASR/conformer_ctc/.transformer.py.swp | Bin 69632 -> 69632 bytes egs/aishell/ASR/conformer_ctc/transformer.py | 1 - 2 files changed, 1 deletion(-) diff --git a/egs/aishell/ASR/conformer_ctc/.transformer.py.swp b/egs/aishell/ASR/conformer_ctc/.transformer.py.swp index 894e87db3a8b29aeab983556b9973d5667606dad..b8c6618042c628cdb0e2e6d9f9a827e053431500 100644 GIT binary patch delta 107 zcmZozz|ydQMJ&l6%+puFQqO<^2m}}yj=9`U4rkdYwo8CfeDg&C~p)2aq^Va!kgdT5jF+@ga8~w delta 136 zcmZozz|ydQMJ&l6%+puFQqO<^2m}}y>|AaqKV;r0wo8CfV)I1-qWGM|%G4sw$^EM(Hov_iYzzSH9wo8> diff --git a/egs/aishell/ASR/conformer_ctc/transformer.py b/egs/aishell/ASR/conformer_ctc/transformer.py index 16eb54599..f8d30e403 100644 --- a/egs/aishell/ASR/conformer_ctc/transformer.py +++ b/egs/aishell/ASR/conformer_ctc/transformer.py @@ -401,7 +401,6 @@ class TransfEncoder(nn.TransformerEncoder): __constants__ = ['norm'] def __init__(self, encoder_layer, num_layers, norm=None, enable_nested_tensor=True, mask_check=True): - print(encoder_layer) super(TransfEncoder, self).__init__( encoder_layer=encoder_layer, num_layers=num_layers,