From ee845f219cb94869b1fba3a049fae1f582f0e48c Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Sat, 25 Feb 2023 16:10:40 +0900 Subject: [PATCH] from local --- .../ASR/conformer_ctc2/.decode.py.swp | Bin 65536 -> 65536 bytes egs/librispeech/ASR/conformer_ctc2/decode.py | 2 +- 2 files changed, 1 insertion(+), 1 deletion(-) diff --git a/egs/librispeech/ASR/conformer_ctc2/.decode.py.swp b/egs/librispeech/ASR/conformer_ctc2/.decode.py.swp index a9159bd882d8dd760edaa85b38e22934f1d07490..b7ad43a159fe12c751ea9cdfae3293c69a1229c7 100644 GIT binary patch delta 146 zcmZo@U}yuCRt4h6EDQ`sf%pgzuiTtj cXvIAF!+%yrrOEgHNN?U(GKFb#MEQyi0Kh#WQvd(} diff --git a/egs/librispeech/ASR/conformer_ctc2/decode.py b/egs/librispeech/ASR/conformer_ctc2/decode.py index d96f1bcf9..55989d59c 100755 --- a/egs/librispeech/ASR/conformer_ctc2/decode.py +++ b/egs/librispeech/ASR/conformer_ctc2/decode.py @@ -417,7 +417,7 @@ def decode_one_batch( if params.method == "greedy-search": unsorted_token_ids = graph_compiler.texts_to_ids(supervisions["text"]) - att_loss = model.decoder_forward( + pred, att_loss = model.decoder_forward( memory, memory_key_padding_mask, token_ids=unsorted_token_ids,