From 6e08de169aabea76f8e7278ae243960a0c273ce7 Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Mon, 9 Jan 2023 19:35:49 +0900 Subject: [PATCH] from local --- .../ASR/incremental_transf/.conformer.py.swp | Bin 110592 -> 110592 bytes .../ASR/incremental_transf/conformer.py | 4 ---- 2 files changed, 4 deletions(-) diff --git a/egs/librispeech/ASR/incremental_transf/.conformer.py.swp b/egs/librispeech/ASR/incremental_transf/.conformer.py.swp index c326555b95ddd6d763743cfa96aa833434478e5a..6cac712ab497bc149237c97a600815d2894b2251 100644 GIT binary patch delta 247 zcmX}mzY76z7zXh7_2WqCzMMfx-DHy#ixPwXftf{OHW*yVz=SJr76zN?%5am(W-!^A zj5-6ehr#fvXLx$w$GU0OO>;cOjnPxNPsPSCI%XpF6OkfnxUm`QJQR(n4JkMWiIy;h zHdH`^AY1}OTj)U-YLEj1PPn;)4rJi_=n9)m+q8ekt|Yle>?VU2o~`QMjD^08J$~kpKVy delta 204 zcmWm7p$TtZHg-}HsK32MZqPI zR7^0ESNJ{k@vgRawbu_N&B0Tdr;=npHV2RnF$!p!_*fk2&upMdUkf= 1.0 we are "fully warmed up". It is used - to turn modules on sequentially. Returns: Return a tuple containing 2 tensors: - embeddings: its shape is (batch_size, output_seq_len, d_model)