From 29fa878fff1d9ee6aa40572511e1a0c3eed12ae0 Mon Sep 17 00:00:00 2001 From: Fangjun Kuang Date: Mon, 6 Jun 2022 17:08:07 +0800 Subject: [PATCH] Fix Emformer for torchscript using torch 1.6.0 (#401) --- egs/librispeech/ASR/pruned_stateless_emformer_rnnt2/emformer.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/egs/librispeech/ASR/pruned_stateless_emformer_rnnt2/emformer.py b/egs/librispeech/ASR/pruned_stateless_emformer_rnnt2/emformer.py index 2ed7dab53..7d4702f11 100644 --- a/egs/librispeech/ASR/pruned_stateless_emformer_rnnt2/emformer.py +++ b/egs/librispeech/ASR/pruned_stateless_emformer_rnnt2/emformer.py @@ -296,7 +296,7 @@ class Emformer(EncoderInterface): Return the initial state of each layer. NOTE: the returned tensors are on the given device. `len(ans) == num_emformer_layers`. """ - if self._init_state: + if len(self._init_state) > 0: # Note(fangjun): It is OK to share the init state as it is # not going to be modified by the model return self._init_state