diff --git a/egs/librispeech/ASR/conformer_ctc/pretrained.py b/egs/librispeech/ASR/conformer_ctc/pretrained.py index 25fd1745f..95029fadb 100755 --- a/egs/librispeech/ASR/conformer_ctc/pretrained.py +++ b/egs/librispeech/ASR/conformer_ctc/pretrained.py @@ -261,7 +261,6 @@ def main(): if params.method in ["whole-lattice-rescoring", "attention-decoder"]: logging.info(f"Loading G from {params.G}") G = k2.Fsa.from_dict(torch.load(params.G, map_location="cpu")) - G = G.to(device) # Add epsilon self-loops to G as we will compose # it with the whole lattice later G = G.to(device)