mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-08-09 10:02:22 +00:00
Remove onnx and onnxruntime from requirements.txt (#640)
* Remove onnx and onnxruntime from requirements.txt
This commit is contained in:
parent
1abf2863bb
commit
7f1c0e07b6
@ -42,6 +42,11 @@ import argparse
|
|||||||
import logging
|
import logging
|
||||||
from typing import List, Optional, Tuple
|
from typing import List, Optional, Tuple
|
||||||
|
|
||||||
|
from icefall import is_module_available
|
||||||
|
|
||||||
|
if not is_module_available("onnxruntime"):
|
||||||
|
raise ValueError("Please 'pip install onnxruntime' first.")
|
||||||
|
|
||||||
import onnxruntime as ort
|
import onnxruntime as ort
|
||||||
import sentencepiece as spm
|
import sentencepiece as spm
|
||||||
import torch
|
import torch
|
||||||
|
@ -24,6 +24,11 @@ with the given torchscript model for the same input.
|
|||||||
import argparse
|
import argparse
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
|
from icefall import is_module_available
|
||||||
|
|
||||||
|
if not is_module_available("onnxruntime"):
|
||||||
|
raise ValueError("Please 'pip install onnxruntime' first.")
|
||||||
|
|
||||||
import onnxruntime as ort
|
import onnxruntime as ort
|
||||||
import torch
|
import torch
|
||||||
|
|
||||||
|
@ -21,6 +21,11 @@ This file is to test that models can be exported to onnx.
|
|||||||
"""
|
"""
|
||||||
import os
|
import os
|
||||||
|
|
||||||
|
from icefall import is_module_available
|
||||||
|
|
||||||
|
if not is_module_available("onnxruntime"):
|
||||||
|
raise ValueError("Please 'pip install onnxruntime' first.")
|
||||||
|
|
||||||
import onnxruntime as ort
|
import onnxruntime as ort
|
||||||
import torch
|
import torch
|
||||||
from conformer import (
|
from conformer import (
|
||||||
|
@ -21,7 +21,6 @@ import k2
|
|||||||
import torch
|
import torch
|
||||||
import torch.nn as nn
|
import torch.nn as nn
|
||||||
from encoder_interface import EncoderInterface
|
from encoder_interface import EncoderInterface
|
||||||
from multi_quantization.prediction import JointCodebookLoss
|
|
||||||
from scaling import ScaledLinear
|
from scaling import ScaledLinear
|
||||||
|
|
||||||
from icefall.utils import add_sos
|
from icefall.utils import add_sos
|
||||||
@ -74,6 +73,14 @@ class Transducer(nn.Module):
|
|||||||
encoder_dim, vocab_size, initial_speed=0.5
|
encoder_dim, vocab_size, initial_speed=0.5
|
||||||
)
|
)
|
||||||
self.simple_lm_proj = ScaledLinear(decoder_dim, vocab_size)
|
self.simple_lm_proj = ScaledLinear(decoder_dim, vocab_size)
|
||||||
|
|
||||||
|
from icefall import is_module_available
|
||||||
|
|
||||||
|
if not is_module_available("multi_quantization"):
|
||||||
|
raise ValueError("Please 'pip install multi_quantization' first.")
|
||||||
|
|
||||||
|
from multi_quantization.prediction import JointCodebookLoss
|
||||||
|
|
||||||
if num_codebooks > 0:
|
if num_codebooks > 0:
|
||||||
self.codebook_loss_net = JointCodebookLoss(
|
self.codebook_loss_net = JointCodebookLoss(
|
||||||
predictor_channels=encoder_dim,
|
predictor_channels=encoder_dim,
|
||||||
|
@ -28,18 +28,21 @@ from typing import List, Tuple
|
|||||||
import numpy as np
|
import numpy as np
|
||||||
import torch
|
import torch
|
||||||
import torch.multiprocessing as mp
|
import torch.multiprocessing as mp
|
||||||
import multi_quantization as quantization
|
|
||||||
|
|
||||||
|
from icefall import is_module_available
|
||||||
|
|
||||||
|
if not is_module_available("multi_quantization"):
|
||||||
|
raise ValueError("Please 'pip install multi_quantization' first.")
|
||||||
|
|
||||||
|
import multi_quantization as quantization
|
||||||
from asr_datamodule import LibriSpeechAsrDataModule
|
from asr_datamodule import LibriSpeechAsrDataModule
|
||||||
from hubert_xlarge import HubertXlargeFineTuned
|
from hubert_xlarge import HubertXlargeFineTuned
|
||||||
from icefall.utils import (
|
|
||||||
AttributeDict,
|
|
||||||
setup_logger,
|
|
||||||
)
|
|
||||||
from lhotse import CutSet, load_manifest
|
from lhotse import CutSet, load_manifest
|
||||||
from lhotse.cut import MonoCut
|
from lhotse.cut import MonoCut
|
||||||
from lhotse.features.io import NumpyHdf5Writer
|
from lhotse.features.io import NumpyHdf5Writer
|
||||||
|
|
||||||
|
from icefall.utils import AttributeDict, setup_logger
|
||||||
|
|
||||||
|
|
||||||
class CodebookIndexExtractor:
|
class CodebookIndexExtractor:
|
||||||
"""
|
"""
|
||||||
|
@ -40,6 +40,11 @@ https://huggingface.co/luomingshuang/icefall_asr_wenetspeech_pruned_transducer_s
|
|||||||
import argparse
|
import argparse
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
|
from icefall import is_module_available
|
||||||
|
|
||||||
|
if not is_module_available("onnxruntime"):
|
||||||
|
raise ValueError("Please 'pip install onnxruntime' first.")
|
||||||
|
|
||||||
import onnxruntime as ort
|
import onnxruntime as ort
|
||||||
import torch
|
import torch
|
||||||
|
|
||||||
|
@ -49,6 +49,12 @@ from typing import List
|
|||||||
import k2
|
import k2
|
||||||
import kaldifeat
|
import kaldifeat
|
||||||
import numpy as np
|
import numpy as np
|
||||||
|
|
||||||
|
from icefall import is_module_available
|
||||||
|
|
||||||
|
if not is_module_available("onnxruntime"):
|
||||||
|
raise ValueError("Please 'pip install onnxruntime' first.")
|
||||||
|
|
||||||
import onnxruntime as ort
|
import onnxruntime as ort
|
||||||
import torch
|
import torch
|
||||||
import torchaudio
|
import torchaudio
|
||||||
|
@ -50,6 +50,7 @@ from .utils import (
|
|||||||
get_executor,
|
get_executor,
|
||||||
get_texts,
|
get_texts,
|
||||||
is_jit_tracing,
|
is_jit_tracing,
|
||||||
|
is_module_available,
|
||||||
l1_norm,
|
l1_norm,
|
||||||
l2_norm,
|
l2_norm,
|
||||||
linf_norm,
|
linf_norm,
|
||||||
|
@ -17,7 +17,7 @@
|
|||||||
from collections import defaultdict
|
from collections import defaultdict
|
||||||
from typing import List, Optional, Tuple
|
from typing import List, Optional, Tuple
|
||||||
|
|
||||||
import kaldifst
|
from icefall.utils import is_module_available
|
||||||
|
|
||||||
|
|
||||||
class NgramLm:
|
class NgramLm:
|
||||||
@ -36,6 +36,11 @@ class NgramLm:
|
|||||||
is_binary:
|
is_binary:
|
||||||
True if the given file is a binary FST.
|
True if the given file is a binary FST.
|
||||||
"""
|
"""
|
||||||
|
if not is_module_available("kaldifst"):
|
||||||
|
raise ValueError("Please 'pip install kaldifst' first.")
|
||||||
|
|
||||||
|
import kaldifst
|
||||||
|
|
||||||
if is_binary:
|
if is_binary:
|
||||||
lm = kaldifst.StdVectorFst.read(fst_filename)
|
lm = kaldifst.StdVectorFst.read(fst_filename)
|
||||||
else:
|
else:
|
||||||
@ -85,6 +90,8 @@ class NgramLm:
|
|||||||
self, state: int, label: int
|
self, state: int, label: int
|
||||||
) -> Tuple[int, float]:
|
) -> Tuple[int, float]:
|
||||||
"""TODO: Add doc."""
|
"""TODO: Add doc."""
|
||||||
|
import kaldifst
|
||||||
|
|
||||||
arc_iter = kaldifst.ArcIterator(self.lm, state)
|
arc_iter = kaldifst.ArcIterator(self.lm, state)
|
||||||
num_arcs = self.lm.num_arcs(state)
|
num_arcs = self.lm.num_arcs(state)
|
||||||
|
|
||||||
|
@ -976,3 +976,17 @@ def display_and_save_batch(
|
|||||||
y = sp.encode(supervisions["text"], out_type=int)
|
y = sp.encode(supervisions["text"], out_type=int)
|
||||||
num_tokens = sum(len(i) for i in y)
|
num_tokens = sum(len(i) for i in y)
|
||||||
logging.info(f"num tokens: {num_tokens}")
|
logging.info(f"num tokens: {num_tokens}")
|
||||||
|
|
||||||
|
|
||||||
|
# `is_module_available` is copied from
|
||||||
|
# https://github.com/pytorch/audio/blob/6bad3a66a7a1c7cc05755e9ee5931b7391d2b94c/torchaudio/_internal/module_utils.py#L9
|
||||||
|
def is_module_available(*modules: str) -> bool:
|
||||||
|
r"""Returns if a top-level module with :attr:`name` exists *without**
|
||||||
|
importing it. This is generally safer than try-catch block around a
|
||||||
|
`import X`.
|
||||||
|
|
||||||
|
Note: "borrowed" from torchaudio:
|
||||||
|
"""
|
||||||
|
import importlib
|
||||||
|
|
||||||
|
return all(importlib.util.find_spec(m) is not None for m in modules)
|
||||||
|
@ -3,9 +3,4 @@ kaldialign
|
|||||||
sentencepiece>=0.1.96
|
sentencepiece>=0.1.96
|
||||||
tensorboard
|
tensorboard
|
||||||
typeguard
|
typeguard
|
||||||
multi_quantization
|
|
||||||
onnx
|
|
||||||
onnxruntime
|
|
||||||
--extra-index-url https://pypi.ngc.nvidia.com
|
|
||||||
dill
|
dill
|
||||||
kaldifst
|
|
||||||
|
@ -16,6 +16,12 @@
|
|||||||
# limitations under the License.
|
# limitations under the License.
|
||||||
|
|
||||||
import graphviz
|
import graphviz
|
||||||
|
|
||||||
|
from icefall import is_module_available
|
||||||
|
|
||||||
|
if not is_module_available("kaldifst"):
|
||||||
|
raise ValueError("Please 'pip install kaldifst' first.")
|
||||||
|
|
||||||
import kaldifst
|
import kaldifst
|
||||||
|
|
||||||
from icefall import NgramLm, NgramLmStateCost
|
from icefall import NgramLm, NgramLmStateCost
|
||||||
|
Loading…
x
Reference in New Issue
Block a user