mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-08-09 01:52:41 +00:00
Merge branch 'k2-fsa:master' into dev/k2ssl
This commit is contained in:
commit
ab44ac0f9e
2
.github/scripts/docker/Dockerfile
vendored
2
.github/scripts/docker/Dockerfile
vendored
@ -49,7 +49,7 @@ RUN pip install --no-cache-dir \
|
|||||||
kaldifst \
|
kaldifst \
|
||||||
kaldilm \
|
kaldilm \
|
||||||
librosa \
|
librosa \
|
||||||
matplotlib \
|
"matplotlib<=3.9.4" \
|
||||||
multi_quantization \
|
multi_quantization \
|
||||||
numba \
|
numba \
|
||||||
"numpy<2.0" \
|
"numpy<2.0" \
|
||||||
|
33
.github/scripts/ljspeech/TTS/run-matcha.sh
vendored
33
.github/scripts/ljspeech/TTS/run-matcha.sh
vendored
@ -77,7 +77,7 @@ function export_onnx() {
|
|||||||
popd
|
popd
|
||||||
|
|
||||||
pushd data/fbank
|
pushd data/fbank
|
||||||
rm -v *.json
|
rm -fv *.json
|
||||||
curl -SL -O https://huggingface.co/csukuangfj/icefall-tts-ljspeech-matcha-en-2024-10-28/resolve/main/data/cmvn.json
|
curl -SL -O https://huggingface.co/csukuangfj/icefall-tts-ljspeech-matcha-en-2024-10-28/resolve/main/data/cmvn.json
|
||||||
popd
|
popd
|
||||||
|
|
||||||
@ -115,6 +115,37 @@ function export_onnx() {
|
|||||||
|
|
||||||
ls -lh /icefall/*.wav
|
ls -lh /icefall/*.wav
|
||||||
soxi /icefall/generated-matcha-tts-steps-6-*.wav
|
soxi /icefall/generated-matcha-tts-steps-6-*.wav
|
||||||
|
|
||||||
|
cp ./model-steps-*.onnx /icefall
|
||||||
|
|
||||||
|
d=matcha-icefall-en_US-ljspeech
|
||||||
|
mkdir $d
|
||||||
|
cp -v data/tokens.txt $d
|
||||||
|
cp model-steps-3.onnx $d
|
||||||
|
pushd $d
|
||||||
|
curl -SL -O https://github.com/k2-fsa/sherpa-onnx/releases/download/tts-models/espeak-ng-data.tar.bz2
|
||||||
|
tar xf espeak-ng-data.tar.bz2
|
||||||
|
rm espeak-ng-data.tar.bz2
|
||||||
|
|
||||||
|
cat >README.md <<EOF
|
||||||
|
# Introduction
|
||||||
|
|
||||||
|
This model is trained using the dataset from
|
||||||
|
https://keithito.com/LJ-Speech-Dataset/
|
||||||
|
|
||||||
|
The dataset contains only 1 female speaker.
|
||||||
|
|
||||||
|
You can find the training code at
|
||||||
|
https://github.com/k2-fsa/icefall/tree/master/egs/ljspeech/TTS#matcha
|
||||||
|
EOF
|
||||||
|
|
||||||
|
ls -lh
|
||||||
|
|
||||||
|
popd
|
||||||
|
|
||||||
|
tar cvjf $d.tar.bz2 $d
|
||||||
|
mv $d.tar.bz2 /icefall
|
||||||
|
mv $d /icefall
|
||||||
}
|
}
|
||||||
|
|
||||||
prepare_data
|
prepare_data
|
||||||
|
74
.github/workflows/ljspeech.yml
vendored
74
.github/workflows/ljspeech.yml
vendored
@ -30,8 +30,8 @@ jobs:
|
|||||||
id: set-matrix
|
id: set-matrix
|
||||||
run: |
|
run: |
|
||||||
# outputting for debugging purposes
|
# outputting for debugging purposes
|
||||||
python ./.github/scripts/docker/generate_build_matrix.py
|
python ./.github/scripts/docker/generate_build_matrix.py --min-torch-version "2.3"
|
||||||
MATRIX=$(python ./.github/scripts/docker/generate_build_matrix.py)
|
MATRIX=$(python ./.github/scripts/docker/generate_build_matrix.py --min-torch-version "2.3")
|
||||||
echo "::set-output name=matrix::${MATRIX}"
|
echo "::set-output name=matrix::${MATRIX}"
|
||||||
|
|
||||||
ljspeech:
|
ljspeech:
|
||||||
@ -70,6 +70,10 @@ jobs:
|
|||||||
cd /icefall
|
cd /icefall
|
||||||
git config --global --add safe.directory /icefall
|
git config --global --add safe.directory /icefall
|
||||||
|
|
||||||
|
pip install "matplotlib<=3.9.4"
|
||||||
|
|
||||||
|
pip list
|
||||||
|
|
||||||
.github/scripts/ljspeech/TTS/run-matcha.sh
|
.github/scripts/ljspeech/TTS/run-matcha.sh
|
||||||
.github/scripts/ljspeech/TTS/run.sh
|
.github/scripts/ljspeech/TTS/run.sh
|
||||||
|
|
||||||
@ -94,3 +98,69 @@ jobs:
|
|||||||
repo_name: k2-fsa/sherpa-onnx
|
repo_name: k2-fsa/sherpa-onnx
|
||||||
repo_token: ${{ secrets.UPLOAD_GH_SHERPA_ONNX_TOKEN }}
|
repo_token: ${{ secrets.UPLOAD_GH_SHERPA_ONNX_TOKEN }}
|
||||||
tag: tts-models
|
tag: tts-models
|
||||||
|
|
||||||
|
- uses: actions/upload-artifact@v4
|
||||||
|
if: matrix.python-version == '3.9' && matrix.torch-version == '2.3.0'
|
||||||
|
with:
|
||||||
|
name: step-2
|
||||||
|
path: ./model-steps-2.onnx
|
||||||
|
|
||||||
|
- uses: actions/upload-artifact@v4
|
||||||
|
if: matrix.python-version == '3.9' && matrix.torch-version == '2.3.0'
|
||||||
|
with:
|
||||||
|
name: step-3
|
||||||
|
path: ./model-steps-3.onnx
|
||||||
|
|
||||||
|
- uses: actions/upload-artifact@v4
|
||||||
|
if: matrix.python-version == '3.9' && matrix.torch-version == '2.3.0'
|
||||||
|
with:
|
||||||
|
name: step-4
|
||||||
|
path: ./model-steps-4.onnx
|
||||||
|
|
||||||
|
- uses: actions/upload-artifact@v4
|
||||||
|
if: matrix.python-version == '3.9' && matrix.torch-version == '2.3.0'
|
||||||
|
with:
|
||||||
|
name: step-5
|
||||||
|
path: ./model-steps-5.onnx
|
||||||
|
|
||||||
|
- uses: actions/upload-artifact@v4
|
||||||
|
if: matrix.python-version == '3.9' && matrix.torch-version == '2.3.0'
|
||||||
|
with:
|
||||||
|
name: step-6
|
||||||
|
path: ./model-steps-6.onnx
|
||||||
|
|
||||||
|
- name: Upload models to huggingface
|
||||||
|
if: matrix.python-version == '3.9' && matrix.torch-version == '2.3.0'
|
||||||
|
shell: bash
|
||||||
|
env:
|
||||||
|
HF_TOKEN: ${{ secrets.HF_TOKEN }}
|
||||||
|
run: |
|
||||||
|
d=matcha-icefall-en_US-ljspeech
|
||||||
|
|
||||||
|
GIT_LFS_SKIP_SMUDGE=1 git clone https://huggingface.co/csukuangfj/$d hf
|
||||||
|
cp -av $d/* hf/
|
||||||
|
|
||||||
|
pushd hf
|
||||||
|
|
||||||
|
git lfs track "cmn_dict"
|
||||||
|
git lfs track "ru_dict"
|
||||||
|
|
||||||
|
git add .
|
||||||
|
|
||||||
|
git config --global user.name "csukuangfj"
|
||||||
|
git config --global user.email "csukuangfj@gmail.com"
|
||||||
|
git config --global lfs.allowincompletepush true
|
||||||
|
|
||||||
|
git commit -m "upload model" && git push https://csukuangfj:${HF_TOKEN}@huggingface.co/csukuangfj/$d main || true
|
||||||
|
popd
|
||||||
|
|
||||||
|
- name: Release exported onnx models
|
||||||
|
if: matrix.python-version == '3.9' && matrix.torch-version == '2.3.0'
|
||||||
|
uses: svenstaro/upload-release-action@v2
|
||||||
|
with:
|
||||||
|
file_glob: true
|
||||||
|
overwrite: true
|
||||||
|
file: matcha-icefall-*.tar.bz2
|
||||||
|
repo_name: k2-fsa/sherpa-onnx
|
||||||
|
repo_token: ${{ secrets.UPLOAD_GH_SHERPA_ONNX_TOKEN }}
|
||||||
|
tag: tts-models
|
||||||
|
@ -32,7 +32,7 @@ class Conformer(Transformer):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension
|
d_model (int): attention dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
num_decoder_layers (int): number of decoder layers
|
num_decoder_layers (int): number of decoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
|
@ -42,7 +42,7 @@ class Conformer(Transformer):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension, also the output dimension
|
d_model (int): attention dimension, also the output dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
num_decoder_layers (int): number of decoder layers
|
num_decoder_layers (int): number of decoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
|
@ -33,7 +33,7 @@ class Conformer(Transformer):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension
|
d_model (int): attention dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
num_decoder_layers (int): number of decoder layers
|
num_decoder_layers (int): number of decoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
|
@ -42,7 +42,7 @@ class Conformer(EncoderInterface):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension, also the output dimension
|
d_model (int): attention dimension, also the output dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
layer_dropout (float): layer-dropout rate.
|
layer_dropout (float): layer-dropout rate.
|
||||||
|
@ -42,7 +42,7 @@ class Conformer(EncoderInterface):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension, also the output dimension
|
d_model (int): attention dimension, also the output dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
layer_dropout (float): layer-dropout rate.
|
layer_dropout (float): layer-dropout rate.
|
||||||
|
@ -42,7 +42,7 @@ class Conformer(EncoderInterface):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension, also the output dimension
|
d_model (int): attention dimension, also the output dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
layer_dropout (float): layer-dropout rate.
|
layer_dropout (float): layer-dropout rate.
|
||||||
|
@ -42,7 +42,7 @@ class Conformer(EncoderInterface):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension, also the output dimension
|
d_model (int): attention dimension, also the output dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
layer_dropout (float): layer-dropout rate.
|
layer_dropout (float): layer-dropout rate.
|
||||||
|
@ -69,7 +69,7 @@ class Conformer(Transformer):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension
|
d_model (int): attention dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
num_decoder_layers (int): number of decoder layers
|
num_decoder_layers (int): number of decoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
|
@ -35,7 +35,7 @@ class Conformer(Transformer):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension
|
d_model (int): attention dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
cnn_module_kernel (int): Kernel size of convolution module
|
cnn_module_kernel (int): Kernel size of convolution module
|
||||||
|
@ -176,6 +176,15 @@ The above command generates the following files:
|
|||||||
|
|
||||||
where the 2 in `model-steps-2.onnx` means it uses 2 steps for the ODE solver.
|
where the 2 in `model-steps-2.onnx` means it uses 2 steps for the ODE solver.
|
||||||
|
|
||||||
|
**HINT**: If you get the following error while running `export_onnx.py`:
|
||||||
|
|
||||||
|
```
|
||||||
|
torch.onnx.errors.UnsupportedOperatorError: Exporting the operator
|
||||||
|
'aten::scaled_dot_product_attention' to ONNX opset version 14 is not supported.
|
||||||
|
```
|
||||||
|
|
||||||
|
please use `torch>=2.2.0`.
|
||||||
|
|
||||||
|
|
||||||
To export the Hifigan vocoder to onnx, please use:
|
To export the Hifigan vocoder to onnx, please use:
|
||||||
|
|
||||||
|
@ -176,12 +176,16 @@ def main():
|
|||||||
"language": "English",
|
"language": "English",
|
||||||
"voice": "en-us",
|
"voice": "en-us",
|
||||||
"has_espeak": 1,
|
"has_espeak": 1,
|
||||||
|
"jieba": 0,
|
||||||
"n_speakers": 1,
|
"n_speakers": 1,
|
||||||
"sample_rate": 22050,
|
"sample_rate": 22050,
|
||||||
"version": 1,
|
"version": 1,
|
||||||
|
"pad_id": tokenizer.pad_id,
|
||||||
"model_author": "icefall",
|
"model_author": "icefall",
|
||||||
"maintainer": "k2-fsa",
|
"maintainer": "k2-fsa",
|
||||||
|
"use_eos_bos": 1,
|
||||||
"dataset": "LJ Speech",
|
"dataset": "LJ Speech",
|
||||||
|
"dataset_url": "https://keithito.com/LJ-Speech-Dataset/",
|
||||||
"num_ode_steps": num_steps,
|
"num_ode_steps": num_steps,
|
||||||
}
|
}
|
||||||
add_meta_data(filename=filename, meta_data=meta_data)
|
add_meta_data(filename=filename, meta_data=meta_data)
|
||||||
|
Loading…
x
Reference in New Issue
Block a user