mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-08-26 18:24:18 +00:00
misc. update
This commit is contained in:
parent
b9d34fb9d9
commit
05e48ca880
@ -14,7 +14,7 @@ The best CER, for CommonVoice 16.1 (cv-corpus-16.1-2023-12-06/zh-HK) is below:
|
|||||||
| modified_beam_search | 0.98 | 1.11 | --epoch 24 --avg 5 |
|
| modified_beam_search | 0.98 | 1.11 | --epoch 24 --avg 5 |
|
||||||
| fast_beam_search | 1.08 | 1.27 | --epoch 24 --avg 5 |
|
| fast_beam_search | 1.08 | 1.27 | --epoch 24 --avg 5 |
|
||||||
|
|
||||||
When doing the cross-corpus validation on MDCC (w/o blank penalty),
|
When doing the cross-corpus validation on [MDCC](https://arxiv.org/abs/2201.02419) (w/o blank penalty),
|
||||||
the best CER is below:
|
the best CER is below:
|
||||||
|
|
||||||
| | Dev | Test | Note |
|
| | Dev | Test | Note |
|
||||||
@ -23,7 +23,7 @@ the best CER is below:
|
|||||||
| modified_beam_search | 39.73 | 39.19| --epoch 24 --avg 5 |
|
| modified_beam_search | 39.73 | 39.19| --epoch 24 --avg 5 |
|
||||||
| fast_beam_search | 42.14 | 41.98| --epoch 24 --avg 5 |
|
| fast_beam_search | 42.14 | 41.98| --epoch 24 --avg 5 |
|
||||||
|
|
||||||
When doing the cross-corpus validation on MDCC (with blank penalty set to 2.2),
|
When doing the cross-corpus validation on [MDCC](https://arxiv.org/abs/2201.02419) (with blank penalty set to 2.2),
|
||||||
the best CER is below:
|
the best CER is below:
|
||||||
|
|
||||||
| | Dev | Test | Note |
|
| | Dev | Test | Note |
|
||||||
@ -68,14 +68,16 @@ Detailed experimental results and pre-trained model are available at:
|
|||||||
<https://huggingface.co/zrjin/icefall-asr-commonvoice-zh-HK-zipformer-2024-03-20>
|
<https://huggingface.co/zrjin/icefall-asr-commonvoice-zh-HK-zipformer-2024-03-20>
|
||||||
|
|
||||||
|
|
||||||
### GigaSpeech BPE training results (Pruned Stateless Transducer 7)
|
### CommonVoice English (en) BPE training results (Pruned Stateless Transducer 7)
|
||||||
|
|
||||||
#### [pruned_transducer_stateless7](./pruned_transducer_stateless7)
|
#### [pruned_transducer_stateless7](./pruned_transducer_stateless7)
|
||||||
|
|
||||||
See #997 for more details.
|
See #997 for more details.
|
||||||
|
|
||||||
Number of model parameters: 70369391, i.e., 70.37 M
|
Number of model parameters: 70369391, i.e., 70.37 M
|
||||||
|
|
||||||
|
Note that the result is obtained using GigaSpeech transcript trained BPE model
|
||||||
|
|
||||||
The best WER, as of 2023-04-17, for Common Voice English 13.0 (cv-corpus-13.0-2023-03-09/en) is below:
|
The best WER, as of 2023-04-17, for Common Voice English 13.0 (cv-corpus-13.0-2023-03-09/en) is below:
|
||||||
|
|
||||||
Results are:
|
Results are:
|
||||||
|
@ -47,6 +47,7 @@ def get_args():
|
|||||||
"--subset",
|
"--subset",
|
||||||
type=str,
|
type=str,
|
||||||
default="train",
|
default="train",
|
||||||
|
choices=["train", "validated", "invalidated"],
|
||||||
help="""Dataset parts to compute fbank. """,
|
help="""Dataset parts to compute fbank. """,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -130,11 +130,11 @@ if __name__ == "__main__":
|
|||||||
norm_lines = [normalize_text(line, lang) for line in lines]
|
norm_lines = [normalize_text(line, lang) for line in lines]
|
||||||
|
|
||||||
text_words_segments = get_word_segments(norm_lines)
|
text_words_segments = get_word_segments(norm_lines)
|
||||||
with open(output_dir / "transcript_words.txt", "w+", encoding="utf-8") as f:
|
with open(output_dir / "transcript_words.txt", "w", encoding="utf-8") as f:
|
||||||
f.writelines(text_words_segments)
|
f.writelines(text_words_segments)
|
||||||
|
|
||||||
words = get_words(text_words_segments)[1:] # remove "\n" from words
|
words = get_words(text_words_segments)[1:] # remove "\n" from words
|
||||||
with open(output_dir / "words_no_ids.txt", "w+", encoding="utf-8") as f:
|
with open(output_dir / "words_no_ids.txt", "w", encoding="utf-8") as f:
|
||||||
f.writelines([word + "\n" for word in sorted(words)])
|
f.writelines([word + "\n" for word in sorted(words)])
|
||||||
|
|
||||||
words = (
|
words = (
|
||||||
@ -143,5 +143,5 @@ if __name__ == "__main__":
|
|||||||
+ ["#0", "<s>", "<\s>"]
|
+ ["#0", "<s>", "<\s>"]
|
||||||
)
|
)
|
||||||
|
|
||||||
with open(output_dir / "words.txt", "w+", encoding="utf-8") as f:
|
with open(output_dir / "words.txt", "w", encoding="utf-8") as f:
|
||||||
f.writelines([f"{word} {i}\n" for i, word in enumerate(words)])
|
f.writelines([f"{word} {i}\n" for i, word in enumerate(words)])
|
||||||
|
Loading…
x
Reference in New Issue
Block a user