From 8d2797d7cd6b6a91f8ceae305e03f2655f2e5d90 Mon Sep 17 00:00:00 2001 From: Fangjun Kuang Date: Fri, 29 Apr 2022 14:13:44 +0800 Subject: [PATCH] Update results. --- README.md | 17 +++++++++++++++++ egs/librispeech/ASR/RESULTS.md | 23 +++++++++++++++++++++++ 2 files changed, 40 insertions(+) diff --git a/README.md b/README.md index 6adba4955..188ca013b 100644 --- a/README.md +++ b/README.md @@ -35,6 +35,9 @@ We do provide a Colab notebook for this recipe. ### LibriSpeech +Please see +for the **latest** results. + We provide 4 models for this recipe: - [conformer CTC model][LibriSpeech_conformer_ctc] @@ -92,6 +95,20 @@ in the decoding. We provide a Colab notebook to run a pre-trained transducer conformer + stateless decoder model: [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1CO1bXJ-2khDckZIW8zjOPHGSKLHpTDlp?usp=sharing) + +#### k2 pruned RNN-T + +| | test-clean | test-other | +|-----|------------|------------| +| WER | 2.57 | 5.95 | + +#### k2 pruned RNN-T + GigaSpeech + +| | test-clean | test-other | +|-----|------------|------------| +| WER | 2.19 | 4.97 | + + ### Aishell We provide two models for this recipe: [conformer CTC model][Aishell_conformer_ctc] diff --git a/egs/librispeech/ASR/RESULTS.md b/egs/librispeech/ASR/RESULTS.md index fc63d385d..871013892 100644 --- a/egs/librispeech/ASR/RESULTS.md +++ b/egs/librispeech/ASR/RESULTS.md @@ -127,6 +127,29 @@ The Nbest oracle WER is computed using the following steps: - 5. The path with the lowest edit distance is the final output and is used to compute the WER +The command to compute the Nbest oracle WER is: + +```bash +for epoch in 27; do + for avg in 10 ; do + for num_paths in 50 100 200 400; do + for nbest_scale in 0.5 0.8 1.0; do + ./pruned_transducer_stateless3/decode.py \ + --epoch $epoch \ + --avg $avg \ + --exp-dir ./pruned_transducer_stateless3/exp \ + --max-duration 600 \ + --decoding-method fast_beam_search_nbest_oracle \ + --num-paths $num_paths \ + --max-states 32 \ + --beam 8 \ + --nbest-scale $nbest_scale + done + done + done +done +``` + ### LibriSpeech BPE training results (Pruned Transducer 2) [pruned_transducer_stateless2](./pruned_transducer_stateless2)