* Add function display_and_save_batch in egs/wenetspeech/ASR/pruned_transducer_stateless2/train.py
* Modify function: display_and_save_batch
* Delete empty line in pruned_transducer_stateless2/train.py
* Modify code format
* Sort result to make it more convenient to compare decoding results
* Add cut_id to recognition results
* add cut_id to results for all recipes
* Fix torch.jit.script
* Fix comments
* Minor fixes
* Fix torch.jit.tracing for Pytorch version before v1.9.0
* add multiprocessing for wenetspeech text segmentation
* Fix preparing char based lang for wenetspeech
* fix style
Co-authored-by: WeijiZhuang <zhuangweiji@xiaomi.com>
* WIP: Support exporting to ONNX format
* Minor fixes.
* Combine encoder/decoder/joiner into a single file.
* Revert merging three onnx models into a single one.
It's quite time consuming to extract a sub-graph from the combined
model. For instance, it takes more than one hour to extract
the encoder model.
* Update CI to test ONNX models.
* Decode with exported models.
* Fix typos.
* Add more doc.
* Remove ncnn as it is not fully tested yet.
* Fix as_strided for streaming conformer.
* add other decoding methods for wenetspeech
* changes for RESULTS.md
* add ngram-lm-scale=0.35 results
* set ngram-lm-scale=0.35 as default
* Update README.md
* add nbest-scale for flie name
* pruned-rnnt5-for-wenetspeech
* style check
* style check
* add streaming conformer
* add streaming decode
* changes codes for fast_beam_search and export cpu jit
* add modified-beam-search for streaming decoding
* add modified-beam-search for streaming decoding
* change for streaming_beam_search.py
* add README.md and RESULTS.md
* change for style_check.yml
* do some changes
* do some changes for export.py
* add some decode commands for usage
* add streaming results on README.md
* add stats about duration and padding proportion
* add for utt_duration
* add stats for other recipes
* add stats for other 2 recipes
* modify doc
* minor change
* ctc attention model with reworked conformer encoder and reworked transformer decoder
* remove unnecessary func
* resolve flake8 conflicts
* fix typos and modify the expr of ScaledEmbedding
* use original beam size
* minor changes to the scripts
* add rnn lm decoding
* minor changes
* check whether q k v weight is None
* check whether q k v weight is None
* check whether q k v weight is None
* style correction
* update results
* update results
* upload the decoding results of rnn-lm to the RESULTS
* upload the decoding results of rnn-lm to the RESULTS
* Update egs/librispeech/ASR/RESULTS.md
Co-authored-by: Fangjun Kuang <csukuangfj@gmail.com>
* Update egs/librispeech/ASR/RESULTS.md
Co-authored-by: Fangjun Kuang <csukuangfj@gmail.com>
* Update egs/librispeech/ASR/RESULTS.md
Co-authored-by: Fangjun Kuang <csukuangfj@gmail.com>
Co-authored-by: Fangjun Kuang <csukuangfj@gmail.com>
* init files
* use average value as memory vector for each chunk
* change tail padding length from right_context_length to chunk_length
* correct the files, ln -> cp
* fix bug in conv_emformer_transducer_stateless2/emformer.py
* fix doc in conv_emformer_transducer_stateless/emformer.py
* refactor init states for stream
* modify .flake8
* fix bug about memory mask when memory_size==0
* add @torch.jit.export for init_states function
* update RESULTS.md
* minor change
* update README.md
* modify doc
* replace torch.div() with <<
* fix bug, >> -> <<
* use i&i-1 to judge if it is a power of 2
* minor fix
* fix error in RESULTS.md
* add pruned transducer stateless5 recipe for tal_csasr
* do some changes for merging
* change for conformer.py
* add wer and cer for Chinese and English respectively
* fix a error for conformer.py