non stable training in some scenarios. The clamping range is set to (-10,2).
Note that this change may cause unexpected effect if you resume
training from a model that is trained without clamping.
* Fix not enough values to unpack error .
* [WIP] Pruned transducer stateless2 for AISHELL-1
* fix the style issue
* code format for black
* add pruned-transducer-stateless2 results for AISHELL-1
* simplify result
* add ScaledLSTM
* add RNNEncoderLayer and RNNEncoder classes in lstm.py
* add RNN and Conv2dSubsampling classes in lstm.py
* hardcode bidirectional=False
* link from pruned_transducer_stateless2
* link scaling.py pruned_transducer_stateless2
* copy from pruned_transducer_stateless2
* modify decode.py pretrained.py test_model.py train.py
* copy streaming decoding files from pruned_transducer_stateless2
* modify streaming decoding files
* simplified code in ScaledLSTM
* flat weights after scaling
* pruned2 -> pruned4
* link __init__.py
* fix style
* remove add_model_arguments
* modify .flake8
* fix style
* fix scale value in scaling.py
* add random combiner for training deeper model
* add using proj_size
* add scaling converter for ScaledLSTM
* support jit trace
* add using averaged model in export.py
* modify test_model.py, test if the model can be successfully exported by jit.trace
* modify pretrained.py
* support streaming decoding
* fix model.py
* Add cut_id to recognition results
* Add cut_id to recognition results
* do not pad in Conv subsampling module; add tail padding during decoding.
* update RESULTS.md
* minor fix
* fix doc
* update README.md
* minor change, filter infinite loss
* remove the condition of raise error
* modify type hint for the return value in model.py
* minor change
* modify RESULTS.md
Co-authored-by: pkufool <wkang.pku@gmail.com>
* Add function display_and_save_batch in egs/wenetspeech/ASR/pruned_transducer_stateless2/train.py
* Modify function: display_and_save_batch
* Delete empty line in pruned_transducer_stateless2/train.py
* Modify code format
* Sort result to make it more convenient to compare decoding results
* Add cut_id to recognition results
* add cut_id to results for all recipes
* Fix torch.jit.script
* Fix comments
* Minor fixes
* Fix torch.jit.tracing for Pytorch version before v1.9.0
* add multiprocessing for wenetspeech text segmentation
* Fix preparing char based lang for wenetspeech
* fix style
Co-authored-by: WeijiZhuang <zhuangweiji@xiaomi.com>
* WIP: Support exporting to ONNX format
* Minor fixes.
* Combine encoder/decoder/joiner into a single file.
* Revert merging three onnx models into a single one.
It's quite time consuming to extract a sub-graph from the combined
model. For instance, it takes more than one hour to extract
the encoder model.
* Update CI to test ONNX models.
* Decode with exported models.
* Fix typos.
* Add more doc.
* Remove ncnn as it is not fully tested yet.
* Fix as_strided for streaming conformer.
* add other decoding methods for wenetspeech
* changes for RESULTS.md
* add ngram-lm-scale=0.35 results
* set ngram-lm-scale=0.35 as default
* Update README.md
* add nbest-scale for flie name
* pruned-rnnt5-for-wenetspeech
* style check
* style check
* add streaming conformer
* add streaming decode
* changes codes for fast_beam_search and export cpu jit
* add modified-beam-search for streaming decoding
* add modified-beam-search for streaming decoding
* change for streaming_beam_search.py
* add README.md and RESULTS.md
* change for style_check.yml
* do some changes
* do some changes for export.py
* add some decode commands for usage
* add streaming results on README.md