1336 Commits

Author SHA1 Message Date
Daniel Povey
8a9bbb93bc Cosmetic fixes 2022-07-24 04:45:57 +08:00
Daniel Povey
966ac36cde Fixes to comments 2022-07-24 04:36:41 +08:00
Daniel Povey
33ffd17515 Some cleanup 2022-07-24 04:22:11 +08:00
Daniel Povey
ddceb7963b Interpolate between iterative estimate of scale, and original value. 2022-07-23 15:27:48 +08:00
Daniel Povey
2c4bdd0ad0 Add _update_param_scales_simple(), add documentation 2022-07-23 14:49:58 +08:00
Daniel Povey
9730352257 Redce smoothing constant slightly 2022-07-23 13:12:31 +08:00
Daniel Povey
e1873fc0bb Tune phase2 again, from 0.005,5.0 to 0.01,40. Epoch 140 is 0.21/0.149 2022-07-23 10:10:01 +08:00
Daniel Povey
0fc58bac56 More tuning, epoch-140 results are 0.23,0.11 2022-07-23 09:52:51 +08:00
Daniel Povey
34a2d331bf Smooth in opposite orientation to G 2022-07-23 09:38:16 +08:00
Daniel Povey
a972655a70 Tuning. 2022-07-23 09:15:49 +08:00
Daniel Povey
b47433b77a Fix bug in smooth_cov, for power==1.0 2022-07-23 09:06:03 +08:00
Daniel Povey
cc388675a9 Bug fix RE rankj 2022-07-23 08:24:59 +08:00
Daniel Povey
dee496145d this version performs way worse but has bugs fixed, can optimize from here. 2022-07-23 08:11:20 +08:00
Daniel Povey
dd10eb140f First version after refactorization and changing the math, where optim.py runs 2022-07-23 06:32:56 +08:00
Quandwang
116d0cf26d
CTC attention model with reworked Conformer encoder and reworked Transformer decoder (#462)
* ctc attention model with reworked conformer encoder and reworked transformer decoder

* remove unnecessary func

* resolve flake8 conflicts

* fix typos and modify the expr of ScaledEmbedding

* use original beam size

* minor changes to the scripts

* add rnn lm decoding

* minor changes

* check whether q k v weight is None

* check whether q k v weight is None

* check whether q k v weight is None

* style correction

* update results

* update results

* upload the decoding results of rnn-lm to the RESULTS

* upload the decoding results of rnn-lm to the RESULTS

* Update egs/librispeech/ASR/RESULTS.md

Co-authored-by: Fangjun Kuang <csukuangfj@gmail.com>

* Update egs/librispeech/ASR/RESULTS.md

Co-authored-by: Fangjun Kuang <csukuangfj@gmail.com>

* Update egs/librispeech/ASR/RESULTS.md

Co-authored-by: Fangjun Kuang <csukuangfj@gmail.com>

Co-authored-by: Fangjun Kuang <csukuangfj@gmail.com>
2022-07-22 15:31:25 +08:00
Daniel Povey
4da4e69fba Draft of new way of smoothing param_rms, diagonalized by grad 2022-07-22 06:37:20 +08:00
Mingshuang Luo
3d2986b4c2
Update conformer.py for aishell4 (#484)
* update conformer.py for aishell4

* update conformer.py

* add strict=False when model.load_state_dict
2022-07-20 21:32:53 +08:00
Daniel Povey
a8696b36fc
Merge pull request #483 from yaozengwei/fix_diagnostic
Fix diagnostic
2022-07-18 23:33:45 -07:00
yaozengwei
a35b28cd8d fix for case of None stats 2022-07-19 14:29:23 +08:00
Daniel Povey
a63afe348a Increase max_lr_factor from 3.0 to 4.0 2022-07-19 06:56:41 +08:00
Daniel Povey
79a2f09f62 Change how formula for max_lr_factor works, and increase factor from 2.5 to 3. 2022-07-19 06:54:49 +08:00
Daniel Povey
525c097130 Increase power from 0.7 to 0.75 2022-07-19 05:44:03 +08:00
Daniel Povey
2dff1161b4 Reduce max_lr_factor from 3.0 to 2.5 2022-07-19 05:15:03 +08:00
Daniel Povey
8bb44b2944 Change param_pow from 0.6 to 0.7 2022-07-19 05:08:32 +08:00
ezerhouni
608473b4eb
Add RNN-LM rescoring in fast beam search (#475) 2022-07-18 16:52:17 +08:00
Mingshuang Luo
aec222e2fe
add compile_lg.py for aishell2 recipe (#481) 2022-07-18 14:36:40 +08:00
Daniel Povey
bb1e1e154a Increasing param_pow to 0.6 and decreasing max_lr_factor from 4 to 3. 2022-07-18 09:06:32 +08:00
Daniel Povey
8db3b48edb Update parameter dependent part of cov more slowly, plus bug fix. 2022-07-18 05:26:55 +08:00
Daniel Povey
198cf2635c Reduce param_pow from 0.5 to 0.4. 2022-07-17 15:35:07 +08:00
Daniel Povey
3857a87b47 Merge branch 'merge_refactor_param_cov_norank1_iter_batch_max4.0_pow0.5_fix2r_lrupdate200_2k_ns' into merge2_refactor_max4.0_pow0.5_200_1k_ma3.0 2022-07-17 15:32:43 +08:00
Daniel Povey
a572eb4e33 Reducing final lr_update_period from 2k to 1k 2022-07-17 12:56:02 +08:00
Daniel Povey
f36ebad618 Remove 2/3 StructuredLinear/StructuredConv1d modules, use linear/conv1d 2022-07-17 06:40:19 +08:00
Daniel Povey
7e88e2a0e9 Increase debug freq; add type to diagnostics and increase precision of mean,rms 2022-07-17 06:40:16 +08:00
Daniel Povey
de1fd91435 Adding max_abs=3.0 to ActivationBalancer modules inside feedoforward modules. 2022-07-16 07:19:26 +08:00
Daniel Povey
23e6d2e6d8 Fix to the fix 2022-07-16 06:53:44 +08:00
Daniel Povey
4c8d77d14a Fix return type 2022-07-15 14:18:07 +08:00
ezerhouni
ffca1ae7fb
[WIP] Rnn-T LM nbest rescoring (#471) 2022-07-15 10:32:54 +08:00
Daniel Povey
68c5935691 Fix bug re param_cov freshness, properly. 2022-07-15 08:33:10 +08:00
Daniel Povey
b6ee698278 Make LR update period less frequent later in training; fix bug with param_cov freshness, was too fresh 2022-07-15 07:59:30 +08:00
Yuekai Zhang
c17233eca7
[Ready] [Recipes] add aishell2 (#465)
* add aishell2

* fix aishell2

* add manifest stats

* update prepare char dict

* fix lint

* setting max duration

* lint

* change context size to 1

* update result

* update hf link

* fix decoding comment

* add more decoding methods

* update result

* change context-size 2 default
2022-07-14 14:46:56 +08:00
Daniel Povey
689441b237 Reduce param_pow from 0.75 to 0.5 2022-07-14 06:08:06 +08:00
Daniel Povey
7f6fe02db9 Fix formula for smoothing (was applying more smoothing than intended, and in the opposite sense to intended), also revert max_rms from 2.0 to 4.0 2022-07-14 06:06:02 +08:00
LIyong.Guo
f8d28f0998
update multi_quantization installation (#469)
* update multi_quantization installation

* Update egs/librispeech/ASR/pruned_transducer_stateless6/train.py

Co-authored-by: Fangjun Kuang <csukuangfj@gmail.com>

Co-authored-by: Fangjun Kuang <csukuangfj@gmail.com>
2022-07-13 21:16:45 +08:00
Daniel Povey
4785245e5c Reduce debug freq 2022-07-13 06:51:23 +08:00
Daniel Povey
d48fe0b99c Change max rms from 10.0 to 4.0 2022-07-13 05:53:35 +08:00
Zengwei Yao
bc2882ddcc
Simplified memory bank for Emformer (#440)
* init files

* use average value as memory vector for each chunk

* change tail padding length from right_context_length to chunk_length

* correct the files, ln -> cp

* fix bug in conv_emformer_transducer_stateless2/emformer.py

* fix doc in conv_emformer_transducer_stateless/emformer.py

* refactor init states for stream

* modify .flake8

* fix bug about memory mask when memory_size==0

* add @torch.jit.export for init_states function

* update RESULTS.md

* minor change

* update README.md

* modify doc

* replace torch.div() with <<

* fix bug, >> -> <<

* use i&i-1 to judge if it is a power of 2

* minor fix

* fix error in RESULTS.md
2022-07-12 19:19:58 +08:00
Daniel Povey
cedfb5a377 Make max eig ratio 10 2022-07-12 13:59:58 +08:00
Daniel Povey
278358bb9f Remove debug code 2022-07-12 08:39:14 +08:00
Daniel Povey
8c44ff26f7 Fix bug in batching code for scalars 2022-07-12 08:36:45 +08:00
Daniel Povey
25cb8308d5 Add max_block_size=512 to PrAdam 2022-07-12 08:35:14 +08:00