marcoyang1998
c74cec59e9
propagate changes from #525 to other librispeech recipes ( #531 )
...
* propaga changes from #525 to other librispeech recipes
* refactor display_and_save_batch to utils
* fixed typo
* reformat code style
2022-08-17 17:18:15 +08:00
Fangjun Kuang
1f7832b93c
Fix loading sampler state dict. ( #421 )
...
* Fix loading sampler state dict.
* skip scan_pessimistic_batches_for_oom if params.start_batch > 0
2022-08-06 10:00:08 +08:00
Zengwei Yao
8203d10be7
Add stats about duration and padding proportion ( #485 )
...
* add stats about duration and padding proportion
* add for utt_duration
* add stats for other recipes
* add stats for other 2 recipes
* modify doc
* minor change
2022-07-25 16:40:43 +08:00
Wei Kang
6e609c67a2
Using streaming conformer as transducer encoder ( #380 )
...
* support streaming in conformer
* Add more documents
* support streaming on pruned_transducer_stateless2; add delay penalty; fixes for decode states
* Minor fixes
* streaming for pruned_transducer_stateless4
* Fix conv cache error, support async streaming decoding
* Fix style
* Fix style
* Fix style
* Add torch.jit.export
* mask the initial cache
* Cutting off invalid frames of encoder_embed output
* fix relative positional encoding in streaming decoding for compution saving
* Minor fixes
* Minor fixes
* Minor fixes
* Minor fixes
* Minor fixes
* Fix jit export for torch 1.6
* Minor fixes for streaming decoding
* Minor fixes on decode stream
* move model parameters to train.py
* make states in forward streaming optional
* update pretrain to support streaming model
* update results.md
* update tensorboard and pre-models
* fix typo
* Fix tests
* remove unused arguments
* add streaming decoding ci
* Minor fix
* Minor fix
* disable right context by default
2022-06-28 00:18:54 +08:00
Zengwei Yao
a42d96dfe0
Fix warmup ( #435 )
...
* fix warmup when scan_pessimistic_batches_for_oom
* delete comments
2022-06-20 13:40:01 +08:00
Quandwang
8512aaf585
fix typos ( #409 )
2022-06-08 20:08:44 +08:00
Daniel Povey
4e23fb2252
Improve diagnostics code memory-wise and accumulate more stats. ( #373 )
...
* Update diagnostics, hopefully print more stats.
# Conflicts:
# egs/librispeech/ASR/pruned_transducer_stateless4b/train.py
* Remove memory-limit options arg
* Remove unnecessary option for diagnostics code, collect on more batches
2022-05-19 11:45:59 +08:00
Fangjun Kuang
32f05c00e3
Save batch to disk on exception. ( #350 )
2022-05-06 17:49:40 +08:00
Fangjun Kuang
e1c3e98980
Save batch to disk on OOM. ( #343 )
...
* Save batch to disk on OOM.
* minor fixes
* Fixes after review.
* Fix style issues.
2022-05-05 15:09:23 +08:00
pehonnet
9a98e6ced6
fix fp16 option in example usage ( #332 )
2022-04-25 18:51:53 +08:00
Mingshuang Luo
93c60a9d30
Code style check for librispeech pruned transducer stateless2 ( #308 )
2022-04-11 22:15:18 +08:00
Wei Kang
7012fd65b5
Support mix precision training on the reworked model ( #305 )
...
* Add mix precision support
* Minor fixes
* Minor fixes
* Minor fixes
2022-04-11 16:49:54 +08:00
Daniel Povey
5078332088
Fix adding learning rate to tensorboard
2022-04-11 14:58:15 +08:00
Daniel Povey
46d52dda10
Fix dir names
2022-04-11 12:03:41 +08:00
Daniel Povey
962cf868c9
Fix import
2022-04-10 15:31:46 +08:00
Daniel Povey
d1e4ae788d
Refactor how learning rate is set.
2022-04-10 15:25:27 +08:00
Daniel Povey
82d58629ea
Implement 2p version of learning rate schedule.
2022-04-10 13:50:31 +08:00
Daniel Povey
da50525ca5
Make lrate rule more symmetric
2022-04-10 13:25:40 +08:00
Daniel Povey
4d41ee0caa
Implement 2o schedule
2022-04-09 18:37:03 +08:00
Daniel Povey
db72aee1f0
Set 2n rule..
2022-04-09 18:15:56 +08:00
Daniel Povey
0f8ee68af2
Fix bug
2022-04-08 16:53:42 +08:00
Daniel Povey
f587cd527d
Change exponential part of lrate to be epoch based
2022-04-08 16:24:21 +08:00
Daniel Povey
6ee32cf7af
Set new scheduler
2022-04-08 16:10:06 +08:00
Daniel Povey
a41e93437c
Change some defaults in LR-setting rule.
2022-04-06 12:36:58 +08:00
Daniel Povey
d1a669162c
Fix bug in lambda
2022-04-05 13:31:52 +08:00
Daniel Povey
ed8eba91e1
Reduce model_warm_step from 4k to 3k
2022-04-05 13:24:09 +08:00
Daniel Povey
c3169222ae
Simplified optimizer, rework somet things..
2022-04-05 13:23:02 +08:00
Daniel Povey
0f5957394b
Fix to reading scheudler from optim
2022-04-05 12:58:43 +08:00
Daniel Povey
1548cc7462
Fix checkpoint-writing
2022-04-05 11:19:40 +08:00
Daniel Povey
234366e51c
Fix type of parameter
2022-04-05 00:18:36 +08:00
Daniel Povey
d1f2f93460
Some fixes..
2022-04-04 22:40:18 +08:00
Daniel Povey
72f4a673b1
First draft of new approach to learning rates + init
2022-04-04 20:21:34 +08:00
Daniel Povey
4929e4cf32
Change how warm-step is set
2022-04-04 17:09:25 +08:00
Daniel Povey
34500afc43
Various bug fixes
2022-04-02 20:06:43 +08:00
Daniel Povey
8be10d3d6c
First draft of model rework
2022-04-02 20:03:21 +08:00
Daniel Povey
eec597fdd5
Merge changes from master
2022-04-02 18:45:20 +08:00
Daniel Povey
709c387ce6
Initial refactoring to remove unnecessary vocab_size
2022-03-30 21:40:22 +08:00
Daniel Povey
4e453a4bf9
Rework conformer, remove some code.
2022-03-29 23:41:13 +08:00
Daniel Povey
11124b03ea
Refactoring and simplifying conformer and frontend
2022-03-29 20:32:14 +08:00
Daniel Povey
262388134d
Increase model_warm_step to 4k
2022-03-27 11:18:16 +08:00
Daniel Povey
d2ed3dfc90
Fix bug
2022-03-25 20:35:11 +08:00
Daniel Povey
4b650e9f01
Make warmup work by scaling layer contributions; leave residual layer-drop
2022-03-25 20:34:33 +08:00
Daniel Povey
aab72bc2a5
Add changes from master to decode.py, train.py
2022-03-24 13:10:54 +08:00
Daniel Povey
9a8aa1f54a
Change how warmup works.
2022-03-22 15:36:20 +08:00
Daniel Povey
4004ca81d8
Increase warm_step (and valid_interval)
2022-03-22 13:32:24 +08:00
Daniel Povey
b82a505dfc
Reduce initial pruned_loss scale from 0.01 to 0.0
2022-03-22 12:30:48 +08:00
Daniel Povey
ccbf8ba086
Incorporate changes from master into pruned_transducer_stateless2.
2022-03-21 21:12:43 +08:00
Daniel Povey
0ee2404ff0
Remove logging code that broke with newer Lhotse; fix bug with pruned_loss
2022-03-19 14:01:45 +08:00
Daniel Povey
2dfcd8f117
Double warm_step
2022-03-18 16:38:36 +08:00
Daniel Povey
cbe6b175d1
Reduce warmup scale on pruned loss form 0.1 to 0.01.
2022-03-17 16:46:59 +08:00