47 Commits

Author SHA1 Message Date
Daniel Povey
425e274c82 Replace norm in ConvolutionModule with a scaling factor. 2022-03-10 16:01:53 +08:00
Daniel Povey
87b843f023 Change exp dir 2022-03-10 14:44:55 +08:00
Daniel Povey
b55472bb42 Replace most normalizations with scales (still have norm in conv) 2022-03-10 14:43:54 +08:00
Daniel Povey
feb20ca84d Merge changes to diagnostics 2022-03-10 10:31:42 +08:00
Daniel Povey
1e5455ba29 Update diagnostics 2022-03-10 10:28:48 +08:00
Daniel Povey
d074cf73c6 Extensions to diagnostics code 2022-03-09 20:37:20 +08:00
Daniel Povey
e2ace9d545 Replace norm on input layer with scale of 0.1. 2022-03-07 11:24:04 +08:00
Daniel Povey
a37d98463a Restore ConvolutionModule to state before changes; change all Swish,Swish(Swish) to SwishOffset. 2022-03-06 11:55:02 +08:00
Daniel Povey
8a8b81cd18 Replace relu with swish-squared. 2022-03-05 22:21:42 +08:00
Daniel Povey
5f2c0a09b7 Convert swish nonlinearities to ReLU 2022-03-05 16:28:24 +08:00
Daniel Povey
0cd14ae739 Fix exp dir 2022-03-05 12:17:09 +08:00
Daniel Povey
65b09dd5f2 Double the threshold in brelu; slightly increase max_factor. 2022-03-05 00:07:14 +08:00
Daniel Povey
74f2b163de Merge diagnostics improvement 2022-03-04 23:15:47 +08:00
Daniel Povey
6252282fd0 Add deriv-balancing code 2022-03-04 20:19:11 +08:00
Daniel Povey
eb3ed54202 Reduce scale from 50 to 20 2022-03-04 15:56:45 +08:00
Daniel Povey
9cc5999829 Fix duplicate Swish; replace norm+swish with swish+exp-scale in convolution module 2022-03-04 15:50:51 +08:00
Daniel Povey
7e88999641 Increase scale from 20 to 50. 2022-03-04 14:31:29 +08:00
Daniel Povey
3207bd98a9 Increase scale on Scale from 4 to 20 2022-03-04 13:16:40 +08:00
Daniel Povey
503f8d521c Fix bug in diagnostics 2022-03-04 13:08:56 +08:00
Daniel Povey
cd216f50b6 Add import 2022-03-04 11:03:01 +08:00
Daniel Povey
bc6c720e25 Combine ExpScale and swish for memory reduction 2022-03-04 10:52:05 +08:00
Daniel Povey
23b3aa233c Double learning rate of exp-scale units 2022-03-04 00:42:37 +08:00
Daniel Povey
5c177fc52b pelu_base->expscale, add 2xExpScale in subsampling, and in feedforward units. 2022-03-03 23:52:03 +08:00
Daniel Povey
3fb559d2f0 Add baseline for the PeLU expt, keeping only the small normalization-related changes. 2022-03-02 18:27:08 +08:00
Daniel Povey
9d1b4ae046 Add pelu to this good-performing setup.. 2022-03-02 16:33:27 +08:00
Daniel Povey
2ff520c800 Improvements to diagnostics (RE those with 1 dim 2022-02-28 12:22:27 +08:00
Daniel Povey
c1063def95 First version of rand-combine iterated-training-like idea. 2022-02-27 17:34:58 +08:00
Daniel Povey
63d8d935d4 Refactor/simplify ConformerEncoder 2022-02-27 13:56:15 +08:00
Daniel Povey
581786a6d3 Adding diagnostics code... 2022-02-27 13:44:43 +08:00
Daniel Povey
2af1b3af98 Remove ReLU in attention 2022-02-14 19:39:19 +08:00
Daniel Povey
d187ad8b73 Change max_frames from 0.2 to 0.15 2022-02-11 16:24:17 +08:00
Daniel Povey
4cd2c02fff Fix num_time_masks code; revert 0.8 to 0.9 2022-02-10 15:53:11 +08:00
Daniel Povey
c170c53006 Change p=0.9 to p=0.8 in SpecAug 2022-02-10 14:59:14 +08:00
Daniel Povey
8aa50df4f0 Change p=0.5->0.9, mask_fraction 0.3->0.2 2022-02-09 22:52:53 +08:00
Daniel Povey
beaf5bfbab Merge specaug change from Mingshuang. 2022-02-08 19:42:23 +08:00
Daniel Povey
a859dcb205 Remove learnable offset, use relu instead. 2022-02-07 12:14:48 +08:00
Daniel Povey
48a764eccf Add min in q,k,v of attention 2022-02-06 21:19:37 +08:00
pkufool
fcd25bdfff Fix torch.nn.Embedding error for torch below 1.8.0 2022-02-06 18:22:56 +08:00
Wei Kang
5ae80dfca7
Minor fixes (#193) 2022-01-27 18:01:17 +08:00
Fangjun Kuang
d6050eb02e Fix calling optimized_transducer after new release. (#182) 2022-01-21 08:18:50 +08:00
Fangjun Kuang
f94ff19bfe
Refactor beam search and update results. (#177) 2022-01-18 16:40:19 +08:00
Fangjun Kuang
4c1b3665ee
Use optimized_transducer to compute transducer loss. (#162)
* WIP: Use optimized_transducer to compute transducer loss.

* Minor fixes.

* Fix decoding.

* Fix decoding.

* Add RESULTS.

* Update RESULTS.

* Update CI.

* Fix sampling rate for yesno recipe.
2022-01-10 11:54:58 +08:00
Fangjun Kuang
14c93add50
Remove batchnorm, weight decay, and SOS from transducer conformer encoder (#155)
* Remove batchnorm, weight decay, and SOS.

* Make --context-size configurable.

* Update results.
2021-12-27 16:01:10 +08:00
Fangjun Kuang
8187d6236c
Minor fix to maximum number of symbols per frame for RNN-T decoding. (#157)
* Minor fix to maximum number of symbols per frame RNN-T decoding.

* Minor fixes.
2021-12-24 21:48:40 +08:00
Fangjun Kuang
fb6a57e9e0
Increase the size of the context in the RNN-T decoder. (#153) 2021-12-23 07:55:02 +08:00
Fangjun Kuang
cb04c8a750
Limit the number of symbols per frame in RNN-T decoding. (#151) 2021-12-18 11:00:42 +08:00
Fangjun Kuang
1d44da845b
RNN-T Conformer training for LibriSpeech (#143)
* Begin to add RNN-T training for librispeech.

* Copy files from conformer_ctc.

Will edit it.

* Use conformer/transformer model as encoder.

* Begin to add training script.

* Add training code.

* Remove long utterances to avoid OOM when a large max_duraiton is used.

* Begin to add decoding script.

* Add decoding script.

* Minor fixes.

* Add beam search.

* Use LSTM layers for the encoder.

Need more tunings.

* Use stateless decoder.

* Minor fixes to make it ready for merge.

* Fix README.

* Update RESULT.md to include RNN-T Conformer.

* Minor fixes.

* Fix tests.

* Minor fixes.

* Minor fixes.

* Fix tests.
2021-12-18 07:42:51 +08:00