Commit Graph

  • 9bb0c7988f add random combiner for training deeper model yaozengwei 2022-07-18 11:50:35 +08:00
  • 8bd700cff2 fix scale value in scaling.py yaozengwei 2022-07-18 10:33:35 +08:00
  • bb1e1e154a Increasing param_pow to 0.6 and decreasing max_lr_factor from 4 to 3. Daniel Povey 2022-07-18 09:06:32 +08:00
  • 8db3b48edb Update parameter dependent part of cov more slowly, plus bug fix. Daniel Povey 2022-07-18 05:26:55 +08:00
  • 1d036a6a5a Support streaming decoding pkufool 2022-07-18 00:03:51 +08:00
  • 3cedbe3678 fix style yaozengwei 2022-07-17 21:40:29 +08:00
  • fd261eca3a Merge remote-tracking branch 'k2-fsa/master' into lstm_new yaozengwei 2022-07-17 21:24:56 +08:00
  • 1b0d2f3592 modify .flake8 yaozengwei 2022-07-17 21:22:00 +08:00
  • c71788ee57 remove add_model_arguments yaozengwei 2022-07-17 21:20:39 +08:00
  • 7c00f92abb fix style yaozengwei 2022-07-17 21:17:45 +08:00
  • 872d2390d2 link __init__.py yaozengwei 2022-07-17 20:39:46 +08:00
  • ce2d817114 pruned2 -> pruned4 yaozengwei 2022-07-17 20:36:20 +08:00
  • 125eac8dee flat weights after scaling yaozengwei 2022-07-17 20:35:29 +08:00
  • 539a9d75d4 simplified code in ScaledLSTM yaozengwei 2022-07-17 17:07:14 +08:00
  • 5c669b7716 modify streaming decoding files yaozengwei 2022-07-17 16:09:24 +08:00
  • 822cc78a9c copy streaming decoding files from pruned_transducer_stateless2 yaozengwei 2022-07-17 15:47:43 +08:00
  • 4a0dea2aa2 modify decode.py pretrained.py test_model.py train.py yaozengwei 2022-07-17 15:38:53 +08:00
  • b1be6ea475 copy from pruned_transducer_stateless2 yaozengwei 2022-07-17 15:37:27 +08:00
  • 89bfb6b9c7 link scaling.py pruned_transducer_stateless2 yaozengwei 2022-07-17 15:35:59 +08:00
  • 198cf2635c Reduce param_pow from 0.5 to 0.4. Daniel Povey 2022-07-17 15:35:07 +08:00
  • d16b9ec15f link from pruned_transducer_stateless2 yaozengwei 2022-07-17 15:32:54 +08:00
  • 3857a87b47 Merge branch 'merge_refactor_param_cov_norank1_iter_batch_max4.0_pow0.5_fix2r_lrupdate200_2k_ns' into merge2_refactor_max4.0_pow0.5_200_1k_ma3.0 Daniel Povey 2022-07-17 15:32:43 +08:00
  • 074bd7da71 hardcode bidirectional=False yaozengwei 2022-07-17 15:31:25 +08:00
  • 2d53f2ef8b add RNN and Conv2dSubsampling classes in lstm.py yaozengwei 2022-07-17 12:59:27 +08:00
  • a572eb4e33 Reducing final lr_update_period from 2k to 1k Daniel Povey 2022-07-17 12:56:02 +08:00
  • f36ebad618 Remove 2/3 StructuredLinear/StructuredConv1d modules, use linear/conv1d Daniel Povey 2022-07-17 04:24:48 +08:00
  • 7e88e2a0e9 Increase debug freq; add type to diagnostics and increase precision of mean,rms Daniel Povey 2022-07-17 04:41:53 +08:00
  • 7c9fcfa5c9 add RNNEncoderLayer and RNNEncoder classes in lstm.py yaozengwei 2022-07-16 22:50:42 +08:00
  • 9165de5f57 add ScaledLSTM yaozengwei 2022-07-16 22:47:05 +08:00
  • 1ebf714fb7 remove hlg related modifications Guo Liyong 2022-07-16 13:37:31 +08:00
  • 604daca1fd Use CTC loss as an auxiliary loss for transducer training. Fangjun Kuang 2022-07-16 08:52:44 +08:00
  • de1fd91435 Adding max_abs=3.0 to ActivationBalancer modules inside feedoforward modules. Daniel Povey 2022-07-16 07:19:26 +08:00
  • 23e6d2e6d8 Fix to the fix Daniel Povey 2022-07-16 06:53:44 +08:00
  • 473efcd531 add self loop to L Guo Liyong 2022-07-16 01:19:04 +08:00
  • 4c8d77d14a Fix return type Daniel Povey 2022-07-15 14:18:07 +08:00
  • 79d1316905 Add RNN-LM rescoring Erwan 2022-07-12 09:01:39 +02:00
  • ffca1ae7fb
    [WIP] Rnn-T LM nbest rescoring (#471) ezerhouni 2022-07-15 04:32:54 +02:00
  • 68c5935691 Fix bug re param_cov freshness, properly. Daniel Povey 2022-07-15 08:33:10 +08:00
  • b6ee698278 Make LR update period less frequent later in training; fix bug with param_cov freshness, was too fresh Daniel Povey 2022-07-15 07:59:30 +08:00
  • 0ce4c654f9 Add temperature in prefix Erwan 2022-07-14 15:16:49 +02:00
  • c17233eca7
    [Ready] [Recipes] add aishell2 (#465) Yuekai Zhang 2022-07-14 14:46:56 +08:00
  • 42eaaf7c86 change context-size 2 default Yuekai Zhang 2022-07-14 06:42:07 +00:00
  • 689441b237 Reduce param_pow from 0.75 to 0.5 Daniel Povey 2022-07-14 06:08:06 +08:00
  • 7f6fe02db9 Fix formula for smoothing (was applying more smoothing than intended, and in the opposite sense to intended), also revert max_rms from 2.0 to 4.0 Daniel Povey 2022-07-14 06:06:02 +08:00
  • 0a99ceb6ba psd algorithm Guo Liyong 2022-07-14 00:01:28 +08:00
  • dfff69b160 update result Yuekai Zhang 2022-07-13 14:10:42 +00:00
  • f8d28f0998
    update multi_quantization installation (#469) LIyong.Guo 2022-07-13 21:16:45 +08:00
  • d0c171d8a6
    Update egs/librispeech/ASR/pruned_transducer_stateless6/train.py LIyong.Guo 2022-07-13 21:03:06 +08:00
  • 3016440035 add more decoding methods Yuekai Zhang 2022-07-13 02:21:47 +00:00
  • 4785245e5c Reduce debug freq Daniel Povey 2022-07-13 06:51:23 +08:00
  • d48fe0b99c Change max rms from 10.0 to 4.0 Daniel Povey 2022-07-13 05:53:35 +08:00
  • bc2882ddcc
    Simplified memory bank for Emformer (#440) Zengwei Yao 2022-07-12 19:19:58 +08:00
  • 1a44724c26 fix error in RESULTS.md yaozengwei 2022-07-12 19:12:50 +08:00
  • 0fcdd15fec
    Merge branch 'k2-fsa:master' into master Zengwei Yao 2022-07-12 15:39:04 +08:00
  • cedfb5a377 Make max eig ratio 10 Daniel Povey 2022-07-12 13:59:58 +08:00
  • 2e619be9b0 fix decoding comment Yuekai Zhang 2022-07-12 02:40:51 +00:00
  • 13067dddb1 update hf link Yuekai Zhang 2022-07-12 02:13:36 +00:00
  • 278358bb9f Remove debug code Daniel Povey 2022-07-12 08:39:14 +08:00
  • 8c44ff26f7 Fix bug in batching code for scalars Daniel Povey 2022-07-12 08:36:45 +08:00
  • 25cb8308d5 Add max_block_size=512 to PrAdam Daniel Povey 2022-07-12 08:35:14 +08:00
  • 41df045773 Simplify formula, getting rid of scalar_exp_avg_sq Daniel Povey 2022-07-11 17:14:12 -07:00
  • 4f0e219523 Bug fix to reproduce past results with max_block_size unset. Daniel Povey 2022-07-11 17:03:32 -07:00
  • 075a2e27d8 Replace max_fullcov_size with max_block_size Daniel Povey 2022-07-11 16:37:01 -07:00
  • 3468c3aa5a Remove ActivationBalancer, unnecessary Daniel Povey 2022-07-11 14:12:24 -07:00
  • 7993c84cd6 Apparently working version, with changed test-code topology Daniel Povey 2022-07-11 13:17:29 -07:00
  • a2b54cca10 update result Yuekai Zhang 2022-07-11 14:35:39 +00:00
  • e5cadba145 Add temperature in doc Erwan 2022-07-11 15:13:46 +02:00
  • fd4deeab95 Add fast_beam_search_with_nbest_rescoring in decode Erwan 2022-07-07 15:20:01 +02:00
  • 2456307acb Add fast_beam_search_nbest Erwan 2022-07-07 15:11:20 +02:00
  • ce26495238
    Rand combine update result (#467) Zengwei Yao 2022-07-11 18:13:31 +08:00
  • 245d39b1bb Still debugging but close to done Daniel Povey 2022-07-11 00:33:37 -07:00
  • 27da50a1f6 Committing partial work.. Daniel Povey 2022-07-10 15:46:32 -07:00
  • d25df4af5e Slight refactoring, preparing for batching. Daniel Povey 2022-07-09 22:24:36 -07:00
  • d9a6180ae0 Bug fix Daniel Povey 2022-07-10 10:20:39 +08:00
  • b7035844a2 Introduce scalar_max, stop eps getting large or small Daniel Povey 2022-07-10 10:13:55 +08:00
  • 2f73434541 Reduce debug frequency Daniel Povey 2022-07-10 06:44:50 +08:00
  • b3bb2dac6f Iterative, more principled way of estimating param_cov Daniel Povey 2022-07-10 06:28:01 +08:00
  • d139c18f22 Max eig of Q limited to 5 times the mean Daniel Povey 2022-07-09 14:30:03 +08:00
  • ffeef4ede4 Remove rank-1 dims, meaning where size==numel(), from processing. Daniel Povey 2022-07-09 13:36:48 +08:00
  • 2fc9eb9789 Respect param_pow Daniel Povey 2022-07-09 12:49:04 +08:00
  • 209acaf6e4 Increase lr_update_period to 200. The update takes about 2 minutes, fore entire model. Daniel Povey 2022-07-09 11:36:54 +08:00
  • 61cab3ab65 introduce grad_cov_period Daniel Povey 2022-07-09 10:29:23 +08:00
  • 35a51bc153 Reduce debug probs Daniel Povey 2022-07-09 10:22:19 +08:00
  • 65bc964854 Fix bug for scalar update Daniel Povey 2022-07-09 10:14:20 +08:00
  • aa2237a793 Bug fix Daniel Povey 2022-07-09 10:11:54 +08:00
  • 50ee414486 Fix train.py for new optimizer Daniel Povey 2022-07-09 10:09:53 +08:00
  • 6810849058 Implement new version of learning method. Does more complete diagonalization of grads than the previous methods. Daniel Povey 2022-07-09 10:02:17 +08:00
  • a9edecd32c Conformed that symmetrizing helps because of interaction with regular update; still meta_lr_scale=0 best :-( Daniel Povey 2022-07-09 05:20:04 +08:00
  • ad68987423 minor fix yaozengwei 2022-07-08 17:28:45 +08:00
  • e3e8b1990c use i&i-1 to judge if it is a power of 2 yaozengwei 2022-07-08 17:16:10 +08:00
  • 2057124004 fix bug, >> -> << yaozengwei 2022-07-08 16:45:12 +08:00
  • 5cfdbd3699 replace torch.div() with << yaozengwei 2022-07-08 16:41:46 +08:00
  • 1a663f4759
    Merge 5ea58a446514f0cd1786c66a5058f7d735a53806 into 6c69c4e2533a1dd8d2d06583e734ed77c7e56752 Zengwei Yao 2022-07-08 15:49:32 +08:00
  • 6c69c4e253
    Support running icefall outside of a git tracked directory. (#470) Fangjun Kuang 2022-07-08 15:03:07 +08:00
  • 73205de3ae Minor fixes. Fangjun Kuang 2022-07-08 15:02:36 +08:00
  • 517febb006 Support running icefall outside of a git tracked directory. Fangjun Kuang 2022-07-08 14:58:31 +08:00
  • aed2093c83 update multi_quantization installation Guo Liyong 2022-07-08 12:20:25 +08:00
  • 52bfb2b018 This works better for reasons I dont understand. transpose is enough, same as symmetrizing. Daniel Povey 2022-07-08 11:53:59 +08:00
  • dc40220951 change context size to 1 Yuekai Zhang 2022-07-08 03:13:09 +00:00
  • e9ab1ddd39 Inconseqeuential config change Daniel Povey 2022-07-08 11:03:16 +08:00