Commit Graph

  • 072776b2a1 Apply layer-skip dropout prob Daniel Povey 2022-10-29 20:11:39 +08:00
  • 9a7979d7b8 Avoid falling off the loop for weird inputs Daniel Povey 2022-10-29 20:03:41 +08:00
  • 05689f6354 Add skip connections as in normal U-net Daniel Povey 2022-10-29 19:47:10 +08:00
  • bba454a349 Make decoder group size equal to 4. Daniel Povey 2022-10-29 17:08:19 +08:00
  • ff03ec88a5 Tuning change to num encoder layers, inspired by relative param importance. Daniel Povey 2022-10-29 15:56:02 +08:00
  • f995426324 Reduce min of bypass_scale from 0.5 to 0.3, and make it not applied in test mode. Daniel Povey 2022-10-29 15:40:47 +08:00
  • 435d0dec71 Reduce dim of linear positional encoding in attention layers. Daniel Povey 2022-10-29 15:31:34 +08:00
  • 96ea4cf1be Have 6 different encoder stacks, U-shaped network. Daniel Povey 2022-10-28 20:36:45 +08:00
  • 7b57a34227 Have 4 encoder stacks Daniel Povey 2022-10-28 19:32:14 +08:00
  • de9a6ebd6c Fix bug RE seq lengths Daniel Povey 2022-10-28 19:26:06 +08:00
  • d7d5188bd9 Refactor how the downsampling is done so that it happens later, but the 1st encoder stack still operates after a subsampling of 2. Daniel Povey 2022-10-28 19:20:21 +08:00
  • 0a89f51dc9 Have a 3rd encoder, at downsampling factor of 8. Daniel Povey 2022-10-28 17:39:29 +08:00
  • ed1b4d5e5d Refactor zipformer for more flexibility so we can change number of encoder layers. Daniel Povey 2022-10-28 17:32:38 +08:00
  • e592a920b4 Merge branch 'scaled_adam_exp198b' into scaled_adam_exp202 Daniel Povey 2022-10-28 13:13:55 +08:00
  • a067fe8026 Fix clamping of epsilon Daniel Povey 2022-10-28 12:50:14 +08:00
  • 7b8a0108ea Merge branch 'scaled_adam_exp188' into scaled_adam_exp198b Daniel Povey 2022-10-28 12:49:36 +08:00
  • b9f6ba1aa2 Remove some unused variables. Daniel Povey 2022-10-28 12:01:45 +08:00
  • c8abba75a9 Update decode.py by copying from pruned_transducer_stateless5 and changing directory name Daniel Povey 2022-10-28 11:19:45 +08:00
  • 6709bf1e63
    Update train.py (#635) Nagendra Goel 2022-10-27 22:23:32 -04:00
  • e08a44dd7f
    Merge pull request #1 from ngoel17/pruned-transducer3-train-bugfix Nagendra Goel 2022-10-27 22:10:13 -04:00
  • 1ae0ee3104
    Update train.py Nagendra Goel 2022-10-27 22:08:02 -04:00
  • 5dfa141ca5 Rename Conformer to Zipformer Daniel Povey 2022-10-27 22:43:46 +08:00
  • 3f05e47447 Rename conformer.py to zipformer.py Daniel Povey 2022-10-27 22:41:48 +08:00
  • be5c687fbd Merging upstream/master Daniel Povey 2022-10-27 21:04:48 +08:00
  • f8c531cd23 Increase bypass_scale min from 0.4 to 0.5 Daniel Povey 2022-10-27 14:59:05 +08:00
  • 2c400115e4 Increase bypass_scale from 0.2 to 0.4. Daniel Povey 2022-10-27 14:30:46 +08:00
  • a7fc6ae38c Increase floor on bypass_scale from 0.1 to 0.2. Daniel Povey 2022-10-27 14:09:34 +08:00
  • 938510ac9f Fix clamping of bypass scale; remove a couple unused variables. Daniel Povey 2022-10-27 14:05:53 +08:00
  • bf37c7ca85 Regularize how we apply the min and max to the eps of BasicNorm Daniel Povey 2022-10-26 12:51:20 +08:00
  • a0507a83a5 Change scalar_max in optim.py from 2.0 to 5.0 Daniel Povey 2022-10-25 22:45:08 +08:00
  • 78f3cba58c Add logging about memory used. Daniel Povey 2022-10-25 19:19:33 +08:00
  • 6a6df19bde Hopefully make penalize_abs_values_gt more memory efficient. Daniel Povey 2022-10-25 18:34:04 +08:00
  • dbfbd8016b Cast to float16 in DoubleSwish forward Daniel Povey 2022-10-25 13:16:00 +08:00
  • 3159b09e8f Make 20 the limit for warmup_count Daniel Povey 2022-10-25 12:58:27 +08:00
  • 6ebff23cb9 Reduce cutoff from 100 to 5 for estimating OOM with warmup Daniel Povey 2022-10-25 12:53:12 +08:00
  • 9da5526659 Changes to more accurately estimate OOM conditions Daniel Povey 2022-10-25 12:49:18 +08:00
  • 1e8984174b Change to warmup schedule. Daniel Povey 2022-10-25 12:27:00 +08:00
  • 36cb279318 More memory efficient backprop for DoubleSwish. Daniel Povey 2022-10-25 12:21:22 +08:00
  • 499ac24ecb
    Install kaldifst for GitHub actions (#632) Fangjun Kuang 2022-10-24 15:07:29 +08:00
  • bc0f3f260e Fix CI errors Fangjun Kuang 2022-10-24 11:30:53 +08:00
  • eee48dd7ac Install kaldifst for GitHub actions Fangjun Kuang 2022-10-24 10:44:11 +08:00
  • 95aaa4a8d2 Store only half precision output for softmax. Daniel Povey 2022-10-23 21:24:46 +08:00
  • d3876e32c4 Make it use float16 if in amp but use clamp to avoid wrapping error Daniel Povey 2022-10-23 21:13:23 +08:00
  • 85657946bb Try a more exact way to round to uint8 that should prevent ever wrapping around to zero Daniel Povey 2022-10-23 20:56:26 +08:00
  • d6aa386552 Fix randn to rand Daniel Povey 2022-10-23 17:19:19 +08:00
  • e586cc319c Change the discretization of the sigmoid to be expectation preserving. Daniel Povey 2022-10-23 17:11:35 +08:00
  • 09cbc9fdab Save some memory in the autograd of DoubleSwish. Daniel Povey 2022-10-23 16:59:43 +08:00
  • 40588d3d8a Revert 179->180 change, i.e. change max_abs for deriv_balancer2 back from 50.0 20.0 Daniel Povey 2022-10-23 16:18:58 +08:00
  • c7ac266f9a
    Merge ebcc8e4e1de7f8e97ce1587264166dc865459371 into 348494888d08d5ddba2baadddcfe7df576d4bed1 kobenaxie 2022-10-23 15:55:35 +09:00
  • 5b9d166cb9 --base-lr0.075->0.5; --lr-epochs 3->3.5 Daniel Povey 2022-10-23 13:56:25 +08:00
  • 0406d0b059 Increase max_abs in ActivationBalancer of conv module from 20 to 50 Daniel Povey 2022-10-23 13:51:51 +08:00
  • 9e86d1f44f reduce initial scale in GradScaler Daniel Povey 2022-10-23 00:14:38 +08:00
  • b7083e7aff Increase default max_factor for ActivationBalancer from 0.02 to 0.04; decrease max_abs in ConvolutionModule.deriv_balancer2 from 100.0 to 20.0 Daniel Povey 2022-10-23 00:09:21 +08:00
  • ad2d3c2b36 Dont print out full non-finite tensor Daniel Povey 2022-10-22 23:03:19 +08:00
  • e0c1dc66da Increase probs of activation balancer and make it decay slower. Daniel Povey 2022-10-22 22:18:38 +08:00
  • 2964628ae1 don't do penalize_values_gt on simple_lm_proj and simple_am_proj; reduce --base-lr from 0.1 to 0.075 Daniel Povey 2022-10-22 21:12:58 +08:00
  • 269b70122e Add hooks.py, had negleted to git add it. Daniel Povey 2022-10-22 20:58:52 +08:00
  • 13ffd8e823 Trying to reduce grad_scale of Whiten() from 0.02 to 0.01. Daniel Povey 2022-10-22 20:30:05 +08:00
  • 466176eeff Use penalize_abs_values_gt, not ActivationBalancer. Daniel Povey 2022-10-22 20:15:21 +08:00
  • 7a55cac346 Increase max_factor in final lm_balancer and am_balancer Daniel Povey 2022-10-22 20:02:54 +08:00
  • 8b3bba9b54 Reduce max_abs on am_balancer Daniel Povey 2022-10-22 19:52:11 +08:00
  • 1908123af9 Adding activation balancers after simple_am_prob and simple_lm_prob Daniel Povey 2022-10-22 19:37:35 +08:00
  • 11886dc4f6 Change base lr to 0.1, also rename from initial lr in train.py Daniel Povey 2022-10-22 18:22:26 +08:00
  • 146626bb85 Renaming in optim.py; remove step() from scan_pessimistic_batches_for_oom in train.py Daniel Povey 2022-10-22 17:44:21 +08:00
  • 525e87a82d Add inf check hooks Daniel Povey 2022-10-22 17:16:29 +08:00
  • e8066b5825 Merge branch 'scaled_adam_exp172' into scaled_adam_exp174 Daniel Povey 2022-10-22 15:44:04 +08:00
  • 9919fb3e1b Increase grad_scale to Whiten module Daniel Povey 2022-10-22 15:32:50 +08:00
  • af0fc31c78 Introduce warmup schedule in optimizer Daniel Povey 2022-10-22 15:15:43 +08:00
  • 069125686e Fixes to logging statements. Daniel Povey 2022-10-22 15:08:07 +08:00
  • 1d4382555c Increase initial-lr from 0.06 to 0.075 and decrease lr-epochs from 3.5 to 3. Daniel Povey 2022-10-22 15:04:08 +08:00
  • 74d775014d Increase initial-lr from 0.05 to 0.06. Daniel Povey 2022-10-22 15:02:07 +08:00
  • aa5f34af64 Cosmetic change Daniel Povey 2022-10-22 15:00:15 +08:00
  • 1ec9fe5c98 Make warmup period decrease scale on simple loss, leaving pruned loss scale constant. Daniel Povey 2022-10-22 14:48:53 +08:00
  • efde3757c7 Reset optimizer state when we change loss function definition. Daniel Povey 2022-10-22 14:30:18 +08:00
  • 84580ec022 Configuration changes: scores limit 5->10, min_prob 0.05->0.1, cur_grad_scale more aggressive increase Daniel Povey 2022-10-22 14:09:53 +08:00
  • 9672dffac2 Merge branch 'scaled_adam_exp168' into scaled_adam_exp169 Daniel Povey 2022-10-22 14:05:07 +08:00
  • 8d1021d131 Remove comparison diagnostics, which were not that useful. Daniel Povey 2022-10-22 13:57:00 +08:00
  • 2e93e5d3b7 Add logging Daniel Povey 2022-10-22 13:52:51 +08:00
  • fd3f21f84d Changes to grad scale logging; increase grad scale more frequently if less than one. Daniel Povey 2022-10-22 13:36:26 +08:00
  • 348494888d
    Add kaldifst to requirements.txt (#631) Fangjun Kuang 2022-10-22 13:14:44 +08:00
  • 83cfc78c0e Add kaldifst to requirements.txt Fangjun Kuang 2022-10-22 13:13:23 +08:00
  • 1d2fe8e3c2 Add more diagnostics to debug gradient scale problems Daniel Povey 2022-10-22 12:49:29 +08:00
  • 3298e18732 Increase limit on logit for SimpleCombiner to 25.0 Daniel Povey 2022-10-21 22:06:05 +08:00
  • e5fe3de17e Also apply limit on logit in SimpleCombiner Daniel Povey 2022-10-21 21:31:52 +08:00
  • bdbd2cfce6 Penalize too large weights in softmax of AttentionDownsample() Daniel Povey 2022-10-21 20:12:36 +08:00
  • 9b671e1c21
    Add Shallow fusion in modified_beam_search (#630) ezerhouni 2022-10-21 10:44:56 +02:00
  • f5f3cf2869 Fix code according to review Erwan 2022-10-21 09:45:41 +02:00
  • 476fb9e9f3 Reduce min_prob of ActivationBalancer from 0.1 to 0.05. Daniel Povey 2022-10-21 15:42:04 +08:00
  • 13411dbee0 Modified beam search with ngram rescoring Erwan 2022-10-20 21:53:41 +02:00
  • 9f68b5717c Reduce the limit on attention weights from 50 to 25. Daniel Povey 2022-10-21 12:13:23 +08:00
  • 347031d12d revert utils pkufool 2022-10-21 11:50:31 +08:00
  • 9c5488ee03 support streaming for aishell stateless2; add delay penalty to aishell pkufool 2022-10-21 11:39:28 +08:00
  • 96bb44eb04 Use shallow fusion for modified-beam-search Fangjun Kuang 2022-10-18 23:42:11 +08:00
  • b7782bbe0c test batch size == 1 without shallow fusion Fangjun Kuang 2022-10-17 17:29:58 +08:00
  • 2802ef7700 Add utility for shallow fusion Fangjun Kuang 2022-10-10 09:20:28 +08:00
  • c5cb52fed1 Remove the use of random_clamp in conformer.py. Daniel Povey 2022-10-20 19:54:38 +08:00
  • dccff6b893 Remove use of RandomGrad Daniel Povey 2022-10-20 19:35:11 +08:00
  • 1018a77410 Use normal implementation of softmax. Daniel Povey 2022-10-20 19:34:10 +08:00
  • 6e6209419c Merge branch 'scaled_adam_exp150' into scaled_adam_exp155 Daniel Povey 2022-10-20 15:04:27 +08:00
  • 4565d43d5c Add hard limit of attention weights to +- 50 Daniel Povey 2022-10-20 14:27:55 +08:00