Change configuration again.. not great performance.

This commit is contained in:
Daniel Povey 2021-09-07 20:58:00 +08:00
parent 573e0582d8
commit d313c27c14

View File

@ -132,7 +132,8 @@ def get_params() -> AttributeDict:
"""
params = AttributeDict(
{
"exp_dir": Path("conformer_lm/exp_2"),
# exp_3, vs. exp_2, is using 5e-04 not 2d-04 as max learning rate.
"exp_dir": Path("conformer_lm/exp_3"),
"lm_dataset": Path("data/lm_training_5000/lm_data.pt"),
"num_tokens": 5000,
"blank_sym": 0,
@ -155,7 +156,7 @@ def get_params() -> AttributeDict:
"attention_dim": 512,
"nhead": 8,
"num_decoder_layers": 6,
"max_lrate": 2.0e-04 # was 5.0e-04, then from start_epoch=9 used max_lrate=2.0e-04, then from start_epoch=11 used 1.0e-04.
"max_lrate": 5.0e-04
}
)