Inconseqeuential config change

This commit is contained in:
Daniel Povey 2022-07-08 11:03:16 +08:00
parent be6680e3ba
commit e9ab1ddd39

View File

@ -67,7 +67,7 @@ class LearnedGradient(Optimizer):
param_min_rms=1.0e-05, param_min_rms=1.0e-05,
param_max_rms=2.0, param_max_rms=2.0,
lr_mat_min=0.01, lr_mat_min=0.01,
lr_mat_max=4.0, lr_mat_max=10.0,
lr_est_period=2, lr_est_period=2,
diagonalize_period=4, diagonalize_period=4,
): ):
@ -474,7 +474,8 @@ class LearnedGradient(Optimizer):
# are the final changes, the only 2 we make in this loop that have # are the final changes, the only 2 we make in this loop that have
# side effects. # side effects.
# delta_scale < 1 will make it update the learning rates faster than it otherwise would. # delta_scale < 1 will make it update the learning rates faster than it otherwise would,
# as we'll reach equilbrium with M less rapidly.
delta_scale=1.0 delta_scale=1.0
delta.add_(this_delta, alpha=-delta_scale*meta_lr*(1-beta1)) delta.add_(this_delta, alpha=-delta_scale*meta_lr*(1-beta1))
# there is no momentum on Q. # there is no momentum on Q.