Change some defaults in LR-setting rule.

This commit is contained in:
Daniel Povey 2022-04-06 12:36:58 +08:00
parent 2545237eb3
commit a41e93437c
2 changed files with 3 additions and 4 deletions

View File

@ -60,7 +60,7 @@ class Eve(Optimizer):
"""
def __init__(self, params, lr=1e-3, betas=(0.9, 0.98), eps=1e-8,
weight_decay=3e-4, target_rms=0.1):
weight_decay=1e-3, target_rms=0.1):
if not 0.0 <= lr:
raise ValueError("Invalid learning rate: {}".format(lr))

View File

@ -161,7 +161,7 @@ def get_parser():
parser.add_argument(
"--lr-power",
type=float,
default=0.5,
default=0.75,
help="Power in LR-setting rule",
)
@ -780,8 +780,7 @@ def run(rank, world_size, args):
optimizer = Eve(
model.parameters(),
lr=params.initial_lr, betas=(0.9, 0.98),
eps=1e-9, weight_decay=3e-04, target_rms=0.1)
lr=params.initial_lr)
scheduler = torch.optim.lr_scheduler.LambdaLR(
optimizer,
lambda step: ((params.lr_num_steps/(step + params.lr_num_steps)) ** params.lr_power))