mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-08-09 01:52:41 +00:00
Change some defaults in LR-setting rule.
This commit is contained in:
parent
2545237eb3
commit
a41e93437c
@ -60,7 +60,7 @@ class Eve(Optimizer):
|
||||
"""
|
||||
|
||||
def __init__(self, params, lr=1e-3, betas=(0.9, 0.98), eps=1e-8,
|
||||
weight_decay=3e-4, target_rms=0.1):
|
||||
weight_decay=1e-3, target_rms=0.1):
|
||||
|
||||
if not 0.0 <= lr:
|
||||
raise ValueError("Invalid learning rate: {}".format(lr))
|
||||
|
@ -161,7 +161,7 @@ def get_parser():
|
||||
parser.add_argument(
|
||||
"--lr-power",
|
||||
type=float,
|
||||
default=0.5,
|
||||
default=0.75,
|
||||
help="Power in LR-setting rule",
|
||||
)
|
||||
|
||||
@ -780,8 +780,7 @@ def run(rank, world_size, args):
|
||||
|
||||
optimizer = Eve(
|
||||
model.parameters(),
|
||||
lr=params.initial_lr, betas=(0.9, 0.98),
|
||||
eps=1e-9, weight_decay=3e-04, target_rms=0.1)
|
||||
lr=params.initial_lr)
|
||||
scheduler = torch.optim.lr_scheduler.LambdaLR(
|
||||
optimizer,
|
||||
lambda step: ((params.lr_num_steps/(step + params.lr_num_steps)) ** params.lr_power))
|
||||
|
Loading…
x
Reference in New Issue
Block a user