Fix parameters_names in train.py (#1121)

This commit is contained in:
Yifan Yang 2023-06-08 16:54:05 +08:00 committed by GitHub
parent 3ae47a4940
commit dca21c2a17
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

View File

@ -1127,7 +1127,16 @@ def run(rank, world_size, args):
logging.info("Using DDP")
model = DDP(model, device_ids=[rank], find_unused_parameters=True)
optimizer = ScaledAdam(model.parameters(), lr=params.base_lr, clipping_scale=2.0)
parameters_names = []
parameters_names.append(
[name_param_pair[0] for name_param_pair in model.named_parameters()]
)
optimizer = ScaledAdam(
model.parameters(),
lr=params.base_lr,
clipping_scale=2.0,
parameters_names=parameters_names,
)
scheduler = Eden(optimizer, params.lr_batches, params.lr_epochs)