from local

This commit is contained in:
dohe0342 2023-05-02 13:47:14 +09:00
parent 630da09184
commit e7d2df8a78
2 changed files with 6 additions and 8 deletions

View File

@ -1584,7 +1584,6 @@ def run_adapter(rank, world_size, args, wb=None):
logging.info("Using DDP") logging.info("Using DDP")
model = DDP(model, device_ids=[rank], find_unused_parameters=True) model = DDP(model, device_ids=[rank], find_unused_parameters=True)
'''
adapter_names = [] adapter_names = []
adapter_param = [] adapter_param = []
for n, p in model.named_parameters(): for n, p in model.named_parameters():
@ -1605,15 +1604,14 @@ def run_adapter(rank, world_size, args, wb=None):
clipping_scale=5.0, clipping_scale=5.0,
parameters_names=[adapter_names], parameters_names=[adapter_names],
) )
'''
#prompt = torch.randn((100, 512), requires_grad=True) #prompt = torch.randn((100, 512), requires_grad=True)
optimizer_adapter = ScaledAdam( #optimizer_adapter = ScaledAdam(
[model.prompt], # [model.prompt],
lr=params.adapter_lr, # lr=params.adapter_lr,
clipping_scale=5.0, # clipping_scale=5.0,
parameters_names=['P'], # parameters_names=['P'],
) #)
scheduler_adapter = Eden(optimizer_adapter, 10000, 7) #params.lr_batche, params.lr_epochs) scheduler_adapter = Eden(optimizer_adapter, 10000, 7) #params.lr_batche, params.lr_epochs)