from local

This commit is contained in:
dohe0342 2023-05-24 13:37:08 +09:00
parent 802e3ebbe8
commit 177bce2db4
2 changed files with 7 additions and 6 deletions

View File

@ -1591,17 +1591,18 @@ def run_adapter(rank, world_size, args, wb=None):
logging.info("Using DDP") logging.info("Using DDP")
model = DDP(model, device_ids=[rank], find_unused_parameters=True) model = DDP(model, device_ids=[rank], find_unused_parameters=True)
lora_module = [] lora_modules = []
for i, module in enumerate(model.modules()): for modules in model.modules():
if isinstance(module, fairseq.modules.multihead_attention.MultiheadAttention): if isinstance(module, fairseq.modules.multihead_attention.MultiheadAttention):
for m in module.modules(): for module in modules.modules():
lora_module.append(LoRAHook(m)) lora_modules.append(LoRAHook(m))
adapter_names = [] adapter_names = []
adapter_param = [] adapter_param = []
for lora in lora_module
'''
for n, p in model.named_parameters(): for n, p in model.named_parameters():
print(n) print(n)
'''
if 'adapters' in n:# or 'joiner' in n or 'simple' in n or 'ctc' in n: if 'adapters' in n:# or 'joiner' in n or 'simple' in n or 'ctc' in n:
adapter_names.append(n) adapter_names.append(n)
adapter_param.append(p) adapter_param.append(p)
@ -1609,7 +1610,7 @@ def run_adapter(rank, world_size, args, wb=None):
p.requires_grad = True p.requires_grad = True
else: else:
p.requires_grad = False p.requires_grad = False
''' '''
optimizer_adapter = ScaledAdam( optimizer_adapter = ScaledAdam(
adapter_param, adapter_param,
lr=params.adapter_lr, lr=params.adapter_lr,