from local
This commit is contained in:
parent
03d1d62967
commit
29a7c8d235
Binary file not shown.
@ -1575,6 +1575,7 @@ def run_adapter(rank, world_size, args, wb=None):
|
|||||||
logging.info("Using DDP")
|
logging.info("Using DDP")
|
||||||
model = DDP(model, device_ids=[rank], find_unused_parameters=True)
|
model = DDP(model, device_ids=[rank], find_unused_parameters=True)
|
||||||
|
|
||||||
|
'''
|
||||||
adapter_names = []
|
adapter_names = []
|
||||||
adapter_param = []
|
adapter_param = []
|
||||||
for n, p in model.named_parameters():
|
for n, p in model.named_parameters():
|
||||||
@ -1585,6 +1586,12 @@ def run_adapter(rank, world_size, args, wb=None):
|
|||||||
p.requires_grad = True
|
p.requires_grad = True
|
||||||
else:
|
else:
|
||||||
p.requires_grad = False
|
p.requires_grad = False
|
||||||
|
'''
|
||||||
|
|
||||||
|
for n, p in model.named_parameters():
|
||||||
|
p.requires_grad = False
|
||||||
|
|
||||||
|
prompt = torch.nn.Parameter(torch.randn(50, 512))
|
||||||
|
|
||||||
optimizer_adapter = ScaledAdam(
|
optimizer_adapter = ScaledAdam(
|
||||||
adapter_param,
|
adapter_param,
|
||||||
|
|||||||
Reference in New Issue
Block a user