mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-12-11 06:55:27 +00:00
from local
This commit is contained in:
parent
630da09184
commit
e7d2df8a78
Binary file not shown.
@ -1584,7 +1584,6 @@ def run_adapter(rank, world_size, args, wb=None):
|
|||||||
logging.info("Using DDP")
|
logging.info("Using DDP")
|
||||||
model = DDP(model, device_ids=[rank], find_unused_parameters=True)
|
model = DDP(model, device_ids=[rank], find_unused_parameters=True)
|
||||||
|
|
||||||
'''
|
|
||||||
adapter_names = []
|
adapter_names = []
|
||||||
adapter_param = []
|
adapter_param = []
|
||||||
for n, p in model.named_parameters():
|
for n, p in model.named_parameters():
|
||||||
@ -1605,15 +1604,14 @@ def run_adapter(rank, world_size, args, wb=None):
|
|||||||
clipping_scale=5.0,
|
clipping_scale=5.0,
|
||||||
parameters_names=[adapter_names],
|
parameters_names=[adapter_names],
|
||||||
)
|
)
|
||||||
'''
|
|
||||||
|
|
||||||
#prompt = torch.randn((100, 512), requires_grad=True)
|
#prompt = torch.randn((100, 512), requires_grad=True)
|
||||||
optimizer_adapter = ScaledAdam(
|
#optimizer_adapter = ScaledAdam(
|
||||||
[model.prompt],
|
# [model.prompt],
|
||||||
lr=params.adapter_lr,
|
# lr=params.adapter_lr,
|
||||||
clipping_scale=5.0,
|
# clipping_scale=5.0,
|
||||||
parameters_names=['P'],
|
# parameters_names=['P'],
|
||||||
)
|
#)
|
||||||
|
|
||||||
scheduler_adapter = Eden(optimizer_adapter, 10000, 7) #params.lr_batche, params.lr_epochs)
|
scheduler_adapter = Eden(optimizer_adapter, 10000, 7) #params.lr_batche, params.lr_epochs)
|
||||||
|
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user