mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-12-11 06:55:27 +00:00
from local
This commit is contained in:
parent
630da09184
commit
e7d2df8a78
Binary file not shown.
@ -1584,7 +1584,6 @@ def run_adapter(rank, world_size, args, wb=None):
|
||||
logging.info("Using DDP")
|
||||
model = DDP(model, device_ids=[rank], find_unused_parameters=True)
|
||||
|
||||
'''
|
||||
adapter_names = []
|
||||
adapter_param = []
|
||||
for n, p in model.named_parameters():
|
||||
@ -1605,15 +1604,14 @@ def run_adapter(rank, world_size, args, wb=None):
|
||||
clipping_scale=5.0,
|
||||
parameters_names=[adapter_names],
|
||||
)
|
||||
'''
|
||||
|
||||
#prompt = torch.randn((100, 512), requires_grad=True)
|
||||
optimizer_adapter = ScaledAdam(
|
||||
[model.prompt],
|
||||
lr=params.adapter_lr,
|
||||
clipping_scale=5.0,
|
||||
parameters_names=['P'],
|
||||
)
|
||||
#optimizer_adapter = ScaledAdam(
|
||||
# [model.prompt],
|
||||
# lr=params.adapter_lr,
|
||||
# clipping_scale=5.0,
|
||||
# parameters_names=['P'],
|
||||
#)
|
||||
|
||||
scheduler_adapter = Eden(optimizer_adapter, 10000, 7) #params.lr_batche, params.lr_epochs)
|
||||
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user