mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-12-11 06:55:27 +00:00
from local
This commit is contained in:
parent
3eb3818fc4
commit
b6d15805fd
Binary file not shown.
@ -1602,16 +1602,15 @@ def run_adapter(rank, world_size, args, wb=None):
|
|||||||
else:
|
else:
|
||||||
p.requires_grad = False
|
p.requires_grad = False
|
||||||
'''
|
'''
|
||||||
|
|
||||||
for n, p in model.named_parameters():
|
|
||||||
p.requires_grad = False
|
|
||||||
|
|
||||||
optimizer_adapter = ScaledAdam(
|
optimizer_adapter = ScaledAdam(
|
||||||
adapter_param,
|
adapter_param,
|
||||||
lr=params.adapter_lr,
|
lr=params.adapter_lr,
|
||||||
clipping_scale=5.0,
|
clipping_scale=5.0,
|
||||||
parameters_names=[adapter_names],
|
parameters_names=[adapter_names],
|
||||||
)
|
)
|
||||||
|
|
||||||
|
#for n, p in model.named_parameters():
|
||||||
|
# p.requires_grad = False
|
||||||
|
|
||||||
#prompt = torch.randn((100, 512), requires_grad=True)
|
#prompt = torch.randn((100, 512), requires_grad=True)
|
||||||
#optimizer_adapter = ScaledAdam(
|
#optimizer_adapter = ScaledAdam(
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user