mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-08-08 09:32:20 +00:00
Fix a bug in finetune.py --use-mux (#1949)
This commit is contained in:
parent
fd8f8780fa
commit
30e7ea4b5a
@ -140,8 +140,8 @@ def add_finetune_arguments(parser: argparse.ArgumentParser):
|
|||||||
type=str2bool,
|
type=str2bool,
|
||||||
default=False,
|
default=False,
|
||||||
help="""
|
help="""
|
||||||
Whether to adapt. If true, we will mix 5% of the new data
|
Whether to adapt. If true, we will mix 5%% of the new data
|
||||||
with 95% of the original data to fine-tune. This is useful
|
with 95%% of the original data to fine-tune. This is useful
|
||||||
if you want to maintain the performance on the original domain
|
if you want to maintain the performance on the original domain
|
||||||
""",
|
""",
|
||||||
)
|
)
|
||||||
@ -1134,7 +1134,7 @@ def train_one_epoch(
|
|||||||
f"Epoch {params.cur_epoch}, "
|
f"Epoch {params.cur_epoch}, "
|
||||||
f"batch {batch_idx}, loss[{loss_info}], "
|
f"batch {batch_idx}, loss[{loss_info}], "
|
||||||
f"tot_loss[{tot_loss}], batch size: {batch_size}, "
|
f"tot_loss[{tot_loss}], batch size: {batch_size}, "
|
||||||
f"lr: {cur_lr:.2e}, "
|
f"lr: {cur_lr: .2e}, "
|
||||||
+ (f"grad_scale: {scaler._scale.item()}" if params.use_fp16 else "")
|
+ (f"grad_scale: {scaler._scale.item()}" if params.use_fp16 else "")
|
||||||
)
|
)
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user