From c13700987f81f0da33043f5c82cdaa18d6e2cb4a Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Fri, 24 Feb 2023 10:35:09 +0900 Subject: [PATCH] from local --- .../ASR/conformer_ctc2/.train.py.swp | Bin 65536 -> 65536 bytes egs/librispeech/ASR/conformer_ctc2/train.py | 3 +-- 2 files changed, 1 insertion(+), 2 deletions(-) diff --git a/egs/librispeech/ASR/conformer_ctc2/.train.py.swp b/egs/librispeech/ASR/conformer_ctc2/.train.py.swp index 1eff68e841c234bb9639af1e4e1c9717675d2d5c..49b927ed90bb02c7ab2bb8a05334ad2cc66982b3 100644 GIT binary patch delta 131 zcmZo@U}4Y1c$=fcS8DL8u`@6jaX?J7nJn1wp3NA@ ZGve6H)~Lfid0&nA5YVf{dV_^8j4l&JkvS7n|HbxEx ahTrU)*&21&Coj0hGWm9g?dIZ+-f{qdtS8d| diff --git a/egs/librispeech/ASR/conformer_ctc2/train.py b/egs/librispeech/ASR/conformer_ctc2/train.py index 68c3ef085..6b3721f7d 100755 --- a/egs/librispeech/ASR/conformer_ctc2/train.py +++ b/egs/librispeech/ASR/conformer_ctc2/train.py @@ -329,7 +329,6 @@ def get_parser() -> argparse.ArgumentParser: "--unused-params", type=str2bool, default=False, - help="Whether to use half precision training.", ) add_model_arguments(parser) @@ -1067,7 +1066,7 @@ def run(rank, world_size, args): model.to(device) if world_size > 1: logging.info("Using DDP") - model = DDP(model, device_ids=[rank], find_unused_parameters=False) + model = DDP(model, device_ids=[rank], find_unused_parameters=params.unused_params) optimizer = optim.Eve(model.parameters(), lr=params.initial_lr) scheduler = optim.Eden(optimizer, params.lr_batches, params.lr_epochs)