diff --git a/egs/librispeech/ASR/pruned_transducer_stateless5/.conformer.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless5/.conformer.py.swp index de85ef1e3..2fac7b89b 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless5/.conformer.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless5/.conformer.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless5/.train.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless5/.train.py.swp index 65041cae8..e4ce674a1 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless5/.train.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless5/.train.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless5/train.py b/egs/librispeech/ASR/pruned_transducer_stateless5/train.py index ff70348e9..378e44612 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless5/train.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless5/train.py @@ -983,7 +983,7 @@ def run(rank, world_size, args): model.to(device) if world_size > 1: logging.info("Using DDP") - model = DDP(model, device_ids=[rank], find_unused_parameters=False) + model = DDP(model, device_ids=[rank], find_unused_parameters=True) optimizer = Eve(model.parameters(), lr=params.initial_lr)