diff --git a/egs/librispeech/ASR/.full_ft.sh.swp b/egs/librispeech/ASR/.full_ft.sh.swp index edbdcbcd2..c737dfb15 100644 Binary files a/egs/librispeech/ASR/.full_ft.sh.swp and b/egs/librispeech/ASR/.full_ft.sh.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.full_ft.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.full_ft.py.swp index e454939ab..cfbbff9c7 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.full_ft.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.full_ft.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/full_ft.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/full_ft.py index 7522912d2..070d8349f 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/full_ft.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/full_ft.py @@ -1262,13 +1262,8 @@ def run(rank, world_size, args, wb=None): num_param = sum([p.numel() for p in model.parameters()]) logging.info(f"Number of model parameters: {num_param}") - - assert params.save_every_n >= params.average_period - model_avg: Optional[nn.Module] = None - if rank == 0: - # model_avg is only used with rank 0 - model_avg = copy.deepcopy(model).to(torch.float64) - + + model_avg = None assert params.start_epoch > 0, params.start_epoch checkpoints = load_checkpoint_if_available( params=params, model=model, model_avg=model_avg