diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp index e650fc3ad..02f6b6aff 100644 Binary files a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp and b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp differ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py index 086c1f25d..06190504f 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py @@ -1097,7 +1097,7 @@ def train_one_epoch( f"grad_scale is too small, exiting: {cur_grad_scale}" ) - if params.batch_idx_train > 4000 and loss > 300: + if params.batch_idx_train > 4000 and loss > 300 and params.wandb: wb.log({"valid/loss": 10000}) raise RunteimError( f"divergence... exiting: loss={loss}"