diff --git a/egs/librispeech/ASR/transducer_stateless/train.py b/egs/librispeech/ASR/transducer_stateless/train.py index baeff6016..186bf402b 100755 --- a/egs/librispeech/ASR/transducer_stateless/train.py +++ b/egs/librispeech/ASR/transducer_stateless/train.py @@ -53,7 +53,6 @@ from lhotse.utils import fix_random_seed from model import Transducer from torch import Tensor from torch.nn.parallel import DistributedDataParallel as DDP -from torch.nn.utils import clip_grad_norm_ from torch.utils.tensorboard import SummaryWriter from transformer import Noam @@ -527,7 +526,6 @@ def train_one_epoch( optimizer.zero_grad() loss.backward() - clip_grad_norm_(model.parameters(), 5.0, 2.0) optimizer.step() if params.print_diagnostics and batch_idx == 30: return @@ -737,7 +735,6 @@ def scan_pessimistic_batches_for_oom( is_training=True, ) loss.backward() - clip_grad_norm_(model.parameters(), 5.0, 2.0) optimizer.step() except RuntimeError as e: if "CUDA out of memory" in str(e):