Merge 65212ee0041db43c826c5331c907730ba7c87cd4 into abd9437e6d5419a497707748eb935e50976c3b7b

This commit is contained in:
Fangjun Kuang 2025-06-27 11:31:06 +00:00 committed by GitHub
commit 0ac15ce71d
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -53,7 +53,6 @@ from lhotse.utils import fix_random_seed
from model import Transducer from model import Transducer
from torch import Tensor from torch import Tensor
from torch.nn.parallel import DistributedDataParallel as DDP from torch.nn.parallel import DistributedDataParallel as DDP
from torch.nn.utils import clip_grad_norm_
from torch.utils.tensorboard import SummaryWriter from torch.utils.tensorboard import SummaryWriter
from transformer import Noam from transformer import Noam
@ -527,7 +526,6 @@ def train_one_epoch(
optimizer.zero_grad() optimizer.zero_grad()
loss.backward() loss.backward()
clip_grad_norm_(model.parameters(), 5.0, 2.0)
optimizer.step() optimizer.step()
if params.print_diagnostics and batch_idx == 30: if params.print_diagnostics and batch_idx == 30:
return return
@ -737,7 +735,6 @@ def scan_pessimistic_batches_for_oom(
is_training=True, is_training=True,
) )
loss.backward() loss.backward()
clip_grad_norm_(model.parameters(), 5.0, 2.0)
optimizer.step() optimizer.step()
except RuntimeError as e: except RuntimeError as e:
if "CUDA out of memory" in str(e): if "CUDA out of memory" in str(e):