Merge 65212ee0041db43c826c5331c907730ba7c87cd4 into abd9437e6d5419a497707748eb935e50976c3b7b

This commit is contained in:
Fangjun Kuang 2025-06-27 11:31:06 +00:00 committed by GitHub
commit 0ac15ce71d
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -53,7 +53,6 @@ from lhotse.utils import fix_random_seed
from model import Transducer
from torch import Tensor
from torch.nn.parallel import DistributedDataParallel as DDP
from torch.nn.utils import clip_grad_norm_
from torch.utils.tensorboard import SummaryWriter
from transformer import Noam
@ -527,7 +526,6 @@ def train_one_epoch(
optimizer.zero_grad()
loss.backward()
clip_grad_norm_(model.parameters(), 5.0, 2.0)
optimizer.step()
if params.print_diagnostics and batch_idx == 30:
return
@ -737,7 +735,6 @@ def scan_pessimistic_batches_for_oom(
is_training=True,
)
loss.backward()
clip_grad_norm_(model.parameters(), 5.0, 2.0)
optimizer.step()
except RuntimeError as e:
if "CUDA out of memory" in str(e):