diff --git a/egs/librispeech/ASR/conformer_ctc2/.train.py.swp b/egs/librispeech/ASR/conformer_ctc2/.train.py.swp index 03759765b..122573310 100644 Binary files a/egs/librispeech/ASR/conformer_ctc2/.train.py.swp and b/egs/librispeech/ASR/conformer_ctc2/.train.py.swp differ diff --git a/egs/librispeech/ASR/conformer_ctc2/train.py b/egs/librispeech/ASR/conformer_ctc2/train.py index 99637a0b5..86e5fb7e9 100755 --- a/egs/librispeech/ASR/conformer_ctc2/train.py +++ b/egs/librispeech/ASR/conformer_ctc2/train.py @@ -685,7 +685,8 @@ def train_one_epoch( params.batch_idx_train += 1 batch_size = len(batch["supervisions"]["text"]) #print(batch["supervisions"]) - batch_name = batch["supervisions"]["id"] + #batch_name = batch["supervisions"]["id"] + batch_name = str(batch_idx) with torch.cuda.amp.autocast(enabled=params.use_fp16): loss, loss_info = compute_loss(