mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-08-09 10:02:22 +00:00
Fix Zipformer (#1132)
* Update model.py * Update train.py * Update decoder.py
This commit is contained in:
parent
947f0614c9
commit
0a465794a8
@ -58,7 +58,6 @@ class Decoder(nn.Module):
|
|||||||
self.embedding = nn.Embedding(
|
self.embedding = nn.Embedding(
|
||||||
num_embeddings=vocab_size,
|
num_embeddings=vocab_size,
|
||||||
embedding_dim=decoder_dim,
|
embedding_dim=decoder_dim,
|
||||||
padding_idx=blank_id,
|
|
||||||
)
|
)
|
||||||
# the balancers are to avoid any drift in the magnitude of the
|
# the balancers are to avoid any drift in the magnitude of the
|
||||||
# embeddings, which would interact badly with parameter averaging.
|
# embeddings, which would interact badly with parameter averaging.
|
||||||
|
@ -333,7 +333,7 @@ class AsrModel(nn.Module):
|
|||||||
simple_loss, pruned_loss = self.forward_transducer(
|
simple_loss, pruned_loss = self.forward_transducer(
|
||||||
encoder_out=encoder_out,
|
encoder_out=encoder_out,
|
||||||
encoder_out_lens=encoder_out_lens,
|
encoder_out_lens=encoder_out_lens,
|
||||||
y=y,
|
y=y.to(x.device),
|
||||||
y_lens=y_lens,
|
y_lens=y_lens,
|
||||||
prune_range=prune_range,
|
prune_range=prune_range,
|
||||||
am_scale=am_scale,
|
am_scale=am_scale,
|
||||||
|
@ -789,7 +789,7 @@ def compute_loss(
|
|||||||
|
|
||||||
texts = batch["supervisions"]["text"]
|
texts = batch["supervisions"]["text"]
|
||||||
y = sp.encode(texts, out_type=int)
|
y = sp.encode(texts, out_type=int)
|
||||||
y = k2.RaggedTensor(y).to(device)
|
y = k2.RaggedTensor(y)
|
||||||
|
|
||||||
with torch.set_grad_enabled(is_training):
|
with torch.set_grad_enabled(is_training):
|
||||||
simple_loss, pruned_loss, ctc_loss = model(
|
simple_loss, pruned_loss, ctc_loss = model(
|
||||||
|
Loading…
x
Reference in New Issue
Block a user