mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-08-08 09:32:20 +00:00
Small fix (#1686)
This commit is contained in:
parent
785f3f0bcf
commit
d65187ec52
@ -636,8 +636,9 @@ class ChunkCausalDepthwiseConv1d(torch.nn.Module):
|
||||
)
|
||||
|
||||
def forward(self, x: Tensor, chunk_size: int = -1) -> Tensor:
|
||||
"""
|
||||
Forward function. Args:
|
||||
"""Forward function.
|
||||
|
||||
Args:
|
||||
x: a Tensor of shape (batch_size, channels, seq_len)
|
||||
chunk_size: the chunk size, in frames; does not have to divide seq_len exactly.
|
||||
"""
|
||||
|
@ -406,7 +406,7 @@ def get_parser():
|
||||
"--context-size",
|
||||
type=int,
|
||||
default=2,
|
||||
help="The context size in the decoder. 1 means bigram; " "2 means tri-gram",
|
||||
help="The context size in the decoder. 1 means bigram; 2 means tri-gram",
|
||||
)
|
||||
|
||||
parser.add_argument(
|
||||
@ -429,7 +429,7 @@ def get_parser():
|
||||
"--am-scale",
|
||||
type=float,
|
||||
default=0.0,
|
||||
help="The scale to smooth the loss with am (output of encoder network)" "part.",
|
||||
help="The scale to smooth the loss with am (output of encoder network) part.",
|
||||
)
|
||||
|
||||
parser.add_argument(
|
||||
@ -848,7 +848,7 @@ def compute_loss(
|
||||
True for training. False for validation. When it is True, this
|
||||
function enables autograd during computation; when it is False, it
|
||||
disables autograd.
|
||||
warmup: a floating point value which increases throughout training;
|
||||
warmup: a floating point value which increases throughout training;
|
||||
values >= 1.0 are fully warmed up and have all modules present.
|
||||
"""
|
||||
device = model.device if isinstance(model, DDP) else next(model.parameters()).device
|
||||
|
Loading…
x
Reference in New Issue
Block a user