mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-12-10 22:45:27 +00:00
revert unrelated transformer.py diffs from rebase
This commit is contained in:
parent
c23af2ea1a
commit
ed79fa3c04
@ -545,7 +545,6 @@ class TransformerDecoderLayer(nn.Module):
|
|||||||
memory_mask: Optional[torch.Tensor] = None,
|
memory_mask: Optional[torch.Tensor] = None,
|
||||||
tgt_key_padding_mask: Optional[torch.Tensor] = None,
|
tgt_key_padding_mask: Optional[torch.Tensor] = None,
|
||||||
memory_key_padding_mask: Optional[torch.Tensor] = None,
|
memory_key_padding_mask: Optional[torch.Tensor] = None,
|
||||||
**kwargs,
|
|
||||||
) -> torch.Tensor:
|
) -> torch.Tensor:
|
||||||
"""Pass the inputs (and mask) through the decoder layer.
|
"""Pass the inputs (and mask) through the decoder layer.
|
||||||
|
|
||||||
|
|||||||
@ -549,7 +549,6 @@ class TransformerDecoderLayer(nn.Module):
|
|||||||
memory_mask: Optional[torch.Tensor] = None,
|
memory_mask: Optional[torch.Tensor] = None,
|
||||||
tgt_key_padding_mask: Optional[torch.Tensor] = None,
|
tgt_key_padding_mask: Optional[torch.Tensor] = None,
|
||||||
memory_key_padding_mask: Optional[torch.Tensor] = None,
|
memory_key_padding_mask: Optional[torch.Tensor] = None,
|
||||||
**kwargs,
|
|
||||||
) -> torch.Tensor:
|
) -> torch.Tensor:
|
||||||
"""Pass the inputs (and mask) through the decoder layer.
|
"""Pass the inputs (and mask) through the decoder layer.
|
||||||
|
|
||||||
|
|||||||
@ -549,7 +549,6 @@ class TransformerDecoderLayer(nn.Module):
|
|||||||
memory_mask: Optional[torch.Tensor] = None,
|
memory_mask: Optional[torch.Tensor] = None,
|
||||||
tgt_key_padding_mask: Optional[torch.Tensor] = None,
|
tgt_key_padding_mask: Optional[torch.Tensor] = None,
|
||||||
memory_key_padding_mask: Optional[torch.Tensor] = None,
|
memory_key_padding_mask: Optional[torch.Tensor] = None,
|
||||||
**kwargs,
|
|
||||||
) -> torch.Tensor:
|
) -> torch.Tensor:
|
||||||
"""Pass the inputs (and mask) through the decoder layer.
|
"""Pass the inputs (and mask) through the decoder layer.
|
||||||
|
|
||||||
|
|||||||
@ -550,7 +550,6 @@ class TransformerDecoderLayer(nn.Module):
|
|||||||
tgt_key_padding_mask: Optional[torch.Tensor] = None,
|
tgt_key_padding_mask: Optional[torch.Tensor] = None,
|
||||||
memory_key_padding_mask: Optional[torch.Tensor] = None,
|
memory_key_padding_mask: Optional[torch.Tensor] = None,
|
||||||
warmup: float = 1.0,
|
warmup: float = 1.0,
|
||||||
**kwargs,
|
|
||||||
) -> torch.Tensor:
|
) -> torch.Tensor:
|
||||||
"""Pass the inputs (and mask) through the decoder layer.
|
"""Pass the inputs (and mask) through the decoder layer.
|
||||||
|
|
||||||
|
|||||||
@ -537,7 +537,6 @@ class TransformerDecoderLayer(nn.Module):
|
|||||||
memory_mask: Optional[torch.Tensor] = None,
|
memory_mask: Optional[torch.Tensor] = None,
|
||||||
tgt_key_padding_mask: Optional[torch.Tensor] = None,
|
tgt_key_padding_mask: Optional[torch.Tensor] = None,
|
||||||
memory_key_padding_mask: Optional[torch.Tensor] = None,
|
memory_key_padding_mask: Optional[torch.Tensor] = None,
|
||||||
**kwargs,
|
|
||||||
) -> torch.Tensor:
|
) -> torch.Tensor:
|
||||||
"""Pass the inputs (and mask) through the decoder layer.
|
"""Pass the inputs (and mask) through the decoder layer.
|
||||||
|
|
||||||
|
|||||||
@ -567,7 +567,6 @@ class TransformerDecoderLayer(nn.Module):
|
|||||||
memory_mask: Optional[torch.Tensor] = None,
|
memory_mask: Optional[torch.Tensor] = None,
|
||||||
tgt_key_padding_mask: Optional[torch.Tensor] = None,
|
tgt_key_padding_mask: Optional[torch.Tensor] = None,
|
||||||
memory_key_padding_mask: Optional[torch.Tensor] = None,
|
memory_key_padding_mask: Optional[torch.Tensor] = None,
|
||||||
**kwargs,
|
|
||||||
) -> torch.Tensor:
|
) -> torch.Tensor:
|
||||||
"""Pass the inputs (and mask) through the decoder layer.
|
"""Pass the inputs (and mask) through the decoder layer.
|
||||||
|
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user