mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-08-08 09:32:20 +00:00
Update conformer.py (#1857)
* Update conformer.py feedforward dimention -> feedforward dimension * Update conformer.py feedforward dimention -> feedforward dimension * Update conformer.py feedforward dimention -> feedforward dimension * Update conformer.py feedforward dimention -> feedforward dimension * Update conformer.py feedforward dimention -> feedforward dimension * Update conformer.py feedforward dimention -> feedforward dimension * Update conformer.py feedforward dimention -> feedforward dimension * Update conformer.py feedforward dimention -> feedforward dimension * Update conformer.py feedforward dimention -> feedforward dimension
This commit is contained in:
parent
3b263539cd
commit
3b6d54007b
@ -32,7 +32,7 @@ class Conformer(Transformer):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension
|
d_model (int): attention dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
num_decoder_layers (int): number of decoder layers
|
num_decoder_layers (int): number of decoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
|
@ -42,7 +42,7 @@ class Conformer(Transformer):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension, also the output dimension
|
d_model (int): attention dimension, also the output dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
num_decoder_layers (int): number of decoder layers
|
num_decoder_layers (int): number of decoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
|
@ -33,7 +33,7 @@ class Conformer(Transformer):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension
|
d_model (int): attention dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
num_decoder_layers (int): number of decoder layers
|
num_decoder_layers (int): number of decoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
|
@ -42,7 +42,7 @@ class Conformer(EncoderInterface):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension, also the output dimension
|
d_model (int): attention dimension, also the output dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
layer_dropout (float): layer-dropout rate.
|
layer_dropout (float): layer-dropout rate.
|
||||||
|
@ -42,7 +42,7 @@ class Conformer(EncoderInterface):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension, also the output dimension
|
d_model (int): attention dimension, also the output dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
layer_dropout (float): layer-dropout rate.
|
layer_dropout (float): layer-dropout rate.
|
||||||
|
@ -42,7 +42,7 @@ class Conformer(EncoderInterface):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension, also the output dimension
|
d_model (int): attention dimension, also the output dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
layer_dropout (float): layer-dropout rate.
|
layer_dropout (float): layer-dropout rate.
|
||||||
|
@ -42,7 +42,7 @@ class Conformer(EncoderInterface):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension, also the output dimension
|
d_model (int): attention dimension, also the output dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
layer_dropout (float): layer-dropout rate.
|
layer_dropout (float): layer-dropout rate.
|
||||||
|
@ -69,7 +69,7 @@ class Conformer(Transformer):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension
|
d_model (int): attention dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
num_decoder_layers (int): number of decoder layers
|
num_decoder_layers (int): number of decoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
|
@ -35,7 +35,7 @@ class Conformer(Transformer):
|
|||||||
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
subsampling_factor (int): subsampling factor of encoder (the convolution layers before transformers)
|
||||||
d_model (int): attention dimension
|
d_model (int): attention dimension
|
||||||
nhead (int): number of head
|
nhead (int): number of head
|
||||||
dim_feedforward (int): feedforward dimention
|
dim_feedforward (int): feedforward dimension
|
||||||
num_encoder_layers (int): number of encoder layers
|
num_encoder_layers (int): number of encoder layers
|
||||||
dropout (float): dropout rate
|
dropout (float): dropout rate
|
||||||
cnn_module_kernel (int): Kernel size of convolution module
|
cnn_module_kernel (int): Kernel size of convolution module
|
||||||
|
Loading…
x
Reference in New Issue
Block a user