mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-12-11 06:55:27 +00:00
Revert zipformer.py to status on previous commit
This commit is contained in:
parent
e952598677
commit
62b42887b4
@ -207,15 +207,6 @@ class Zipformer(EncoderInterface):
|
|||||||
attention_share_layers=attention_share_layers[i],
|
attention_share_layers=attention_share_layers[i],
|
||||||
)
|
)
|
||||||
|
|
||||||
# modify the layerdrop schedule with an extra schedule that takes longer
|
|
||||||
# to warm up for the less-downsampled layers; this encourages the more
|
|
||||||
# heavily downsampled layers to learn something.
|
|
||||||
|
|
||||||
extra_layerdrop = ScheduledFloat((0.0, 0.2), (20000.0 / downsampling_factor[i], 0.0))
|
|
||||||
for layer in encoder.layers:
|
|
||||||
# we can add objects of type ScheduledFloat.
|
|
||||||
layer.layer_skip_rate = layer.layer_skip_rate + extra_layerdrop
|
|
||||||
|
|
||||||
if downsampling_factor[i] != 1:
|
if downsampling_factor[i] != 1:
|
||||||
encoder = DownsampledZipformerEncoder(
|
encoder = DownsampledZipformerEncoder(
|
||||||
encoder,
|
encoder,
|
||||||
@ -229,9 +220,6 @@ class Zipformer(EncoderInterface):
|
|||||||
encoder.lr_scale = downsampling_factor[i] ** -0.25
|
encoder.lr_scale = downsampling_factor[i] ** -0.25
|
||||||
|
|
||||||
encoders.append(encoder)
|
encoders.append(encoder)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
self.encoders = nn.ModuleList(encoders)
|
self.encoders = nn.ModuleList(encoders)
|
||||||
|
|
||||||
# initializes self.skip_layers and self.skip_modules
|
# initializes self.skip_layers and self.skip_modules
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user