mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-12-11 06:55:27 +00:00
Use dropout in attention, on attn weights.
This commit is contained in:
parent
24aea947d2
commit
ce3f59d9c7
@ -168,7 +168,7 @@ class ConformerEncoderLayer(nn.Module):
|
|||||||
self.d_model = d_model
|
self.d_model = d_model
|
||||||
|
|
||||||
self.self_attn = RelPositionMultiheadAttention(
|
self.self_attn = RelPositionMultiheadAttention(
|
||||||
d_model, nhead, dropout=0.0
|
d_model, nhead, dropout=dropout,
|
||||||
)
|
)
|
||||||
|
|
||||||
self.feed_forward = nn.Sequential(
|
self.feed_forward = nn.Sequential(
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user