mirror of
https://github.com/k2-fsa/icefall.git
synced 2025-12-11 06:55:27 +00:00
Use dropout in attention, on attn weights.
This commit is contained in:
parent
24aea947d2
commit
ce3f59d9c7
@ -168,7 +168,7 @@ class ConformerEncoderLayer(nn.Module):
|
||||
self.d_model = d_model
|
||||
|
||||
self.self_attn = RelPositionMultiheadAttention(
|
||||
d_model, nhead, dropout=0.0
|
||||
d_model, nhead, dropout=dropout,
|
||||
)
|
||||
|
||||
self.feed_forward = nn.Sequential(
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user