diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py b/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py index 1401b623a..3ae7e8d61 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/conformer.py @@ -405,7 +405,7 @@ class RelPositionalEncoding(torch.nn.Module): class RelPositionMultiheadAttention(nn.Module): - r"""Multi-Head Attention layer with relative position encoding + r"""Multi-Head Attention layer with simplified relative position encoding See reference: "Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context"