diff --git a/egs/librispeech/ASR/conformer_ctc2/attention.py b/egs/librispeech/ASR/conformer_ctc2/attention.py index b5929ecf6..1375d7245 100644 --- a/egs/librispeech/ASR/conformer_ctc2/attention.py +++ b/egs/librispeech/ASR/conformer_ctc2/attention.py @@ -183,13 +183,19 @@ class MultiheadAttention(nn.Module): if not self._qkv_same_embed_dim: q_proj_weight = ( - self.q_proj_weight.get_weight() if self.q_proj_weight is not None else None + self.q_proj_weight.get_weight() + if self.q_proj_weight is not None + else None ) k_proj_weight = ( - self.k_proj_weight.get_weight() if self.k_proj_weight is not None else None + self.k_proj_weight.get_weight() + if self.k_proj_weight is not None + else None ) v_proj_weight = ( - self.v_proj_weight.get_weight() if self.v_proj_weight is not None else None + self.v_proj_weight.get_weight() + if self.v_proj_weight is not None + else None ) ( attn_output,