diff --git a/egs/librispeech/WSASR/README.md b/egs/librispeech/WSASR/README.md
index 6fe973028..db8e77d1c 100644
--- a/egs/librispeech/WSASR/README.md
+++ b/egs/librispeech/WSASR/README.md
@@ -11,15 +11,15 @@ We propose BTC/OTC to directly train an ASR system leveraging weak supervision,
Examples of error in the transcript. The grey box is the verbatim transcript and the red box is the inaccurate transcript. Inaccurate words are marked in bold.
@@ -27,11 +27,10 @@ We propose BTC/OTC to directly train an ASR system leveraging weak supervision,
This is achieved by using a special token $\star$ to model uncertainties (i.e., substitution errors, insertion errors, and deletion errors)
within the WFST framework during training.\
we modify $G(\mathbf{y})$ by adding self-loop arcs into each state and bypass arcs into each arc.
-
-
+
- OTC WFST representations of the transcript "a b"
-
+
+
After composing the modified WFST $G_{\text{otc}}(\mathbf{y})$ with $L$ and $T$, the OTC training graph is shown in this figure:
@@ -43,7 +42,7 @@ After composing the modified WFST $G_{\text{otc}}(\mathbf{y})$ with $L$ and $T$,
The $\star$ is represented as the average probability of all non-blank tokens.