diff --git a/egs/librispeech/WSASR/README.md b/egs/librispeech/WSASR/README.md index 6fe973028..db8e77d1c 100644 --- a/egs/librispeech/WSASR/README.md +++ b/egs/librispeech/WSASR/README.md @@ -11,15 +11,15 @@ We propose BTC/OTC to directly train an ASR system leveraging weak supervision,
Image 1 -
Substitution error
+
Image 2 -
Insertion error
+
Image 3 -
Deletion error
+
Examples of error in the transcript. The grey box is the verbatim transcript and the red box is the inaccurate transcript. Inaccurate words are marked in bold.

@@ -27,11 +27,10 @@ We propose BTC/OTC to directly train an ASR system leveraging weak supervision, This is achieved by using a special token $\star$ to model uncertainties (i.e., substitution errors, insertion errors, and deletion errors) within the WFST framework during training.\ we modify $G(\mathbf{y})$ by adding self-loop arcs into each state and bypass arcs into each arc. -
-
+

Image Alt Text -

OTC WFST representations of the transcript "a b"
-
+ +

After composing the modified WFST $G_{\text{otc}}(\mathbf{y})$ with $L$ and $T$, the OTC training graph is shown in this figure: @@ -43,7 +42,7 @@ After composing the modified WFST $G_{\text{otc}}(\mathbf{y})$ with $L$ and $T$, The $\star$ is represented as the average probability of all non-blank tokens.

- Image Alt Text + Image Alt Text OTC emission WFST