diff --git a/egs/librispeech/WSASR/conformer_ctc2/train.py b/egs/librispeech/WSASR/conformer_ctc2/train.py index a45fc2c47..58108d1b9 100755 --- a/egs/librispeech/WSASR/conformer_ctc2/train.py +++ b/egs/librispeech/WSASR/conformer_ctc2/train.py @@ -31,7 +31,7 @@ export CUDA_VISIBLE_DEVICES="0,1,2,3" --exp-dir conformer_ctc2/exp \ --lang-dir data/lang_bpe_200 \ --otc-token "" \ - --num-features 80 \ + --num-features 768 \ --allow-bypass-arc true \ --allow-self-loop-arc true \ --initial-bypass-weight -19 \ @@ -383,8 +383,8 @@ def get_params() -> AttributeDict: - warm_step: The warm_step for Noam optimizer. """ parser = get_parser() + LibriSpeechAsrDataModule.add_arguments(parser) args = parser.parse_args() - feature_dim = args.num_features params = AttributeDict( { "best_train_loss": float("inf"), @@ -397,7 +397,7 @@ def get_params() -> AttributeDict: "valid_interval": 800, # For the 100h subset, use 800 "alignment_interval": 25, # parameters for conformer - "feature_dim": feature_dim, + "feature_dim": args.num_features, "subsampling_factor": 2, "encoder_dim": 512, "nhead": 8,