export PYTHONPATH=$PYTHONPATH:/mnt/samsung-t7/yuekai/asr/icefall_llm # pip install k2==1.24.3.dev20230524+cuda11.8.torch2.0.1 -f https://k2-fsa.github.io/k2/cuda.html # pip install -r whisper/requirements.txt export CUDA_VISIBLE_DEVICES=0,1 torchrun --nproc_per_node 2 ./whisper_llm_zh/train.py \ --max-duration 20 \ --exp-dir ./whisper_llm_zh/exp_test \ --speech-encoder-path-or-name tiny \ --llm-path-or-name Qwen/Qwen1.5-0.5B-Chat \ --manifest-dir data/fbank \ --deepspeed \ --deepspeed_config ./whisper_llm_zh/ds_config_zero1.json \ --use-flash-attn False # python3 ./whisper_llm_zh/decode.py \ # --max-duration 80 \ # --exp-dir ./whisper_llm_zh/exp_qwen_0.5B \ # --speech-encoder-path-or-name /mnt/samsung-t7/yuekai/asr/v1.1/whisper-large-v2-multi-hans-zh-epoch-3-avg-10.pt \ # --llm-path-or-name Qwen/Qwen1.5-0.5B-Chat \ # --epoch 1 --avg 1 \ # --manifest-dir data/fbank \ # --use-flash-attn False