diff --git a/egs/librispeech/ASR/.distillation_with_hubert.sh.swp b/egs/librispeech/ASR/.distillation_with_hubert.sh.swp index 57bd8a209..e3c0229ae 100644 Binary files a/egs/librispeech/ASR/.distillation_with_hubert.sh.swp and b/egs/librispeech/ASR/.distillation_with_hubert.sh.swp differ diff --git a/egs/librispeech/ASR/distillation_with_hubert.sh b/egs/librispeech/ASR/distillation_with_hubert.sh index 56f86ab11..d7d599161 100755 --- a/egs/librispeech/ASR/distillation_with_hubert.sh +++ b/egs/librispeech/ASR/distillation_with_hubert.sh @@ -50,7 +50,7 @@ full_libri=False # "True" -> stage 0 and stage 1 would be skipped, # and directly download the extracted codebook indexes for distillation # "False" -> start from scratch -use_extracted_codebook=True +use_extracted_codebook=False # teacher_model_id can be one of # "hubert_xtralarge_ll60k_finetune_ls960" -> fine-tuned model, it is the one we currently use.