diff --git a/egs/librispeech/ASR/.distillation_with_hubert.sh.swp b/egs/librispeech/ASR/.distillation_with_hubert.sh.swp index f18e25dd2..71895ea43 100644 Binary files a/egs/librispeech/ASR/.distillation_with_hubert.sh.swp and b/egs/librispeech/ASR/.distillation_with_hubert.sh.swp differ diff --git a/egs/librispeech/ASR/distillation_with_hubert.sh b/egs/librispeech/ASR/distillation_with_hubert.sh index 5075cc9e7..3e1163d85 100755 --- a/egs/librispeech/ASR/distillation_with_hubert.sh +++ b/egs/librispeech/ASR/distillation_with_hubert.sh @@ -50,7 +50,7 @@ full_libri=False # "True" -> stage 0 and stage 1 would be skipped, # and directly download the extracted codebook indexes for distillation # "False" -> start from scratch -use_extracted_codebook=False +use_extracted_codebook=True # teacher_model_id can be one of # "hubert_xtralarge_ll60k_finetune_ls960" -> fine-tuned model, it is the one we currently use.