From 3913ea7616f6a20d2584e85d1c222f9ac75c91f4 Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Wed, 26 Apr 2023 17:42:35 +0900 Subject: [PATCH] from local --- egs/librispeech/ASR/.full_ft_all.sh.swp | Bin 12288 -> 12288 bytes .../.data2vec_encoder.py.swp | Bin 16384 -> 16384 bytes .../data2vec_encoder.py | 6 +++--- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/egs/librispeech/ASR/.full_ft_all.sh.swp b/egs/librispeech/ASR/.full_ft_all.sh.swp index 9d949bbc03a0b4c852acef77ad47ad8328ead908..9f59ff4e283d704958a99a7eb5ca7d60bd253560 100644 GIT binary patch delta 32 mcmZojXh;xGG6?hZRj|}EU;qLE1_qu-9w}n8S8Wu3sSf~%kO=Ys delta 32 mcmZojXh;xGG6?hZRj|}EU;qLE28Ow1?kPufS8No2sSf~*t_f5C diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_encoder.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.data2vec_encoder.py.swp index b005b0ac63e9e0b88662e0ab9bd2f5ad7273ebd4..9db3d6a012af42ea8472a744c016c9501b02aa5e 100644 GIT binary patch delta 245 zcmWN@F-rnr0LS6~d3SlurALC^OT)bleFMqHCi@267ii1Tp|@5ixv8mDEJ(ObH{FA# z;3jX=r9luPg8B^)JkEK}d9%QTR#IFXn#8;BMACO%`oILwct8&q*d0jkc)~T>_zk5m ztg*lxLu9x?7cHD1!cQPoSm70qxW#3lE1i=Dw!XAMi8qXpp^pSd*w>^NOfkS6Vl->x XD)o&q#qsGf%s;Ko|7=>MXJz~kj5jEr delta 246 zcmWN@F-rnr0LS6~d)-^QMI3h^ibofB*EbMr7xWu=U%;)#rtU`93YuD@0*BxR9m7>{ zl>&nXT^xc2!_D9Dz*7{aD9pPX+A~RBSq+;eZ&al1ku=9Mo^Xo`wDEf=&G3jGYS>w6 zgC!;y;|3Y}Xd!?Pi+@-8#s}VzV}MIn9i5XCtPi9W78qfO3|-W5j6X+u#S5-+1rNS+ Z_Y;@F>C~F&;WJ50bf3m1D&zDk*aKv-Cl>$! diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_encoder.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_encoder.py index 1470360fe..c32d5f512 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_encoder.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/data2vec_encoder.py @@ -66,9 +66,9 @@ class FairSeqData2VecEncoder(EncoderInterface): model = models[0] model.feature_grad_mult = 0.0 ## for conv network freeze ## prevent overfitting - #model.mask_prob = 0.65 - #model.mask_channel_prob = 0.5 - #model.mask_channel_length = 64 + model.mask_prob = 0.65 + model.mask_channel_prob = 0.5 + model.mask_channel_length = 64 #model.activation_dropout = 0.1 self.encoders = model