From 2cbf5e4bf29f04601ab4917a68df61e199b3a9fc Mon Sep 17 00:00:00 2001 From: dohe0342 Date: Mon, 26 Dec 2022 13:41:13 +0900 Subject: [PATCH] from local --- .../.train.py.swp | Bin 86016 -> 86016 bytes .../train.py | 3 --- 2 files changed, 3 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/.train.py.swp index 78395208d23cc43f33b9ad7d99afa118606ccd47..2c58d569726622118241c66e0ea9f4b0074de19d 100644 GIT binary patch delta 463 zcmYk&JxD?U5Ww+!epJsAe1z1Jn$egDQPhy(P$5)XHVIKdh0vEEfrAbu)cQb&hFqeL z&5WcPf=ZecT8x&+BB-IEEy%Ecv~=);JMM0{OJvN6jM+Zz4F<+s{T{CnHj%f^gPFsu z{ajNb`mN?v!nqHvxnvE}^%?#~B025xMp2`b6Y%MkN=W~xxb&&vwp{FKbHEz?HyuuE zE@dfRYZAG{2G%hOH*9!r6xqifR^W#UBMf-Ah?H@R1p5A&Eq2!n8-XcD(RMC2Pr>x|FK> E2mZ`fHvj+t delta 465 zcmXZYK`4W97{~GFZyVmR^3rl3HbpI^6)Bv=h_%Uq9PGdjPUEnp9n_XMON;lSwj2lt z=@o}vq@*0kL7{0|NiI7$aIxZBJoTyn)AOt6-;>PPk{O$G-a9=OaP@V0g{l+TYCj0~ zo;b6*ily%LRuwTn^oCnU2h;kBzevo~3)f2X(~iTFDi%XVpBT~Wz*0F{lTJUBUDIkvreIKm-P zSV0s%O%l?^BVR3&g@Gt0(1~)D$Rkc+VgqqZp$B#da+M-i*u^Sl5JUiF8_N-g55wq3 zGrp~({K_hlK^kGSz==jRk}-iw4vq%Q1HlHm=-i(LDZ|)Xw)iRY;lcp;pVD HSGj%xX~$H~ diff --git a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py index 9be0dfce1..f5471e72c 100755 --- a/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless_d2v_v2/train.py @@ -1465,7 +1465,6 @@ def run_adapter(rank, world_size, args, wb=None): """ params = get_params() params.update(vars(args)) - #params.warm_step *= params.accum_grads fix_random_seed(params.seed) if world_size > 1: @@ -1495,8 +1494,6 @@ def run_adapter(rank, world_size, args, wb=None): logging.info("About to create model") model = get_transducer_model(params) - logging.info(model) - exit() num_param = sum([p.numel() for p in model.parameters()]) logging.info(f"Number of model parameters: {num_param}")