From 75e872ea574eddd425709cb00f6bb993b94668e3 Mon Sep 17 00:00:00 2001 From: Daniel Povey Date: Fri, 8 Jul 2022 09:13:54 +0800 Subject: [PATCH] Fix bug in getting denom in proj update --- egs/librispeech/ASR/pruned_transducer_stateless7/optim.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/egs/librispeech/ASR/pruned_transducer_stateless7/optim.py b/egs/librispeech/ASR/pruned_transducer_stateless7/optim.py index 06812b44e..d8613ccce 100644 --- a/egs/librispeech/ASR/pruned_transducer_stateless7/optim.py +++ b/egs/librispeech/ASR/pruned_transducer_stateless7/optim.py @@ -109,7 +109,6 @@ class LearnedGradient(Optimizer): for group in self.param_groups: lr = group["lr"] - meta_lr_scale = group["meta_lr_scale"] size_lr = lr * group["size_lr_scale"] beta1, beta2 = group["betas"] lr_mat_min = group["lr_mat_min"] @@ -454,7 +453,7 @@ class LearnedGradient(Optimizer): # for now just estimate proj2_grad_var from proj2_grad; later, we # may try aggregating it across iterations. This is like exp_avg_sq # in Adam. - proj2_grad_var = self._get_matrix_var(proj_grad, group["eps"]) + proj2_grad_var = self._get_matrix_var(proj2_grad, group["eps"]) denom = proj2_grad_var.sqrt()