Update optim.py (#1292)

This commit is contained in:
zr_jin 2023-10-09 17:39:23 +08:00 committed by GitHub
parent ce08230ade
commit fefffc02f6
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

View File

@ -491,6 +491,12 @@ class ScaledAdam(BatchedOptimizer):
if self.show_dominant_parameters:
assert p.shape[0] == len(param_names)
self._show_gradient_dominating_parameter(tuples, tot_sumsq)
if ans != ans: # e.g. ans is nan
ans = 0.0
if ans == 0.0:
for p, state, param_names in tuples:
p.grad.zero_() # get rid of infinity()
return ans
def _show_gradient_dominating_parameter(
@ -573,7 +579,7 @@ class ScaledAdam(BatchedOptimizer):
grad = p.grad
if clipping_scale != 1.0:
grad = grad * clipping_scale
grad *= clipping_scale
step = state["step"]
delta = state["delta"]