diff --git a/hivemind/optim/grad_scaler.py b/hivemind/optim/grad_scaler.py index cf0410b5a..e999be91d 100644 --- a/hivemind/optim/grad_scaler.py +++ b/hivemind/optim/grad_scaler.py @@ -125,4 +125,4 @@ def _unscale_grads_( def are_grads_finite(self, optimizer: TorchOptimizer, use_cached: bool = False) -> bool: opt_dict = self._found_inf_per_device(optimizer) if use_cached else self._check_inf_per_device(optimizer) - return not sum(v.item() for v in opt_dict.values()) \ No newline at end of file + return not sum(v.item() for v in opt_dict.values())