From 3f9ce87578dc953374b5e5cf9e0ff0b12ab1db21 Mon Sep 17 00:00:00 2001 From: yhna940 Date: Wed, 5 Apr 2023 00:37:54 +0900 Subject: [PATCH] Fix typo --- colossalai/tensor/param_op_hook.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/colossalai/tensor/param_op_hook.py b/colossalai/tensor/param_op_hook.py index ed705da0eb0d..9c2e0d4adbf1 100644 --- a/colossalai/tensor/param_op_hook.py +++ b/colossalai/tensor/param_op_hook.py @@ -168,12 +168,12 @@ def _get_grad_args(*args): # if there is no grad tensor, the backward of PreFwdPostBwd can't be triggered arg_zero = args[0] if not isinstance(arg_zero, tuple): - raise NotImplementedError("Some torch function is incompatible because of its complcated inputs.") + raise NotImplementedError("Some torch function is incompatible because of its complicated inputs.") check_grad_flag = False for obj in arg_zero: check_grad_flag |= _is_grad_tensor(obj) if not check_grad_flag: - raise NotImplementedError("Some torch function is incompatible because of its complcated inputs.") + raise NotImplementedError("Some torch function is incompatible because of its complicated inputs.") return arg_zero, args[1:]