From d28fb67eb7b12420342502add4e2cb8b4db3d423 Mon Sep 17 00:00:00 2001 From: jiangchangting1 Date: Fri, 10 Nov 2023 03:41:06 +0000 Subject: [PATCH] update debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/wrap_npu_custom.py. Signed-off-by: jiangchangting1 --- .../ptdbg_ascend/hook_module/wrap_npu_custom.py | 11 ++++++++++- 1 file changed, 10 insertions(+), 1 deletion(-) diff --git a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/wrap_npu_custom.py b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/wrap_npu_custom.py index f2e1e8f9d..2b90a8f5b 100644 --- a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/wrap_npu_custom.py +++ b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/wrap_npu_custom.py @@ -30,6 +30,15 @@ with FileOpen(yaml_path, 'r') as f: WrapNpuOps = yaml.safe_load(f).get('torch_npu') +def get_npu_ops(): + global WrapNpuOps + if torch_without_guard_version: + _npu_ops = dir(torch.ops.npu) + else: + _npu_ops = dir(torch_npu._C._VariableFunctionsClass) + return set(WrapNpuOps) & set(_npu_ops) + + class HOOKNpuOP(object): pass @@ -58,6 +67,6 @@ def wrap_npu_op(op_name, hook): def wrap_npu_ops_and_bind(hook): - _npu_ops = WrapNpuOps + _npu_ops = get_npu_ops() for op_name in _npu_ops: setattr(HOOKNpuOP, "wrap_" + str(op_name), wrap_npu_op(op_name, hook)) -- Gitee