From 91c579ad1e430efd1985d4e17c75b1fbefab5851 Mon Sep 17 00:00:00 2001 From: jiangchangting1 Date: Tue, 14 Nov 2023 03:08:24 +0000 Subject: [PATCH 01/10] update debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py. Signed-off-by: jiangchangting1 --- .../src/python/ptdbg_ascend/hook_module/api_registry.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py index f4f21ea899..f16861834a 100644 --- a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py +++ b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py @@ -122,7 +122,8 @@ class ApiRegistry: if attr_name.startswith("wrap_"): self.distributed_hook_attr[attr_name[5:]] = getattr(wrap_distributed.HOOKDistributedOP, attr_name) if not is_gpu and not torch_without_guard_version: - self.store_ori_attr(torch_npu.distributed, get_distributed_ops(), self.npu_distributed_ori_attr) + self.store_ori_attr(torch_npu.distributed, set(dir(torch_npu.distributed)) & get_distributed_ops(), + self.npu_distributed_ori_attr) self.npu_distributed_hook_attr[attr_name[5:]] = getattr(wrap_distributed.HOOKDistributedOP, attr_name) -- Gitee From f059d160822a0293a25a8504e7b93a8b7875a5e9 Mon Sep 17 00:00:00 2001 From: jiangchangting1 Date: Tue, 14 Nov 2023 08:33:30 +0000 Subject: [PATCH 02/10] update debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py. Signed-off-by: jiangchangting1 --- .../src/python/ptdbg_ascend/hook_module/api_registry.py | 7 +++---- 1 file changed, 3 insertions(+), 4 deletions(-) diff --git a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py index f16861834a..4155581d12 100644 --- a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py +++ b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py @@ -24,7 +24,7 @@ from .wrap_tensor import get_tensor_ops from .wrap_vf import get_vf_ops from .wrap_distributed import get_distributed_ops from .wrap_aten import get_aten_ops -from ..common.utils import torch_without_guard_version +from ..common.utils import torch_without_guard_version, npu_distributed_api torch_version_above_2 = torch.__version__.split('+')[0] > '2.0' try: @@ -121,9 +121,8 @@ class ApiRegistry: for attr_name in dir(wrap_distributed.HOOKDistributedOP): if attr_name.startswith("wrap_"): self.distributed_hook_attr[attr_name[5:]] = getattr(wrap_distributed.HOOKDistributedOP, attr_name) - if not is_gpu and not torch_without_guard_version: - self.store_ori_attr(torch_npu.distributed, set(dir(torch_npu.distributed)) & get_distributed_ops(), - self.npu_distributed_ori_attr) + if not is_gpu and not torch_without_guard_version and attr_name in npu_distributed_api: + self.store_ori_attr(torch_npu.distributed, get_distributed_ops(), self.npu_distributed_ori_attr) self.npu_distributed_hook_attr[attr_name[5:]] = getattr(wrap_distributed.HOOKDistributedOP, attr_name) -- Gitee From f9b78e28e561a57bd8f019b29d37e34f8eb30f09 Mon Sep 17 00:00:00 2001 From: jiangchangting1 Date: Tue, 14 Nov 2023 08:37:40 +0000 Subject: [PATCH 03/10] update debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py. Signed-off-by: jiangchangting1 --- .../ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py index 6e63b7cf88..b74c3edd2a 100644 --- a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py +++ b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py @@ -51,7 +51,7 @@ if not is_gpu and not torch_without_guard_version: device = collections.namedtuple('device', ['type', 'index']) prefixes = ['api_stack', 'list', 'range', 'acl'] - +npu_distributed_api = ['wrap_isend', 'wrap_irecv'] class Const: """ -- Gitee From 7440780f7c8c518828b1d963fce19bf62ba8305d Mon Sep 17 00:00:00 2001 From: jiangchangting1 Date: Tue, 14 Nov 2023 11:13:17 +0000 Subject: [PATCH 04/10] update debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py. Signed-off-by: jiangchangting1 --- .../ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py index b74c3edd2a..d1a07c43ac 100644 --- a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py +++ b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py @@ -51,7 +51,7 @@ if not is_gpu and not torch_without_guard_version: device = collections.namedtuple('device', ['type', 'index']) prefixes = ['api_stack', 'list', 'range', 'acl'] -npu_distributed_api = ['wrap_isend', 'wrap_irecv'] +npu_distributed_api = ['isend', 'irecv'] class Const: """ -- Gitee From fe3b9e1bca5aa1c8396a95cc644743aab33b4a72 Mon Sep 17 00:00:00 2001 From: jiangchangting1 Date: Tue, 14 Nov 2023 11:16:08 +0000 Subject: [PATCH 05/10] update debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py. Signed-off-by: jiangchangting1 --- .../src/python/ptdbg_ascend/hook_module/api_registry.py | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py index 4155581d12..33fb4db6d2 100644 --- a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py +++ b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py @@ -24,6 +24,7 @@ from .wrap_tensor import get_tensor_ops from .wrap_vf import get_vf_ops from .wrap_distributed import get_distributed_ops from .wrap_aten import get_aten_ops +from .wrap_npu_custom import get_npu_ops from ..common.utils import torch_without_guard_version, npu_distributed_api torch_version_above_2 = torch.__version__.split('+')[0] > '2.0' @@ -121,8 +122,8 @@ class ApiRegistry: for attr_name in dir(wrap_distributed.HOOKDistributedOP): if attr_name.startswith("wrap_"): self.distributed_hook_attr[attr_name[5:]] = getattr(wrap_distributed.HOOKDistributedOP, attr_name) - if not is_gpu and not torch_without_guard_version and attr_name in npu_distributed_api: - self.store_ori_attr(torch_npu.distributed, get_distributed_ops(), self.npu_distributed_ori_attr) + if not is_gpu and not torch_without_guard_version and attr_name[5:] in npu_distributed_api: + self.store_ori_attr(torch_npu.distributed, npu_distributed_api, self.npu_distributed_ori_attr) self.npu_distributed_hook_attr[attr_name[5:]] = getattr(wrap_distributed.HOOKDistributedOP, attr_name) @@ -140,7 +141,7 @@ class ApiRegistry: self.vf_hook_attr[attr_name[5:]] = getattr(wrap_vf.HOOKVfOP, attr_name) if not is_gpu: - self.store_ori_attr(torch_npu, WrapNpuOps, self.torch_npu_ori_attr) + self.store_ori_attr(torch_npu, get_npu_ops, self.torch_npu_ori_attr) wrap_npu_custom.wrap_npu_ops_and_bind(hook) for attr_name in dir(wrap_npu_custom.HOOKNpuOP): if attr_name.startswith("wrap_"): -- Gitee From 92185f28e99874f32d7867770dee394528a7cd30 Mon Sep 17 00:00:00 2001 From: jiangchangting1 Date: Tue, 14 Nov 2023 11:27:57 +0000 Subject: [PATCH 06/10] update debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py. Signed-off-by: jiangchangting1 --- .../src/python/ptdbg_ascend/hook_module/api_registry.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py index 33fb4db6d2..df643ac303 100644 --- a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py +++ b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py @@ -141,7 +141,7 @@ class ApiRegistry: self.vf_hook_attr[attr_name[5:]] = getattr(wrap_vf.HOOKVfOP, attr_name) if not is_gpu: - self.store_ori_attr(torch_npu, get_npu_ops, self.torch_npu_ori_attr) + self.store_ori_attr(torch_npu, get_npu_ops(), self.torch_npu_ori_attr) wrap_npu_custom.wrap_npu_ops_and_bind(hook) for attr_name in dir(wrap_npu_custom.HOOKNpuOP): if attr_name.startswith("wrap_"): -- Gitee From f3cacd5b397d9832d6fd60dde9e5142a3be46a67 Mon Sep 17 00:00:00 2001 From: jiangchangting1 Date: Tue, 14 Nov 2023 11:32:58 +0000 Subject: [PATCH 07/10] update debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py. Signed-off-by: jiangchangting1 --- .../ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py | 1 + 1 file changed, 1 insertion(+) diff --git a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py index d1a07c43ac..21cae3c52a 100644 --- a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py +++ b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/common/utils.py @@ -53,6 +53,7 @@ device = collections.namedtuple('device', ['type', 'index']) prefixes = ['api_stack', 'list', 'range', 'acl'] npu_distributed_api = ['isend', 'irecv'] + class Const: """ Class for const -- Gitee From d4d42438aff4fd8ab8314ad041b8bf280a7ff60a Mon Sep 17 00:00:00 2001 From: jiangchangting1 Date: Tue, 14 Nov 2023 11:40:23 +0000 Subject: [PATCH 08/10] update debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py. Signed-off-by: jiangchangting1 --- .../src/python/ptdbg_ascend/hook_module/api_registry.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py index df643ac303..3d4d79e550 100644 --- a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py +++ b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py @@ -24,7 +24,6 @@ from .wrap_tensor import get_tensor_ops from .wrap_vf import get_vf_ops from .wrap_distributed import get_distributed_ops from .wrap_aten import get_aten_ops -from .wrap_npu_custom import get_npu_ops from ..common.utils import torch_without_guard_version, npu_distributed_api torch_version_above_2 = torch.__version__.split('+')[0] > '2.0' @@ -35,7 +34,7 @@ except ImportError: else: is_gpu = False from . import wrap_npu_custom - from .wrap_npu_custom import WrapNpuOps + from .wrap_npu_custom import WrapNpuOps, get_npu_ops class ApiRegistry: -- Gitee From 4109aa60052724b0d40731bb6953dcbc67fba651 Mon Sep 17 00:00:00 2001 From: jiangchangting1 Date: Tue, 14 Nov 2023 11:45:12 +0000 Subject: [PATCH 09/10] update debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py. Signed-off-by: jiangchangting1 --- .../src/python/ptdbg_ascend/hook_module/api_registry.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py index 3d4d79e550..7639a7d58e 100644 --- a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py +++ b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py @@ -34,7 +34,7 @@ except ImportError: else: is_gpu = False from . import wrap_npu_custom - from .wrap_npu_custom import WrapNpuOps, get_npu_ops + from .wrap_npu_custom import get_npu_ops class ApiRegistry: -- Gitee From 771ceba6866be7e08ad18775464ef4061ae05798 Mon Sep 17 00:00:00 2001 From: jiangchangting1 Date: Tue, 14 Nov 2023 12:21:11 +0000 Subject: [PATCH 10/10] update debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py. Signed-off-by: jiangchangting1 --- .../src/python/ptdbg_ascend/hook_module/api_registry.py | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py index 7639a7d58e..064103aead 100644 --- a/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py +++ b/debug/accuracy_tools/ptdbg_ascend/src/python/ptdbg_ascend/hook_module/api_registry.py @@ -118,11 +118,12 @@ class ApiRegistry: self.store_ori_attr(dist, get_distributed_ops(), self.distributed_ori_attr) wrap_distributed.wrap_distributed_ops_and_bind(hook) + if not is_gpu and not torch_without_guard_version: + self.store_ori_attr(torch_npu.distributed, npu_distributed_api, self.npu_distributed_ori_attr) for attr_name in dir(wrap_distributed.HOOKDistributedOP): if attr_name.startswith("wrap_"): self.distributed_hook_attr[attr_name[5:]] = getattr(wrap_distributed.HOOKDistributedOP, attr_name) - if not is_gpu and not torch_without_guard_version and attr_name[5:] in npu_distributed_api: - self.store_ori_attr(torch_npu.distributed, npu_distributed_api, self.npu_distributed_ori_attr) + if not is_gpu and not torch_without_guard_version and attr_name[5:] in npu_distributed_api: self.npu_distributed_hook_attr[attr_name[5:]] = getattr(wrap_distributed.HOOKDistributedOP, attr_name) -- Gitee