From ca3028cdbcb4d7f174898ca5738c7eb682d28d18 Mon Sep 17 00:00:00 2001 From: 13611516375 Date: Thu, 27 Mar 2025 02:17:49 +0000 Subject: [PATCH 01/26] =?UTF-8?q?=E6=8F=90=E4=BA=A4Atb=E9=80=82=E9=85=8D?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .../AtbAddSample/AddOp/CMakeLists.txt | 69 ++ .../AtbAddSample/AddOp/CMakePresets.json | 63 ++ atb_operator/AtbAddSample/AddOp/README.md | 199 +++++ atb_operator/AtbAddSample/AddOp/build.sh | 37 + .../AtbAddSample/AddOp/cmake/config.cmake | 25 + .../AtbAddSample/AddOp/cmake/func.cmake | 228 +++++ .../AtbAddSample/AddOp/cmake/intf.cmake | 26 + .../AtbAddSample/AddOp/cmake/makeself.cmake | 17 + .../AtbAddSample/AddOp/cmake/util/__init__.py | 8 + .../cmake/util/ascendc_bin_param_build.py | 129 +++ .../AddOp/cmake/util/ascendc_impl_build.py | 446 ++++++++++ .../AddOp/cmake/util/ascendc_ops_config.py | 114 +++ .../AddOp/cmake/util/ascendc_replay_build.py | 65 ++ .../AddOp/cmake/util/batch_replay_impl.temp | 117 +++ .../AddOp/cmake/util/code_channel_infer.py | 115 +++ .../AddOp/cmake/util/const_var.py | 33 + .../cmake/util/gen_impl_and_mrege_json.sh | 57 ++ .../AddOp/cmake/util/gen_ops_filter.sh | 62 ++ .../AddOp/cmake/util/gen_version_info.sh | 6 + .../AddOp/cmake/util/insert_op_info.py | 36 + .../cmake/util/insert_simplified_keys.py | 248 ++++++ .../AddOp/cmake/util/kernel_entry.py | 115 +++ .../AddOp/cmake/util/kernel_impl.temp | 10 + .../AddOp/cmake/util/makeself/COPYING | 339 ++++++++ .../AddOp/cmake/util/makeself/README.md | 246 ++++++ .../AddOp/cmake/util/makeself/VERSION | 1 + .../AddOp/cmake/util/makeself/make-release.sh | 9 + .../cmake/util/makeself/makeself-header.sh | 660 ++++++++++++++ .../AddOp/cmake/util/makeself/makeself.1 | 110 +++ .../AddOp/cmake/util/makeself/makeself.lsm | 16 + .../AddOp/cmake/util/makeself/makeself.sh | 822 ++++++++++++++++++ .../AddOp/cmake/util/makeself/run-tests.sh | 8 + .../AddOp/cmake/util/merge_aicpu_info_json.sh | 31 + .../AddOp/cmake/util/opdesc_parser.py | 260 ++++++ .../AddOp/cmake/util/parse_ini_to_json.py | 338 +++++++ .../AddOp/cmake/util/preset_parse.py | 23 + .../AddOp/cmake/util/replay_codegen.py | 105 +++ .../AddOp/cmake/util/replay_impl.temp | 120 +++ .../AddOp/cmake/util/tiling_data_def_build.py | 84 ++ .../AddOp/framework/CMakeLists.txt | 11 + .../AddOp/framework/tf_plugin/CMakeLists.txt | 14 + .../tf_plugin/tensorflow_add_custom_plugin.cc | 23 + .../AtbAddSample/AddOp/op_host/CMakeLists.txt | 82 ++ .../AtbAddSample/AddOp/op_host/add_custom.cpp | 117 +++ .../AddOp/op_host/add_custom_tiling.h | 22 + .../AddOp/op_kernel/CMakeLists.txt | 61 ++ .../AddOp/op_kernel/add_custom.cpp | 134 +++ .../AtbAddSample/AddOp/scripts/install.sh | 318 +++++++ .../AtbAddSample/AddOp/scripts/upgrade.sh | 151 ++++ atb_operator/AtbAddSample/readme.md | 0 .../AtbEyeSample/EyeOp/CMakeLists.txt | 69 ++ .../AtbEyeSample/EyeOp/CMakePresets.json | 63 ++ atb_operator/AtbEyeSample/EyeOp/README.md | 216 +++++ atb_operator/AtbEyeSample/EyeOp/build.sh | 80 ++ .../AtbEyeSample/EyeOp/cmake/config.cmake | 25 + .../AtbEyeSample/EyeOp/cmake/func.cmake | 192 ++++ .../AtbEyeSample/EyeOp/cmake/intf.cmake | 26 + .../AtbEyeSample/EyeOp/cmake/makeself.cmake | 17 + .../AtbEyeSample/EyeOp/cmake/util/__init__.py | 8 + .../cmake/util/ascendc_bin_param_build.py | 127 +++ .../EyeOp/cmake/util/ascendc_impl_build.py | 501 +++++++++++ .../EyeOp/cmake/util/ascendc_ops_config.py | 120 +++ .../EyeOp/cmake/util/ascendc_replay_build.py | 65 ++ .../EyeOp/cmake/util/batch_replay_impl.temp | 117 +++ .../EyeOp/cmake/util/code_channel_infer.py | 58 ++ .../EyeOp/cmake/util/const_var.py | 33 + .../cmake/util/gen_impl_and_mrege_json.sh | 20 + .../EyeOp/cmake/util/gen_ops_filter.sh | 62 ++ .../EyeOp/cmake/util/gen_version_info.sh | 6 + .../EyeOp/cmake/util/insert_op_info.py | 36 + .../cmake/util/insert_simplified_keys.py | 242 ++++++ .../EyeOp/cmake/util/kernel_entry.py | 115 +++ .../EyeOp/cmake/util/kernel_impl.temp | 10 + .../EyeOp/cmake/util/makeself/COPYING | 339 ++++++++ .../EyeOp/cmake/util/makeself/README.md | 246 ++++++ .../EyeOp/cmake/util/makeself/VERSION | 1 + .../EyeOp/cmake/util/makeself/make-release.sh | 9 + .../cmake/util/makeself/makeself-header.sh | 660 ++++++++++++++ .../EyeOp/cmake/util/makeself/makeself.1 | 110 +++ .../EyeOp/cmake/util/makeself/makeself.lsm | 16 + .../EyeOp/cmake/util/makeself/makeself.sh | 822 ++++++++++++++++++ .../EyeOp/cmake/util/makeself/run-tests.sh | 8 + .../EyeOp/cmake/util/merge_aicpu_info_json.sh | 31 + .../EyeOp/cmake/util/opdesc_parser.py | 298 +++++++ .../EyeOp/cmake/util/parse_ini_to_json.py | 338 +++++++ .../EyeOp/cmake/util/preset_parse.py | 23 + .../EyeOp/cmake/util/replay_codegen.py | 105 +++ .../EyeOp/cmake/util/replay_impl.temp | 120 +++ .../EyeOp/cmake/util/tiling_data_def_build.py | 87 ++ .../EyeOp/framework/CMakeLists.txt | 11 + .../EyeOp/framework/tf_plugin/CMakeLists.txt | 14 + .../tf_plugin/tensorflow_eye_plugin.cc | 12 + .../AtbEyeSample/EyeOp/op_host/CMakeLists.txt | 82 ++ .../AtbEyeSample/EyeOp/op_host/eye.cpp | 173 ++++ .../AtbEyeSample/EyeOp/op_host/eye_tiling.h | 31 + .../EyeOp/op_kernel/CMakeLists.txt | 68 ++ .../AtbEyeSample/EyeOp/op_kernel/eye.cpp | 91 ++ .../AtbEyeSample/EyeOp/scripts/install.sh | 318 +++++++ .../AtbEyeSample/EyeOp/scripts/upgrade.sh | 151 ++++ .../AtbTrilSample/TrilOp/CMakeLists.txt | 69 ++ .../AtbTrilSample/TrilOp/CMakePresets.json | 63 ++ atb_operator/AtbTrilSample/TrilOp/README.md | 193 ++++ atb_operator/AtbTrilSample/TrilOp/build.sh | 74 ++ .../AtbTrilSample/TrilOp/cmake/config.cmake | 25 + .../AtbTrilSample/TrilOp/cmake/func.cmake | 192 ++++ .../AtbTrilSample/TrilOp/cmake/intf.cmake | 26 + .../AtbTrilSample/TrilOp/cmake/makeself.cmake | 17 + .../TrilOp/cmake/util/__init__.py | 8 + .../cmake/util/ascendc_bin_param_build.py | 127 +++ .../TrilOp/cmake/util/ascendc_impl_build.py | 496 +++++++++++ .../TrilOp/cmake/util/ascendc_ops_config.py | 114 +++ .../TrilOp/cmake/util/ascendc_replay_build.py | 65 ++ .../TrilOp/cmake/util/batch_replay_impl.temp | 117 +++ .../TrilOp/cmake/util/code_channel_infer.py | 58 ++ .../TrilOp/cmake/util/const_var.py | 33 + .../cmake/util/gen_impl_and_mrege_json.sh | 20 + .../TrilOp/cmake/util/gen_ops_filter.sh | 62 ++ .../TrilOp/cmake/util/gen_version_info.sh | 6 + .../TrilOp/cmake/util/insert_op_info.py | 36 + .../cmake/util/insert_simplified_keys.py | 242 ++++++ .../TrilOp/cmake/util/kernel_entry.py | 115 +++ .../TrilOp/cmake/util/kernel_impl.temp | 10 + .../TrilOp/cmake/util/makeself/COPYING | 339 ++++++++ .../TrilOp/cmake/util/makeself/README.md | 246 ++++++ .../TrilOp/cmake/util/makeself/VERSION | 1 + .../cmake/util/makeself/make-release.sh | 9 + .../cmake/util/makeself/makeself-header.sh | 660 ++++++++++++++ .../TrilOp/cmake/util/makeself/makeself.1 | 110 +++ .../TrilOp/cmake/util/makeself/makeself.lsm | 16 + .../TrilOp/cmake/util/makeself/makeself.sh | 822 ++++++++++++++++++ .../TrilOp/cmake/util/makeself/run-tests.sh | 8 + .../cmake/util/merge_aicpu_info_json.sh | 31 + .../TrilOp/cmake/util/opdesc_parser.py | 277 ++++++ .../TrilOp/cmake/util/parse_ini_to_json.py | 338 +++++++ .../TrilOp/cmake/util/preset_parse.py | 23 + .../TrilOp/cmake/util/replay_codegen.py | 105 +++ .../TrilOp/cmake/util/replay_impl.temp | 120 +++ .../cmake/util/tiling_data_def_build.py | 87 ++ .../TrilOp/framework/CMakeLists.txt | 11 + .../TrilOp/framework/tf_plugin/CMakeLists.txt | 14 + .../tf_plugin/tensorflow_tril_plugin.cc | 20 + .../TrilOp/op_host/CMakeLists.txt | 82 ++ .../AtbTrilSample/TrilOp/op_host/tril.cpp | 215 +++++ .../TrilOp/op_host/tril_tiling.h | 34 + .../TrilOp/op_kernel/CMakeLists.txt | 68 ++ .../AtbTrilSample/TrilOp/op_kernel/tril.cpp | 278 ++++++ .../AtbTrilSample/TrilOp/scripts/install.sh | 318 +++++++ .../AtbTrilSample/TrilOp/scripts/upgrade.sh | 151 ++++ atb_operator/README.md | 6 + 149 files changed, 19259 insertions(+) create mode 100644 atb_operator/AtbAddSample/AddOp/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AddOp/CMakePresets.json create mode 100644 atb_operator/AtbAddSample/AddOp/README.md create mode 100644 atb_operator/AtbAddSample/AddOp/build.sh create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/config.cmake create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/func.cmake create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/intf.cmake create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/makeself.cmake create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/__init__.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_bin_param_build.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_impl_build.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_ops_config.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_replay_build.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/batch_replay_impl.temp create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/code_channel_infer.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/const_var.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/gen_impl_and_mrege_json.sh create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/gen_ops_filter.sh create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/gen_version_info.sh create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/insert_op_info.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/insert_simplified_keys.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/kernel_entry.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/kernel_impl.temp create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/COPYING create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/README.md create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/VERSION create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/make-release.sh create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself-header.sh create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.1 create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.lsm create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.sh create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/run-tests.sh create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/merge_aicpu_info_json.sh create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/opdesc_parser.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/parse_ini_to_json.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/preset_parse.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/replay_codegen.py create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/replay_impl.temp create mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/tiling_data_def_build.py create mode 100644 atb_operator/AtbAddSample/AddOp/framework/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AddOp/framework/tf_plugin/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AddOp/framework/tf_plugin/tensorflow_add_custom_plugin.cc create mode 100644 atb_operator/AtbAddSample/AddOp/op_host/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AddOp/op_host/add_custom.cpp create mode 100644 atb_operator/AtbAddSample/AddOp/op_host/add_custom_tiling.h create mode 100644 atb_operator/AtbAddSample/AddOp/op_kernel/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AddOp/op_kernel/add_custom.cpp create mode 100644 atb_operator/AtbAddSample/AddOp/scripts/install.sh create mode 100644 atb_operator/AtbAddSample/AddOp/scripts/upgrade.sh create mode 100644 atb_operator/AtbAddSample/readme.md create mode 100755 atb_operator/AtbEyeSample/EyeOp/CMakeLists.txt create mode 100755 atb_operator/AtbEyeSample/EyeOp/CMakePresets.json create mode 100644 atb_operator/AtbEyeSample/EyeOp/README.md create mode 100755 atb_operator/AtbEyeSample/EyeOp/build.sh create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/config.cmake create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/func.cmake create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/intf.cmake create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/makeself.cmake create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/__init__.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_bin_param_build.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_impl_build.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_ops_config.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_replay_build.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/batch_replay_impl.temp create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/code_channel_infer.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/const_var.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_impl_and_mrege_json.sh create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_ops_filter.sh create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_version_info.sh create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_op_info.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_simplified_keys.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_entry.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_impl.temp create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/COPYING create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/README.md create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/VERSION create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/make-release.sh create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself-header.sh create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.1 create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.lsm create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.sh create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/run-tests.sh create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/merge_aicpu_info_json.sh create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/opdesc_parser.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/parse_ini_to_json.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/preset_parse.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_codegen.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_impl.temp create mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/tiling_data_def_build.py create mode 100755 atb_operator/AtbEyeSample/EyeOp/framework/CMakeLists.txt create mode 100755 atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/CMakeLists.txt create mode 100755 atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/tensorflow_eye_plugin.cc create mode 100755 atb_operator/AtbEyeSample/EyeOp/op_host/CMakeLists.txt create mode 100755 atb_operator/AtbEyeSample/EyeOp/op_host/eye.cpp create mode 100755 atb_operator/AtbEyeSample/EyeOp/op_host/eye_tiling.h create mode 100755 atb_operator/AtbEyeSample/EyeOp/op_kernel/CMakeLists.txt create mode 100755 atb_operator/AtbEyeSample/EyeOp/op_kernel/eye.cpp create mode 100755 atb_operator/AtbEyeSample/EyeOp/scripts/install.sh create mode 100755 atb_operator/AtbEyeSample/EyeOp/scripts/upgrade.sh create mode 100755 atb_operator/AtbTrilSample/TrilOp/CMakeLists.txt create mode 100755 atb_operator/AtbTrilSample/TrilOp/CMakePresets.json create mode 100644 atb_operator/AtbTrilSample/TrilOp/README.md create mode 100755 atb_operator/AtbTrilSample/TrilOp/build.sh create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/config.cmake create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/func.cmake create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/intf.cmake create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/makeself.cmake create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/__init__.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_bin_param_build.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_impl_build.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_ops_config.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_replay_build.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/batch_replay_impl.temp create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/code_channel_infer.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/const_var.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_impl_and_mrege_json.sh create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_ops_filter.sh create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_version_info.sh create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_op_info.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_simplified_keys.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_entry.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_impl.temp create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/COPYING create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/README.md create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/VERSION create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/make-release.sh create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself-header.sh create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.1 create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.lsm create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.sh create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/run-tests.sh create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/merge_aicpu_info_json.sh create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/opdesc_parser.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/parse_ini_to_json.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/preset_parse.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_codegen.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_impl.temp create mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/tiling_data_def_build.py create mode 100755 atb_operator/AtbTrilSample/TrilOp/framework/CMakeLists.txt create mode 100755 atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/CMakeLists.txt create mode 100755 atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/tensorflow_tril_plugin.cc create mode 100755 atb_operator/AtbTrilSample/TrilOp/op_host/CMakeLists.txt create mode 100755 atb_operator/AtbTrilSample/TrilOp/op_host/tril.cpp create mode 100755 atb_operator/AtbTrilSample/TrilOp/op_host/tril_tiling.h create mode 100755 atb_operator/AtbTrilSample/TrilOp/op_kernel/CMakeLists.txt create mode 100755 atb_operator/AtbTrilSample/TrilOp/op_kernel/tril.cpp create mode 100755 atb_operator/AtbTrilSample/TrilOp/scripts/install.sh create mode 100755 atb_operator/AtbTrilSample/TrilOp/scripts/upgrade.sh create mode 100644 atb_operator/README.md diff --git a/atb_operator/AtbAddSample/AddOp/CMakeLists.txt b/atb_operator/AtbAddSample/AddOp/CMakeLists.txt new file mode 100644 index 000000000..584132d80 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/CMakeLists.txt @@ -0,0 +1,69 @@ +cmake_minimum_required(VERSION 3.16.0) +project(opp) +if(ENABLE_CROSS_COMPILE) + if(${CMAKE_SYSTEM_PROCESSOR} STREQUAL x86_64) + set(CROSS_COMPILE_PLATFORM aarch64) + else() + set(CROSS_COMPILE_PLATFORM x86_64) + endif() + set(PLATFORM ${CMAKE_SYSTEM_PROCESSOR}) + set(CMAKE_COMPILE_COMPILER_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/linux/${CROSS_COMPILE_PLATFORM}/) + set(CMAKE_COMPILE_RUNTIME_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/${CROSS_COMPILE_PLATFORM}/) + set(CMAKE_SYSTEM_PROCESSOR ${CROSS_COMPILE_PLATFORM}) + set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) + set(CMAKE_CXX_COMPILER ${CMAKE_CROSS_PLATFORM_COMPILER}) +else() + set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) +endif() + +include(cmake/config.cmake) +include(cmake/func.cmake) +include(cmake/intf.cmake) + +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/framework) + add_subdirectory(framework) +endif() +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_host) + add_subdirectory(op_host) +endif() +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_kernel) + add_subdirectory(op_kernel) +endif() +if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) + add_subdirectory(testcases) +endif() + +# modify vendor_name in install.sh and upgrade.sh +add_custom_command(OUTPUT ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh + COMMAND mkdir -p ${CMAKE_BINARY_DIR}/scripts + COMMAND cp -r ${CMAKE_SOURCE_DIR}/scripts/* ${CMAKE_BINARY_DIR}/scripts/ + COMMAND sed -i "s/vendor_name=customize/vendor_name=${vendor_name}/g" ${CMAKE_BINARY_DIR}/scripts/* +) +add_custom_target(modify_vendor ALL DEPENDS ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh) +install(DIRECTORY ${CMAKE_BINARY_DIR}/scripts/ DESTINATION . FILE_PERMISSIONS OWNER_EXECUTE OWNER_READ GROUP_READ) + +install(FILES ${CMAKE_SOURCE_DIR}/custom.proto DESTINATION packages OPTIONAL) + +get_system_info(SYSTEM_INFO) + +# gen version.info +add_custom_target(gen_version_info ALL + COMMAND bash ${CMAKE_CURRENT_SOURCE_DIR}/cmake/util/gen_version_info.sh ${ASCEND_CANN_PACKAGE_PATH} ${CMAKE_CURRENT_BINARY_DIR} +) + +install(FILES ${CMAKE_CURRENT_BINARY_DIR}/version.info + DESTINATION packages/vendors/${vendor_name}/) + +# CPack config +set(CPACK_PACKAGE_NAME ${CMAKE_PROJECT_NAME}) +set(CPACK_PACKAGE_VERSION ${CMAKE_PROJECT_VERSION}) +set(CPACK_PACKAGE_DESCRIPTION "CPack opp project") +set(CPACK_PACKAGE_DESCRIPTION_SUMMARY "CPack opp project") +set(CPACK_PACKAGE_DIRECTORY ${CMAKE_INSTALL_PREFIX}) +set(CPACK_PACKAGE_FILE_NAME "custom_opp_${SYSTEM_INFO}.run") +set(CPACK_GENERATOR External) +set(CPACK_CMAKE_GENERATOR "Unix Makefiles") +set(CPACK_EXTERNAL_ENABLE_STAGING TRUE) +set(CPACK_EXTERNAL_PACKAGE_SCRIPT ${CMAKE_SOURCE_DIR}/cmake/makeself.cmake) +set(CPACK_EXTERNAL_BUILT_PACKAGES ${CPACK_PACKAGE_DIRECTORY}/_CPack_Packages/Linux/External/${CPACK_PACKAGE_FILE_NAME}/${CPACK_PACKAGE_FILE_NAME}) +include(CPack) diff --git a/atb_operator/AtbAddSample/AddOp/CMakePresets.json b/atb_operator/AtbAddSample/AddOp/CMakePresets.json new file mode 100644 index 000000000..ddec9431b --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/CMakePresets.json @@ -0,0 +1,63 @@ +{ + "version": 1, + "cmakeMinimumRequired": { + "major": 3, + "minor": 19, + "patch": 0 + }, + "configurePresets": [ + { + "name": "default", + "displayName": "Default Config", + "description": "Default build using Unix Makefiles generator", + "generator": "Unix Makefiles", + "binaryDir": "${sourceDir}/build_out", + "cacheVariables": { + "CMAKE_BUILD_TYPE": { + "type": "STRING", + "value": "Release" + }, + "ENABLE_SOURCE_PACKAGE": { + "type": "BOOL", + "value": "True" + }, + "ENABLE_BINARY_PACKAGE": { + "type": "BOOL", + "value": "True" + }, + "ASCEND_COMPUTE_UNIT": { + "type": "STRING", + "value": "ascend310b;ascend910b" + }, + "ENABLE_TEST": { + "type": "BOOL", + "value": "True" + }, + "vendor_name": { + "type": "STRING", + "value": "customize" + }, + "ASCEND_CANN_PACKAGE_PATH": { + "type": "PATH", + "value": "/usr/local/Ascend/ascend-toolkit/latest" + }, + "ASCEND_PYTHON_EXECUTABLE": { + "type": "STRING", + "value": "python3" + }, + "CMAKE_INSTALL_PREFIX": { + "type": "PATH", + "value": "${sourceDir}/build_out" + }, + "ENABLE_CROSS_COMPILE": { + "type": "BOOL", + "value": "False" + }, + "CMAKE_CROSS_PLATFORM_COMPILER": { + "type": "PATH", + "value": "/usr/bin/aarch64-linux-gnu-g++" + } + } + } + ] +} diff --git a/atb_operator/AtbAddSample/AddOp/README.md b/atb_operator/AtbAddSample/AddOp/README.md new file mode 100644 index 000000000..f694535dd --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/README.md @@ -0,0 +1,199 @@ +# 概述 +本样例基于AddOp算子工程,介绍了单算子工程及aclnn接口说明。 + +## 算子工程介绍 +### 算子描述 +Add算子实现了两个数据相加,返回相加结果的功能,该算子实现了非32B对齐shape下的算子功能,完成了算子的泛化实现。 +对应的数学表达式为: +``` +z = x + y +``` +### 算子规格描述 + + + + + + + + + + + +
算子类型(OpType)Add
算子输入nameshapedata typeformat
x-float32,float16,int32,int8ND
y-float32,float16,int32,int8ND
算子输出z-float32,float16,int32,int8ND
核函数名add_custom
+ + +### 算子工程文件介绍 +其中,算子工程目录AddOp包含算子实现的模板文件、编译脚本等,如下所示: +``` +├── AddOp //Add自定义算子工程 +│ ├── cmake +│ ├── framework // 算子插件实现文件目录,单算子模型文件的生成不依赖算子适配插件,无需关注 +│ ├── op_host // host侧实现文件 +│ ├── op_kernel // kernel侧实现文件 +│ ├── scripts // 自定义算子工程打包相关脚本所在目录 +│ ├── build.sh // 编译入口脚本 +│ ├── CMakeLists.txt // 算子工程的CMakeLists.txt +│ └── CMakePresets.json // 编译配置项 +``` + +### 编译运行样例算子 +针对自定义算子工程,编译运行包含如下步骤: +- 编译自定义算子工程生成算子安装包; +- 安装自定义算子到算子库中; +- 调用执行自定义算子; + +详细操作如下所示。 +#### 1. 获取源码包 +编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 +#### 2. 编译算子工程 + 编译自定义算子工程,构建生成自定义算子包。 + + - 执行如下命令,切换到算子工程AddOp目录。 + + ```bash + cd ${git_clone_path}/samples/atb_operator/AtbAddSample/AddOp + ``` + + - 修改CMakePresets.json中ASCEND_CANN_PACKAGE_PATH为CANN软件包安装后的实际路径。 + + + ```json + { + …… + "configurePresets": [ + { + …… + "ASCEND_CANN_PACKAGE_PATH": { + "type": "PATH", + "value": "/usr/local/Ascend/ascend-toolkit/latest" //请替换为CANN软件包安装后的实际路径。eg:/home/HwHiAiUser/Ascend/ascend-toolkit/latest + }, + …… + } + ] + } + ``` + - 在算子工程AddOp目录下执行如下命令,进行算子工程编译。 + + ```bash + ./build.sh + ``` +编译成功后,会在当前目录下创建build_out目录,并在build_out目录下生成自定义算子安装包custom_opp_\_\.run,例如“custom_opp_ubuntu_x86_64.run”。 + +备注:如果要使用dump调试功能,需要移除op_host内和CMakeLists.txt内的Atlas 200/500 A2 推理产品的配置。 + +#### 3. 部署算子包 + +执行如下命令,在自定义算子安装包所在路径下,安装自定义算子包。 + ```bash + cd build_out + ./custom_opp__.run + ``` +命令执行成功后,自定义算子包中的相关文件将部署至当前环境的OPP算子库的vendors/customize目录中。 + +#### 4. 配置环境变量 + + 请根据当前环境上CANN开发套件包的[安装方式](https://hiascend.com/document/redirect/CannCommunityInstSoftware),选择对应配置环境变量的命令。 + - 默认路径,root用户安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=/usr/local/Ascend/ascend-toolkit/latest + ``` + - 默认路径,非root用户安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=$HOME/Ascend/ascend-toolkit/latest + ``` + - 指定路径install_path,安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=${install_path}/ascend-toolkit/latest + ``` + +## aclnn接口说明 + +自定义算子编译部署后,会自动生成单算子API,可以直接在应用程序中调用。算子API的形式一般定义为“两段式接口”,形如: + ```cpp + aclnnStatus aclnnAddOpGetWorkspaceSize(const aclTensor *x, const aclTensor *y, const alcTensor *out, uint64_t workspaceSize, aclOpExecutor **executor); + aclnnStatus aclnnAddOp(void *workspace, int64_t workspaceSize, aclOpExecutor **executor, aclrtStream stream); + ``` +其中aclnnAddOpGetWorkspaceSize为第一段接口,主要用于计算本次API调用计算过程中需要多少的workspace内存。获取到本次API计算需要的workspace大小之后,按照workspaceSize大小申请Device侧内存,然后调用第二段接口aclnnAddOp执行计算。具体参考[AscendCL单算子调用](https://hiascend.com/document/redirect/CannCommunityAscendCInVorkSingleOp)>单算子API执行 章节。 + +### 功能描述 +* 算子功能:完成加法计算。 +* 计算公式: +``` +z = x + y +``` + +#### 参数说明 +##### aclnnAddOpGetWorkspaceSize: + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
接口参数描述
aclnnAddOpGetWorkspaceSizex输入x的Tensor,支持flaot/half/int8/int32类型,ND排布格式
y输入y的Tensor,支持flaot/half/int8/int32类型,ND排布格式
out输出z的Tensor,支持flaot/half/int8/int32类型,ND排布格式
workspaceSize返回需要在Device侧申请的workspace大小
executor返回op执行器,包含了算子计算流程
+ + +| 返回值 | 描述 | +| ---- | ---- | +| ACL_SUCCESS | 成功 | +| ACL_ERROR | 失败 | + +##### aclnnAddOp + + + + + + + + + + + + + + + + + + + + + + + +
接口参数描述
aclnnAddOpworkspace输入Device侧申请的workspace内存
workspaceSizeworkspace大小
executor输入op执行器,包含了算子计算流程
stream输入执行流
+ +| 返回值 | 描述 | +| ---- | ---- | +| ACL_SUCCESS | 成功 | +| ACL_ERROR | 失败 | + + +## 更新说明 +| 时间 | 更新事项 | +|----|------| +| 2025/03/27 | 新增README | \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AddOp/build.sh b/atb_operator/AtbAddSample/AddOp/build.sh new file mode 100644 index 000000000..4be96d7d8 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/build.sh @@ -0,0 +1,37 @@ +#!/bin/bash +script_path=$(realpath $(dirname $0)) + + +mkdir -p build_out +rm -rf build_out/* +cd build_out + +cmake_version=$(cmake --version | grep "cmake version" | awk '{print $3}') +if [ "$cmake_version" \< "3.19.0" ] ; then + opts=$(python3 $script_path/cmake/util/preset_parse.py $script_path/CMakePresets.json) + echo $opts + cmake .. $opts +else + cmake .. --preset=default +fi +target=package +if [ "$1"x != ""x ]; then target=$1; fi + +cmake --build . --target $target -j16 +if [ $? -ne 0 ]; then exit 1; fi + +if [ $target = "package" ]; then + if test -d ./op_kernel/binary ; then + ./cust*.run + if [ $? -ne 0 ]; then exit 1; fi + cmake --build . --target binary -j16 + if [ $? -ne 0 ]; then exit 1; fi + cmake --build . --target $target -j16 + fi +fi + +# for debug +# cd build_out +# make +# cpack +# verbose append -v diff --git a/atb_operator/AtbAddSample/AddOp/cmake/config.cmake b/atb_operator/AtbAddSample/AddOp/cmake/config.cmake new file mode 100644 index 000000000..886119daa --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/config.cmake @@ -0,0 +1,25 @@ + +set(CMAKE_CXX_FLAGS_DEBUG "") +set(CMAKE_CXX_FLAGS_RELEASE "") + +if (NOT DEFINED vendor_name) + set(vendor_name customize CACHE STRING "") +endif() +if (NOT DEFINED ASCEND_CANN_PACKAGE_PATH) + set(ASCEND_CANN_PACKAGE_PATH /usr/local/Ascend/latest CACHE PATH "") +endif() +if (NOT DEFINED ASCEND_PYTHON_EXECUTABLE) + set(ASCEND_PYTHON_EXECUTABLE python3 CACHE STRING "") +endif() +if (NOT DEFINED ASCEND_COMPUTE_UNIT) + message(FATAL_ERROR "ASCEND_COMPUTE_UNIT not set in CMakePreset.json ! +") +endif() +set(ASCEND_TENSOR_COMPILER_PATH ${ASCEND_CANN_PACKAGE_PATH}/compiler) +set(ASCEND_CCEC_COMPILER_PATH ${ASCEND_TENSOR_COMPILER_PATH}/ccec_compiler/bin) +set(ASCEND_AUTOGEN_PATH ${CMAKE_BINARY_DIR}/autogen) +set(ASCEND_FRAMEWORK_TYPE tensorflow) +file(MAKE_DIRECTORY ${ASCEND_AUTOGEN_PATH}) +set(CUSTOM_COMPILE_OPTIONS "custom_compile_options.ini") +execute_process(COMMAND rm -rf ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} + COMMAND touch ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS}) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/func.cmake b/atb_operator/AtbAddSample/AddOp/cmake/func.cmake new file mode 100644 index 000000000..ad187e7d6 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/func.cmake @@ -0,0 +1,228 @@ + +function(get_system_info SYSTEM_INFO) + if (UNIX) + execute_process(COMMAND grep -i ^id= /etc/os-release OUTPUT_VARIABLE TEMP) + string(REGEX REPLACE "\n|id=|ID=|\"" "" SYSTEM_NAME ${TEMP}) + set(${SYSTEM_INFO} ${SYSTEM_NAME}_${CMAKE_SYSTEM_PROCESSOR} PARENT_SCOPE) + elseif (WIN32) + message(STATUS "System is Windows. Only for pre-build.") + else () + message(FATAL_ERROR "${CMAKE_SYSTEM_NAME} not support.") + endif () +endfunction() + +function(opbuild) + message(STATUS "Opbuild generating sources") + cmake_parse_arguments(OPBUILD "" "OUT_DIR;PROJECT_NAME;ACCESS_PREFIX" "OPS_SRC" ${ARGN}) + execute_process(COMMAND ${CMAKE_COMPILE} -g -fPIC -shared -std=c++11 ${OPBUILD_OPS_SRC} -D_GLIBCXX_USE_CXX11_ABI=0 + -I ${ASCEND_CANN_PACKAGE_PATH}/include -L ${ASCEND_CANN_PACKAGE_PATH}/lib64 -lexe_graph -lregister -ltiling_api + -o ${OPBUILD_OUT_DIR}/libascend_all_ops.so + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("build ops lib info: ${EXEC_INFO}") + message("build ops lib error: ${EXEC_ERROR}") + message(FATAL_ERROR "opbuild run failed!") + endif() + set(proj_env "") + set(prefix_env "") + if (NOT "${OPBUILD_PROJECT_NAME}x" STREQUAL "x") + set(proj_env "OPS_PROJECT_NAME=${OPBUILD_PROJECT_NAME}") + endif() + if (NOT "${OPBUILD_ACCESS_PREFIX}x" STREQUAL "x") + set(prefix_env "OPS_DIRECT_ACCESS_PREFIX=${OPBUILD_ACCESS_PREFIX}") + endif() + execute_process(COMMAND ${proj_env} ${prefix_env} ${ASCEND_CANN_PACKAGE_PATH}/toolkit/tools/opbuild/op_build + ${OPBUILD_OUT_DIR}/libascend_all_ops.so ${OPBUILD_OUT_DIR} + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("opbuild ops info: ${EXEC_INFO}") + message("opbuild ops error: ${EXEC_ERROR}") + endif() + message(STATUS "Opbuild generating sources - done") +endfunction() + +function(add_ops_info_target) + cmake_parse_arguments(OPINFO "" "TARGET;OPS_INFO;OUTPUT;INSTALL_DIR" "" ${ARGN}) + get_filename_component(opinfo_file_path "${OPINFO_OUTPUT}" DIRECTORY) + add_custom_command(OUTPUT ${OPINFO_OUTPUT} + COMMAND mkdir -p ${opinfo_file_path} + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/parse_ini_to_json.py + ${OPINFO_OPS_INFO} ${OPINFO_OUTPUT} + ) + add_custom_target(${OPINFO_TARGET} ALL + DEPENDS ${OPINFO_OUTPUT} + ) + install(FILES ${OPINFO_OUTPUT} + DESTINATION ${OPINFO_INSTALL_DIR} + ) +endfunction() + +function(add_ops_compile_options OP_TYPE) + cmake_parse_arguments(OP_COMPILE "" "OP_TYPE" "COMPUTE_UNIT;OPTIONS" ${ARGN}) + file(APPEND ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} + "${OP_TYPE},${OP_COMPILE_COMPUTE_UNIT},${OP_COMPILE_OPTIONS}\n") +endfunction() + +function(add_ops_impl_target) + cmake_parse_arguments(OPIMPL "" "TARGET;OPS_INFO;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) + add_custom_command(OUTPUT ${OPIMPL_OUT_DIR}/.impl_timestamp + COMMAND mkdir -m 700 -p ${OPIMPL_OUT_DIR}/dynamic + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py + ${OPIMPL_OPS_INFO} + \"${OPIMPL_OPS_BATCH}\" \"${OPIMPL_OPS_ITERATE}\" + ${OPIMPL_IMPL_DIR} + ${OPIMPL_OUT_DIR}/dynamic + ${ASCEND_AUTOGEN_PATH} + + COMMAND rm -rf ${OPIMPL_OUT_DIR}/.impl_timestamp + COMMAND touch ${OPIMPL_OUT_DIR}/.impl_timestamp + DEPENDS ${OPIMPL_OPS_INFO} + ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py + ) + add_custom_target(${OPIMPL_TARGET} ALL + DEPENDS ${OPIMPL_OUT_DIR}/.impl_timestamp) + if (${ENABLE_SOURCE_PACKAGE}) + install(DIRECTORY ${OPIMPL_OUT_DIR}/dynamic + DESTINATION ${OPIMPL_INSTALL_DIR} + ) + endif() +endfunction() + +function(add_ops_replay_targets) + cmake_parse_arguments(OPREPLAY "" "OPS_INFO;COMPUTE_UNIT;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) + # ccec compile options + set(ccec_base_opts -c -O2 --cce-aicore-only -mllvm -cce-aicore-function-stack-size=16000 + -mllvm -cce-aicore-record-overflow=false -std=c++17) + set(ccec_extopts_ascend310p --cce-aicore-arch=dav-m200 -mllvm -cce-aicore-fp-ceiling=2) + set(ccec_extopts_ascend910 --cce-aicore-arch=dav-c100) + set(ccec_extopts_ascend910b --cce-aicore-arch=dav-c220-cube) + file(MAKE_DIRECTORY ${OPREPLAY_OUT_DIR}) + execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_replay_build.py + ${OPREPLAY_OPS_INFO} + "${OPREPLAY_OPS_BATCH}" "${OPREPLAY_OPS_ITERATE}" + ${OPREPLAY_IMPL_DIR} + ${OPREPLAY_OUT_DIR} + ${OPREPLAY_COMPUTE_UNIT} + ) + file(GLOB replay_kernel_entries ${OPREPLAY_OUT_DIR}/*.cce) + if (NOT "${replay_kernel_entries}x" STREQUAL "x") + foreach(replay_kernel_file ${replay_kernel_entries}) + get_filename_component(replay_kernel_file_name "${replay_kernel_file}" NAME) + string(REPLACE "_entry.cce" "" op_kerne_name ${replay_kernel_file_name}) + file(GLOB replay_lib_src ${OPREPLAY_OUT_DIR}/${op_kerne_name}*.cpp) + set(OP_TILING_DATA_H_PATH ${OPREPLAY_OUT_DIR}/${op_kerne_name}_tiling_data.h) + add_library(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} SHARED ${replay_lib_src}) + if(EXISTS ${OP_TILING_DATA_H_PATH}) + target_compile_options(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE + -include ${OP_TILING_DATA_H_PATH} + ) + endif() + target_compile_definitions(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE + ${op_kerne_name}=${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} + ) + target_compile_options(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE + -D__ASCENDC_REPLAY__ + ) + target_link_libraries(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE intf_pub + tikreplaylib::${OPREPLAY_COMPUTE_UNIT} + register + ) + add_custom_command(OUTPUT ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o + COMMAND ccec ${ccec_base_opts} ${ccec_extopts_${OPREPLAY_COMPUTE_UNIT}} ${replay_kernel_file} + -o ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o + DEPENDS ${replay_kernel_file} + ) + add_custom_target(replay_kernel_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} ALL + DEPENDS ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o + ) + install(TARGETS replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_replay + ) + install(FILES ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o + DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_replay + ) + endforeach() + endif() +endfunction() + +function(add_npu_support_target) + cmake_parse_arguments(NPUSUP "" "TARGET;OPS_INFO_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) + get_filename_component(npu_sup_file_path "${NPUSUP_OUT_DIR}" DIRECTORY) + add_custom_command(OUTPUT ${NPUSUP_OUT_DIR}/npu_supported_ops.json + COMMAND mkdir -p ${NPUSUP_OUT_DIR} + COMMAND ${CMAKE_SOURCE_DIR}/cmake/util/gen_ops_filter.sh + ${NPUSUP_OPS_INFO_DIR} + ${NPUSUP_OUT_DIR} + ) + add_custom_target(npu_supported_ops ALL + DEPENDS ${NPUSUP_OUT_DIR}/npu_supported_ops.json + ) + install(FILES ${NPUSUP_OUT_DIR}/npu_supported_ops.json + DESTINATION ${NPUSUP_INSTALL_DIR} + ) +endfunction() + +function(add_bin_compile_target) + cmake_parse_arguments(BINCMP "" "TARGET;OPS_INFO;COMPUTE_UNIT;IMPL_DIR;ADP_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/src) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/gen) + execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_bin_param_build.py + ${BINCMP_OPS_INFO} ${BINCMP_OUT_DIR}/gen ${BINCMP_COMPUTE_UNIT} + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("ops binary compile scripts gen info: ${EXEC_INFO}") + message("ops binary compile scripts gen error: ${EXEC_ERROR}") + message(FATAL_ERROR "ops binary compile scripts gen failed!") + endif() + if (NOT TARGET binary) + add_custom_target(binary) + endif() + add_custom_target(${BINCMP_TARGET} + COMMAND cp -r ${BINCMP_IMPL_DIR}/*.* ${BINCMP_OUT_DIR}/src + ) + add_custom_target(${BINCMP_TARGET}_gen_ops_config + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/insert_simplified_keys.py -p ${BINCMP_OUT_DIR}/bin + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_ops_config.py -p ${BINCMP_OUT_DIR}/bin + -s ${BINCMP_COMPUTE_UNIT} + ) + add_dependencies(binary ${BINCMP_TARGET}_gen_ops_config) + file(GLOB bin_scripts ${BINCMP_OUT_DIR}/gen/*.sh) + foreach(bin_script ${bin_scripts}) + get_filename_component(bin_file ${bin_script} NAME_WE) + string(REPLACE "-" ";" bin_sep ${bin_file}) + list(GET bin_sep 0 op_type) + list(GET bin_sep 1 op_file) + list(GET bin_sep 2 op_index) + if (NOT TARGET ${BINCMP_TARGET}_${op_file}_copy) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file}) + add_custom_target(${BINCMP_TARGET}_${op_file}_copy + COMMAND cp ${BINCMP_ADP_DIR}/${op_file}.py ${BINCMP_OUT_DIR}/src/${op_type}.py + ) + install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} + DESTINATION ${BINCMP_INSTALL_DIR}/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) + install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json + DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT}/ OPTIONAL + ) + endif() + add_custom_target(${BINCMP_TARGET}_${op_file}_${op_index} + COMMAND export HI_PYTHON=${ASCEND_PYTHON_EXECUTABLE} && bash ${bin_script} ${BINCMP_OUT_DIR}/src/${op_type}.py ${BINCMP_OUT_DIR}/bin/${op_file} + WORKING_DIRECTORY ${BINCMP_OUT_DIR} + ) + add_dependencies(${BINCMP_TARGET}_${op_file}_${op_index} ${BINCMP_TARGET} ${BINCMP_TARGET}_${op_file}_copy) + add_dependencies(${BINCMP_TARGET}_gen_ops_config ${BINCMP_TARGET}_${op_file}_${op_index}) + endforeach() + install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json + DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) +endfunction() diff --git a/atb_operator/AtbAddSample/AddOp/cmake/intf.cmake b/atb_operator/AtbAddSample/AddOp/cmake/intf.cmake new file mode 100644 index 000000000..2f362c396 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/intf.cmake @@ -0,0 +1,26 @@ + +add_library(intf_pub INTERFACE) +target_compile_options(intf_pub INTERFACE + -fPIC + -fvisibility=hidden + -fvisibility-inlines-hidden + $<$:-O2> + $<$:-O0 -g> + $<$:-std=c++11> + $<$,$>:-ftrapv -fstack-check> + $<$:-pthread -Wfloat-equal -Wshadow -Wformat=2 -Wno-deprecated -Wextra> + $,-fstack-protector-strong,-fstack-protector-all> +) +target_compile_definitions(intf_pub INTERFACE + _GLIBCXX_USE_CXX11_ABI=0 + $<$:_FORTIFY_SOURCE=2> +) +target_include_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/include) +target_link_options(intf_pub INTERFACE + $<$,EXECUTABLE>:-pie> + $<$:-s> + -Wl,-z,relro + -Wl,-z,now + -Wl,-z,noexecstack +) +target_link_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/lib64) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/makeself.cmake b/atb_operator/AtbAddSample/AddOp/cmake/makeself.cmake new file mode 100644 index 000000000..48c565bfb --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/makeself.cmake @@ -0,0 +1,17 @@ +execute_process(COMMAND chmod +x ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh) +execute_process(COMMAND ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh + --header ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself-header.sh + --help-header ./help.info + --gzip --complevel 4 --nomd5 --sha256 + ./ ${CPACK_PACKAGE_FILE_NAME} "version:1.0" ./install.sh + WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} + RESULT_VARIABLE EXEC_RESULT + ERROR_VARIABLE EXEC_ERROR +) +if (NOT "${EXEC_RESULT}x" STREQUAL "0x") + message(FATAL_ERROR "CPack Command error: ${EXEC_RESULT}\n${EXEC_ERROR}") +endif() +execute_process(COMMAND cp ${CPACK_EXTERNAL_BUILT_PACKAGES} ${CPACK_PACKAGE_DIRECTORY}/ + COMMAND echo "Copy ${CPACK_EXTERNAL_BUILT_PACKAGES} to ${CPACK_PACKAGE_DIRECTORY}/" + WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} +) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/__init__.py b/atb_operator/AtbAddSample/AddOp/cmake/util/__init__.py new file mode 100644 index 000000000..c4ddc893a --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/__init__.py @@ -0,0 +1,8 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- + +import sys +import os + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) +sys.path.append(PYF_PATH) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_bin_param_build.py b/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_bin_param_build.py new file mode 100644 index 000000000..decf34544 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_bin_param_build.py @@ -0,0 +1,129 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import json +import hashlib +import const_var +import opdesc_parser + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + + +class BinParamBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + super().__init__(op_type) + self.soc = '' + self.out_path = '' + + def set_soc_version(self: any, soc: str): + self.soc = soc + + def set_out_path(self: any, out_path: str): + self.out_path = out_path + + def gen_input_json(self: any): + key_map = {} + count = len(self.input_dtype[0].split(',')) + for i in range(0, count): + inputs = [] + outputs = [] + attrs = [] + op_node = {} + for idx in range(0, len(self.input_name)): + idtypes = self.input_dtype[idx].split(',') + ifmts = self.input_fmt[idx].split(',') + itype = self.input_type[idx] + para = {} + para['name'] = self.input_name[idx] + para['index'] = idx + para['dtype'] = idtypes[i] + para['format'] = ifmts[i] + para['paramType'] = itype + para['shape'] = [-2] + if itype == 'dynamic': + inputs.append([para]) + else: + inputs.append(para) + for idx in range(0, len(self.output_name)): + odtypes = self.output_dtype[idx].split(',') + ofmts = self.output_fmt[idx].split(',') + otype = self.output_type[idx] + para = {} + para['name'] = self.output_name[idx] + para['index'] = idx + para['dtype'] = odtypes[i] + para['format'] = ofmts[i] + para['paramType'] = otype + para['shape'] = [-2] + if otype == 'dynamic': + outputs.append([para]) + else: + outputs.append(para) + for attr in self.attr_list: + att = {} + att['name'] = attr + atype = self.attr_val.get(attr).get('type').lower() + atype = atype.replace('list', 'list_') + att['dtype'] = atype + att['value'] = const_var.ATTR_DEF_VAL.get(atype) + attrs.append(att) + op_node['bin_filename'] = '' + op_node['inputs'] = inputs + op_node['outputs'] = outputs + if len(attrs) > 0: + op_node['attrs'] = attrs + param = {} + param['op_type'] = self.op_type + param['op_list'] = [op_node] + objstr = json.dumps(param, indent=' ') + md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() + while key_map.get(md5sum) is not None: + objstr += '1' + md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() + key_map[md5sum] = md5sum + bin_file = self.op_type + '_' + md5sum + op_node['bin_filename'] = bin_file + param_file = os.path.join(self.out_path, bin_file + '_param.json') + param_file = os.path.realpath(param_file) + with os.fdopen(os.open(param_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + json.dump(param, fd, indent=' ') + self._write_buld_cmd(param_file, bin_file, i) + + + def _write_buld_cmd(self: any, param_file: str, bin_file: str, index: int): + hard_soc = const_var.SOC_MAP_EXT.get(self.soc) + if not hard_soc: + hard_soc = soc.capitalize() + name_com = [self.op_type, self.op_file, str(index)] + compile_file = os.path.join(self.out_path, '-'.join(name_com) + '.sh') + compile_file = os.path.realpath(compile_file) + with os.fdopen(os.open(compile_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + fd.write('#!/bin/bash\n') + fd.write('echo "[{}] Generating {} ..."\n'.format(hard_soc, bin_file)) + cmd = const_var.BIN_CMD.format(fun=self.op_intf, soc=hard_soc, param=param_file, impl='""') + fd.write(cmd) + chk = const_var.CHK_CMD.format(res_file=bin_file + '.json') + fd.write(chk) + chk = const_var.CHK_CMD.format(res_file=bin_file + '.o') + fd.write(chk) + fd.write('echo "[{}] Generating {} Done"\n'.format(hard_soc, bin_file)) + + +def gen_bin_param_file(cfgfile: str, out_dir: str, soc: str): + op_descs = opdesc_parser.get_op_desc(cfgfile, [], [], BinParamBuilder, None) + for op_desc in op_descs: + op_desc.set_soc_version(soc) + op_desc.set_out_path(out_dir) + op_desc.gen_input_json() + + +if __name__ == '__main__': + if len(sys.argv) <= 3: + raise RuntimeError('arguments must greater than 3') + gen_bin_param_file(sys.argv[1], sys.argv[2], sys.argv[3]) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_impl_build.py b/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_impl_build.py new file mode 100644 index 000000000..7fe177da1 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_impl_build.py @@ -0,0 +1,446 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import re +import stat +import opdesc_parser +import const_var + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + +IMPL_HEAD = ''' +import os, sys +import ctypes +import json +import shutil +from tbe.common.platform import get_soc_spec +from tbe.common.utils import para_check +from tbe.tikcpp import compile_op, replay_op, check_op_cap, generalize_op_params, get_code_channel, OpInfo +from tbe.common.buildcfg import get_default_build_config +from impl.util.platform_adapter import tbe_register +from tbe.common.buildcfg import get_current_build_config +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + +DTYPE_MAP = {"float32": ["DT_FLOAT", "float"], + "float16": ["DT_FLOAT16", "half"], + "int8": ["DT_INT8", "int8_t"], + "int16": ["DT_INT16", "int16_t"], + "int32": ["DT_INT32", "int32_t"], + "int64": ["DT_INT64", "int64_t"], + "uint1": ["DT_UINT1", "uint8_t"], + "uint8": ["DT_UINT8", "uint8_t"], + "uint16": ["DT_UINT16", "uint16_t"], + "uint32": ["DT_UINT32", "uint32_t"], + "uint64": ["DT_UINT64", "uint64_t"], + "bool": ["DT_BOOL", "bool"], + "double": ["DT_DOUBLE", "double"], + "dual": ["DT_DUAL", "unknown"], + "dual_sub_int8": ["DT_DUAL_SUB_INT8", "unknown"], + "dual_sub_uint8": ["DT_DUAL_SUB_UINT8", "unknown"], + "string": ["DT_STRING", "unknown"], + "complex64": ["DT_COMPLEX64", "unknown"], + "complex128": ["DT_COMPLEX128", "unknown"], + "qint8": ["DT_QINT8", "unknown"], + "qint16": ["DT_QINT16", "unknown"], + "qint32": ["DT_QINT32", "unknown"], + "quint8": ["DT_QUINT8", "unknown"], + "quint16": ["DT_QUINT16", "unknown"], + "resource": ["DT_RESOURCE", "unknown"], + "string_ref": ["DT_STRING_REF", "unknown"], + "int4": ["DT_INT4", "int8_t"], + "bfloat16": ["DT_BF16", "bfloat16_t"]} + +def get_dtype_fmt_options(__inputs__, __outputs__): + options = [] + for x in __inputs__ + __outputs__: + x_n = x.get("param_name").upper() + x_fmt = x.get("format") + x_dtype = x.get("dtype") + options.append("-DDTYPE_{n}={t}".format(n=x_n, t=DTYPE_MAP.get(x_dtype)[1])) + options.append("-DORIG_DTYPE_{n}={ot}".format(n=x_n, ot=DTYPE_MAP.get(x_dtype)[0])) + options.append("-DFORMAT_{n}=FORMAT_{f}".format(n=x_n, f=x_fmt)) + return options + +def load_dso(so_path): + try: + ctypes.CDLL(so_path) + except OSError as error : + print(error) + raise RuntimeError("cannot open %s" %(so_path)) + else: + print("load so succ ", so_path) + +def get_shortsoc_compile_option(compile_option_list: list, shortsoc:str): + compile_options = [] + if shortsoc in compile_option_list: + compile_options = compile_option_list[shortsoc] + elif '__ALLSOC__' in compile_option_list: + compile_options = compile_option_list['__ALLSOC__'] + return compile_options + +''' + +IMPL_API = ''' +@tbe_register.register_operator("{}") +@para_check.check_op_params({}) +def {}({}, kernel_name="{}", impl_mode=""): + if get_current_build_config("enable_op_prebuild"): + return + __inputs__, __outputs__, __attrs__ = _build_args({}) + options = get_dtype_fmt_options(__inputs__, __outputs__) + options += ["-x", "cce"] + ccec = os.environ.get('CCEC_REAL_PATH') + if ccec is None: + ccec = shutil.which("ccec") + if ccec != None: + ccec_path = os.path.dirname(ccec) + tikcpp_path = os.path.realpath(os.path.join(ccec_path, "..", "..", "tikcpp")) + else: + tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") + options.append("-I" + tikcpp_path) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) + options.append("-I" + os.path.join(PYF_PATH, "..", "ascendc", "common")) + if impl_mode == "high_performance": + options.append("-DHIGH_PERFORMANCE=1") + elif impl_mode == "high_precision": + options.append("-DHIGH_PRECISION=1") + if get_default_build_config("enable_deterministic_mode") == 1: + options.append("-DDETEMINISTIC_MODE=1") + + custom_compile_options = {}, + custom_all_compile_options = {}, + soc_version = get_soc_spec("SOC_VERSION") + soc_short = get_soc_spec("SHORT_SOC_VERSION").lower() + custom_compile_options_soc = get_shortsoc_compile_option(custom_compile_options[0], soc_short) + custom_all_compile_options_soc = get_shortsoc_compile_option(custom_all_compile_options[0], soc_short) + options += custom_all_compile_options_soc + options += custom_compile_options_soc + + origin_func_name = "{}" + ascendc_src_dir = "{}" + ascendc_src_file = "{}" + src = os.path.join(PYF_PATH, "..", "ascendc", ascendc_src_dir, ascendc_src_file) + if not os.path.exists(src): + src = os.path.join(PYF_PATH, ascendc_src_file) +''' + +REPLAY_OP_API = ''' + print("start replay Acend C Operator {}, kernel name is {}") + tikreplay_codegen_path = tikcpp_path + "/tikreplaylib/lib" + tikreplay_stub_path = tikcpp_path + "/tikreplaylib/lib/" + soc_version + print("start load libtikreplaylib_codegen.so and libtikreplaylib_stub.so") + codegen_so_path = tikreplay_codegen_path + "/libtikreplaylib_codegen.so" + replaystub_so_path = tikreplay_stub_path + "/libtikreplaylib_stub.so" + if PYF_PATH.endswith("dynamic"): + op_replay_path = os.path.join(PYF_PATH, "..", "..", "op_replay") + else: + op_replay_path = os.path.join(PYF_PATH, "..", "op_replay") + replayapi_so_path = os.path.join(op_replay_path, "libreplay_{}_" + soc_short + ".so") + load_dso(codegen_so_path) + load_dso(replaystub_so_path) + load_dso(replayapi_so_path) + op_type = "{}" + entry_obj = os.path.join(op_replay_path, "{}_entry_" + soc_short + ".o") + code_channel = get_code_channel(src, kernel_name, op_type, options) + op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ + attrs = __attrs__, impl_mode = impl_mode) + res, msg = replay_op(op_info, entry_obj, code_channel, src, options) + if not res: + print("call replay op failed for %s and get into call compile op" %(msg)) + compile_op(src, origin_func_name, op_info, options, code_channel, '{}') +''' + +COMPILE_OP_API = ''' + print("start compile Ascend C operator {}. kernel name is {}") + op_type = "{}" + code_channel = get_code_channel(src, kernel_name, op_type, options) + op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ + attrs = __attrs__, impl_mode = impl_mode, origin_inputs=[{}], origin_outputs = [{}]) + compile_op(src, origin_func_name, op_info, options, code_channel, '{}') +''' + +SUP_API = ''' +def {}({}, impl_mode=""): + __inputs__, __outputs__, __attrs__ = _build_args({}) + ret_str = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) + ret_dict = json.loads(ret_str) + err_code = ret_dict.get("ret_code") + sup = "Unknown" + reason = "Unknown reason" + if err_code is not None: + if err_code is 0: + sup = "True" + reason = "" + elif err_code is 1: + sup = "False" + reason = ret_dict.get("reason") + else: + sup = "Unknown" + reason = ret_dict.get("reason") + return sup, reason +''' +CAP_API = ''' +def {}({}, impl_mode=""): + __inputs__, __outputs__, __attrs__ = _build_args({}) + result = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) + return result.decode("utf-8") +''' +GLZ_API = ''' +@tbe_register.register_param_generalization("{}") +def {}_generalization({}, generalize_config=None): + __inputs__, __outputs__, __attrs__ = _build_args({}) + ret_str = generalize_op_params("{}", __inputs__, __outputs__, __attrs__, generalize_config) + return [json.loads(ret_str)] +''' + +ATTR_DEFAULT = {'bool': 'False', 'int': '0', 'float': '0.0', 'listInt': '[]', + 'listFloat': '[]', 'listBool': '[]', 'listListInt': '[[]]', 'str': ''} + + +def optype_snake(origin_str): + temp_str = origin_str[0].lower() + origin_str[1:] + new_str = re.sub(r'([A-Z])', r'_\1', temp_str).lower() + return new_str + + +class AdpBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + self.argsname = [] + self.argsdefv = [] + self.op_compile_option:str = '{}' + super().__init__(op_type) + + + def write_adapt(self: any, impl_path, path: str, op_compile_option_all:list = None): + self._build_paradefault() + if impl_path != "": + src_file = os.path.join(impl_path, self.op_file + '.cpp') + if not os.path.exists(src_file): + return + out_path = os.path.abspath(path) + if self.dynamic_shape and not out_path.endswith('dynamic'): + out_path = os.path.join(path, 'dynamic') + os.makedirs(out_path, mode=0o700, exist_ok=True) + adpfile = os.path.join(out_path, self.op_file + '.py') + self._gen_op_compile_option(op_compile_option_all) + with os.fdopen(os.open(adpfile, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + self._write_head(fd) + self._write_argparse(fd) + self._write_impl(fd) + if self.op_chk_support: + self._write_cap('check_supported', fd) + self._write_cap('get_op_support_info', fd) + if self.op_fmt_sel: + self._write_cap('op_select_format', fd) + self._write_cap('get_op_specific_info', fd) + if self.op_range_limit == 'limited' or self.op_range_limit == 'dynamic': + self._write_glz(fd) + + + def _gen_op_compile_option(self:any, op_compile_option_all:list =None): + if op_compile_option_all is not None: + if self.op_type in op_compile_option_all: + self.op_compile_option = op_compile_option_all[self.op_type] + elif "__all__" in op_compile_option_all: + self.op_compile_option = op_compile_option_all["__all__"] + + + def _ip_argpack(self: any, default: bool = True) -> list: + args = [] + for i in range(len(self.input_name)): + arg = self.input_name[i] + if default and self.argsdefv[i] is not None: + arg += '=' + self.argsdefv[i] + args.append(arg) + return args + + def _op_argpack(self: any, default: bool = True) -> list: + args = [] + argidx = len(self.input_name) + for i in range(len(self.output_name)): + arg = self.output_name[i] + if default and self.argsdefv[i + argidx] is not None: + arg += '=' + self.argsdefv[i + argidx] + args.append(arg) + return args + + def _attr_argpack(self: any, default: bool = True) -> list: + args = [] + argidx = len(self.input_name) + len(self.output_name) + for i in range(len(self.attr_list)): + att = self.attr_list[i] + arg = att + if default and self.argsdefv[i + argidx] is not None: + if self.attr_val.get(att).get('type') == 'str': + arg += '="' + self.argsdefv[i + argidx] + '"' + elif self.attr_val.get(att).get('type') == 'bool': + arg += '=' + self.argsdefv[i + argidx].capitalize() + else: + arg += '=' + self.argsdefv[i + argidx] + args.append(arg) + return args + + def _build_paralist(self: any, default: bool = True) -> str: + args = [] + args.extend(self._ip_argpack(default)) + args.extend(self._op_argpack(default)) + args.extend(self._attr_argpack(default)) + return ', '.join(args) + + def _io_parachk(self: any, types: list, type_name: str) -> list: + chk = [] + for iot in types: + if iot == 'optional': + ptype = 'OPTION' + else: + ptype = iot.upper() + chk.append('para_check.{}_{}'.format(ptype, type_name)) + return chk + + def _attr_parachk(self: any) -> list: + chk = [] + for att in self.attr_list: + if self.attr_val.get(att).get('paramType') == 'optional': + pt = 'OPTION' + else: + pt = self.attr_val.get(att).get('paramType').upper() + att_type = self.attr_val.get(att).get('type').upper() + att_type = att_type.replace('LIST', 'LIST_') + chk.append('para_check.{}_ATTR_{}'.format(pt, att_type)) + return chk + + def _build_parachk(self: any) -> str: + chk = [] + chk.extend(self._io_parachk(self.input_type, 'INPUT')) + chk.extend(self._io_parachk(self.output_type, 'OUTPUT')) + chk.extend(self._attr_parachk()) + chk.append('para_check.KERNEL_NAME') + return ', '.join(chk) + + def _build_paradefault(self: any): + optional = False + argtypes = [] + argtypes.extend(self.input_type) + argtypes.extend(self.output_type) + for atype in argtypes: + if atype == 'optional': + optional = True + if optional: + self.argsdefv.append('None') + else: + self.argsdefv.append(None) + for attr in self.attr_list: + atype = self.attr_val.get(attr).get('paramType') + if atype == 'optional': + optional = True + attrval = self.attr_val.get(attr).get('defaultValue') + if attrval is not None: + optional = True + if type == "bool": + attrval = attrval.capitalize() + elif type == "str": + attrval = "\"" + attrval + "\"" + self.argsdefv.append(attrval) + continue + if optional: + self.argsdefv.append(ATTR_DEFAULT.get(self.attr_val.get(attr).get('type'))) + else: + self.argsdefv.append(None) + + def _write_head(self: any, fd: object): + fd.write(IMPL_HEAD) + + def _write_argparse(self: any, fd: object): + args = self._build_paralist(False) + fd.write('def _build_args({}):\n'.format(args)) + fd.write(' __inputs__ = []\n') + fd.write(' for arg in [{}]:\n'.format(', '.join(self.input_name))) + fd.write(' if arg != None:\n') + fd.write(' if isinstance(arg, (list, tuple)):\n') + fd.write(' if len(arg) == 0:\n') + fd.write(' continue\n') + fd.write(' __inputs__.append(arg[0])\n') + fd.write(' else:\n') + fd.write(' __inputs__.append(arg)\n') + fd.write(' __outputs__ = []\n') + fd.write(' for arg in [{}]:\n'.format(', '.join(self.output_name))) + fd.write(' if arg != None:\n') + fd.write(' if isinstance(arg, (list, tuple)):\n') + fd.write(' if len(arg) == 0:\n') + fd.write(' continue\n') + fd.write(' __outputs__.append(arg[0])\n') + fd.write(' else:\n') + fd.write(' __outputs__.append(arg)\n') + fd.write(' __attrs__ = []\n') + for attr in self.attr_list: + fd.write(' if {} != None:\n'.format(attr)) + fd.write(' attr = {}\n') + fd.write(' attr["name"] = "{}"\n'.format(attr)) + fd.write(' attr["dtype"] = "{}"\n'.format(self.attr_val.get(attr).get('type'))) + fd.write(' attr["value"] = {}\n'.format(attr)) + fd.write(' __attrs__.append(attr)\n') + fd.write(' return __inputs__, __outputs__, __attrs__\n') + + def _write_impl(self: any, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + pchk = self._build_parachk() + if len(self.kern_name) > 0: + kern_name = self.kern_name + else: + kern_name = self.op_intf + src = self.op_file + '.cpp' + fd.write(IMPL_API.format(self.op_type, pchk, self.op_intf, argsdef, kern_name, argsval,\ + self.custom_compile_options, self.custom_all_compile_options, self.op_intf,\ + optype_snake(self.op_type), src)) + if self.op_replay_flag: + fd.write(REPLAY_OP_API.format(self.op_type, kern_name, self.op_file, self.op_type, self.op_file,\ + self.op_compile_option)) + else: + fd.write(COMPILE_OP_API.format(self.op_type, kern_name, self.op_type, ', '.join(self.input_name),\ + ', '.join(self.output_name), self.op_compile_option)) + + def _write_cap(self: any, cap_name: str, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + if cap_name == 'check_supported': + fd.write(SUP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) + else: + fd.write(CAP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) + + def _write_glz(self: any, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + fd.write(GLZ_API.format(self.op_type, self.op_intf, argsdef, argsval, self.op_type)) + + +def write_scripts(cfgfile: str, cfgs: dict, dirs: dict, ops: list = None, op_compile_option:list = None): + batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') + iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') + file_map = {} + op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, AdpBuilder,\ + ops, dirs.get(const_var.AUTO_GEN_DIR)) + for op_desc in op_descs: + op_desc.write_adapt(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), op_compile_option) + file_map[op_desc.op_type] = op_desc.op_file + return file_map + +if __name__ == '__main__': + if len(sys.argv) <= 6: + raise RuntimeError('arguments must greater equal than 6') + rep_cfg = {} + rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] + rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] + cfg_dir = {} + cfg_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] + cfg_dir[const_var.CFG_OUT_DIR] = sys.argv[5] + cfg_dir[const_var.AUTO_GEN_DIR] = sys.argv[6] + write_scripts(cfgfile=sys.argv[1], cfgs=rep_cfg, dirs=cfg_dir) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_ops_config.py b/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_ops_config.py new file mode 100644 index 000000000..7a97180be --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_ops_config.py @@ -0,0 +1,114 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import glob +import json +import argparse +import const_var + + +def load_json(json_file: str): + with open(json_file, encoding='utf-8') as file: + json_content = json.load(file) + return json_content + + +def get_specified_suffix_file(root_dir, suffix): + specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) + all_suffix_files = glob.glob(specified_suffix, recursive=True) + return all_suffix_files + + +def add_simplified_config(op_type, key, core_type, objfile, config): + simple_cfg = config.get('binary_info_config.json') + op_cfg = simple_cfg.get(op_type) + if not op_cfg: + op_cfg = {} + op_cfg['dynamicRankSupport'] = True + op_cfg['simplifiedKeyMode'] = 0 + op_cfg['binaryList'] = [] + simple_cfg[op_type] = op_cfg + bin_list = op_cfg.get('binaryList') + bin_list.append({'coreType': core_type, 'simplifiedKey': key, 'binPath': objfile}) + + +def add_op_config(op_file, bin_info, config): + op_cfg = config.get(op_file) + if not op_cfg: + op_cfg = {} + op_cfg['binList'] = [] + config[op_file] = op_cfg + op_cfg.get('binList').append(bin_info) + + +def gen_ops_config(json_file, soc, config): + core_type_map = {"MIX": 0, "AiCore": 1, "VectorCore": 2} + contents = load_json(json_file) + if ('binFileName' not in contents) or ('supportInfo' not in contents): + return + json_base_name = os.path.basename(json_file) + op_dir = os.path.basename(os.path.dirname(json_file)) + support_info = contents.get('supportInfo') + bin_name = contents.get('binFileName') + bin_suffix = contents.get('binFileSuffix') + core_type = core_type_map.get(contents.get("coreType")) + bin_file_name = bin_name + bin_suffix + op_type = bin_name.split('_')[0] + op_file = op_dir + '.json' + bin_info = {} + keys = support_info.get('simplifiedKey') + if keys: + bin_info['simplifiedKey'] = keys + for key in keys: + add_simplified_config(op_type, key, core_type, os.path.join(soc, op_dir, bin_file_name), config) + bin_info['staticKey'] = support_info.get('staticKey') + bin_info['int64Mode'] = support_info.get('int64Mode') + bin_info['inputs'] = support_info.get('inputs') + bin_info['outputs'] = support_info.get('outputs') + if support_info.get('attrs'): + bin_info['attrs'] = support_info.get('attrs') + bin_info['binInfo'] = {'jsonFilePath': os.path.join(soc, op_dir, json_base_name)} + add_op_config(op_file, bin_info, config) + + +def gen_all_config(root_dir, soc): + suffix = 'json' + config = {} + config['binary_info_config.json'] = {} + all_json_files = get_specified_suffix_file(root_dir, suffix) + for _json in all_json_files: + gen_ops_config(_json, soc, config) + for cfg_key in config.keys(): + cfg_file = os.path.join(root_dir, cfg_key) + with os.fdopen(os.open(cfg_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + json.dump(config.get(cfg_key), fd, indent=' ') + + +def args_prase(): + parser = argparse.ArgumentParser() + parser.add_argument('-p', + '--path', + nargs='?', + required=True, + help='Parse the path of the json file.') + parser.add_argument('-s', + '--soc', + nargs='?', + required=True, + help='Parse the soc_version of ops.') + return parser.parse_args() + + +def main(): + args = args_prase() + gen_all_config(args.path, args.soc) + + +if __name__ == '__main__': + main() diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_replay_build.py b/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_replay_build.py new file mode 100644 index 000000000..1cac7d911 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_replay_build.py @@ -0,0 +1,65 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import opdesc_parser +import replay_codegen +import const_var +from replay_codegen import ReplayCodeGenParams + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + + +class ReplayBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + super().__init__(op_type) + + def gen_replay_source(self: any, impl_path: str, out_path: str, ops_product: str): + if not self.op_replay_flag: + print('{} replay not enabled'.format(self.op_type)) + return + argn = len(self.input_name) + len(self.output_name) + 1 + if self.op_replay_batch: + print('{} replay in batch mode'.format(self.op_type)) + else: + print('{} replay in normal mode'.format(self.op_type)) + if impl_path.endswith('op_kernel'): + implf = os.path.join(impl_path, self.op_file + '.cpp') + tiling_file = os.path.join(impl_path, "../op_host", self.op_file + '_tiling.h') + else: + if self.dynamic_shape: + dyn_path = 'dynamic' + else: + dyn_path = '' + implf = os.path.join(impl_path, dyn_path, self.op_file + '.cpp') + tiling_file = os.path.join(impl_path, "../../op_tiling", self.op_file + '_tiling.h') + rep_conf = replay_codegen.ReplayCodeGen(ReplayCodeGenParams(self.op_type, implf, tiling_file, self.op_file, \ + self.op_intf, argn, self.op_replay_batch, self.max_block_dim, self.max_shape_size)) + rep_conf.set_batch(self.op_replay_batch) + rep_conf.set_outdir(out_path) + rep_conf.gen_replay(ops_product) + + +def gen_replay(cfgfile: str, cfgs: dict, dirs: dict, ops_product: str, ops: list = None): + batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') + iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') + op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, ReplayBuilder, ops) + for op_desc in op_descs: + op_desc.gen_replay_source(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), ops_product) + + +if __name__ == '__main__': + if len(sys.argv) <= 6: + raise RuntimeError('arguments must greater than 6') + rep_cfg = {} + rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] + rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] + rep_dir = {} + rep_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] + rep_dir[const_var.CFG_OUT_DIR] = sys.argv[5] + gen_replay(sys.argv[1], rep_cfg, rep_dir, sys.argv[6]) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/batch_replay_impl.temp b/atb_operator/AtbAddSample/AddOp/cmake/util/batch_replay_impl.temp new file mode 100644 index 000000000..0e8834664 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/batch_replay_impl.temp @@ -0,0 +1,117 @@ +#include +#include +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#include "register/op_check.h" +#define __ASCENDC_REPLAY_CODE__ +#include + +using namespace std; +using namespace optiling; +using namespace AscendCReplay; + +extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); +extern "C" int elf_batch_append(char *elf, uint32_t elfSize, char *jit, int kernum, char *atext[], int alen[], + int atlen, const char* kernelname[]); + +#define KERNEL_N 1 +#define ARG_N (__ARG_NUM__) +#define MAX_L (1024 * 1024 * 100) +#define MAX_E (1024 * 1024) + +int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) +{ + // gen type 1 : direct call codes 0: load .o file + if (param.gentype < 0 || param.gentype > 1) { + printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); + return 0; + } else if (param.gentype == 1 && param.objptr == nullptr) { + printf("Error: call replay with direct call mode, but code obj addr is null\n"); + return 0; + } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { + printf("Error: call replay with object file mode, but object file path is null\n"); + return 0; + } + // core_type 0:MIX 1:CUBE 2:VEC + if (core_type < 0 || core_type > 2) { + printf("Error: call replay core type is %d !\n", core_type); + return 0; + } + g_coreType = __CORE_TYPE__; + g_taskRation = param.task_ration; + g_tilingKey = param.tiling_key; + + unsigned char *buf, *jit; + char *kernel[KERNEL_N]; + int len[KERNEL_N]; + block_idx = 0; + block_num = param.block_dim; + g_ubBase = block_num; + uint8_t *code = (uint8_t *)malloc(MAX_L); + uint8_t *pos = code; + struct timespec tp1, tp2; + + clock_gettime(CLOCK_MONOTONIC, &tp1); + if (block_num > 32) { + printf("Error: block_num > 32\n"); + return 0; + } + //__OP_FOPEN__ + for (int i = 0; i < KERNEL_N; i++) { + //__OP_SET_KERNEL__ + for (int j = 0; j < ARG_N; j++) + AddArg(j, ARG_STEP * (j + 1)); +#ifdef FP_CEILING + SetCtrlFloatEnable(); +#else + SetCtrlFloatDisable(); +#endif + CodeInit(pos, true); + __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); + CodeEnd(); + kernel[i] = (char *)pos; + len[i] = CodeLen(); + pos += len[i]; + } + //__OP_FCLOSE__ + clock_gettime(CLOCK_MONOTONIC, &tp2); + buf = (unsigned char *)malloc(MAX_E); + int fd = open(param.entry_file, O_RDONLY); + if (fd < 0) { + printf("[error]: cannot find entry.o : %s\n", param.entry_file); + return 0; + } + uint32_t bufSize = read(fd, buf, MAX_E); + if (bufSize <= 0) { + printf("[error]: entry.o : %s is too small ! \n", param.entry_file); + } + close(fd); + jit = (unsigned char *)malloc(MAX_L); + printf("total code generated %ld\n", pos - code); + int sz = elf_batch_append((char *)buf, bufSize, (char *)jit, KERNEL_N, kernel, len, pos - code, ¶m.kernel_name); + if (tp1.tv_sec != tp2.tv_sec) { + printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); + } else { + printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); + } + printf("new elf size %d\n", sz); + if (param.gentype == 0) { + fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); + (void)write(fd, jit, sz); + close(fd); + free(jit); + } else if (param.gentype == 1) { + *param.objptr = (char*)jit; + } + free(buf); + free(code); + return sz; +} + +REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/code_channel_infer.py b/atb_operator/AtbAddSample/AddOp/cmake/util/code_channel_infer.py new file mode 100644 index 000000000..137c9f39f --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/code_channel_infer.py @@ -0,0 +1,115 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" +import os +import stat +import ctypes +import collections +import shutil +import subprocess +import copy + +"""CODE_* is used to cube/vector api is called in operator code +CODE_MIX means both cube and vector api is called +CODE_CUBE means only cube api is called +CODE_VEC means only vector api is called +""" +CODE_MIX = 0 +CODE_CUBE = 1 +CODE_VEC = 2 + + +def _is_v220(op_product: str): + """return if current soc version is V220 + + Returns: + res: True means V220 + """ + if op_product in ["ascend910b"]: + return True + return False + + +InfoCodeChanelParams = collections.namedtuple('InfoCodeChanelParams',\ +['src_file', 'tiling_header', 'kernel_name', 'outdir', 'op_product', 'compile_options']) + + +def infer_code_channel(params: InfoCodeChanelParams): + """get code channel for v220, return CODE_MIX if soc version is not V220 + + Args: + src_file (str): AscendC operator code file + src_file (str): AscendC operator tiling header file + kernel_name (str): kernel function name + optype (str): operator type + compile_options (list): compile options for ccec cmd + + Raises: + Exception: if not exist L1/L0/UB if code, it's not a aicore code + + Returns: + res (int): CODE_MIX/CODE_CUBE/CODE_VEC + """ + if not _is_v220(params.op_product): + return CODE_MIX + return CODE_VEC + if params.compile_options is None: + compile_options = [] + else: + compile_options = params.compile_options + ccec = shutil.which("ccec") + if ccec is not None: + ccec_path = os.path.dirname(ccec) + tikcpp_path = os.path.realpath(os.path.join(ccec_path, "..", "..", "tikcpp")) + else: + tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") + compile_options.append("-I" + tikcpp_path) + compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) + compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) + compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) + compile_options += ["-include", params.tiling_header] + arch = "dav-c220-cube" + sub_core_type = "AIC" + optional_core = "AiCore" + compile_cmd = [shutil.which("ccec"), '-c', '-O3'] + compile_cmd += compile_options + temp_file_name_tag = "_" + str(os.getpid()) + "_temp.o" + dst_file = os.path.join(kernel_meta_dir, kernel_name + temp_file_name_tag) + compile_cmd += [params.src_file, "--cce-aicore-arch={}".format(arch), + "--cce-aicore-only", "-o", dst_file, + "-mllvm", "-cce-aicore-function-stack-size=16000", + "-mllvm", "-cce-aicore-record-overflow=true", + "-mllvm", "-cce-aicore-addr-transform"] + compile_cmd += ["-std=c++17"] + print('get_code_channel: ', ' '.join(compile_cmd)) + proc = subprocess.Popen( + compile_cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) + (out, _) = proc.communicate() + if proc.returncode != 0: + print('get_code_channel coretype compile error: ', out.decode()) + msg = "compile %s error :%s\n" % (params.src_file, out.decode()) + raise Exception(f"get_code_channel coretype error, msg is{msg}") + objdump_cmd = ['objdump', '-s', '-j', '.text', '{}'.format(dst_file)] + + proc = subprocess.Popen( + objdump_cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) + (out, _) = proc.communicate() + if proc.returncode != 0: + print('get_code_channel objdump error: ', out.decode()) + msg = "get_code_channel objdump %s error :%s\n" % (src_file, out.decode()) + raise Exception(f"get_code_channel objdump error, msg is{msg}") + os.remove(dst_file) + lines = out.decode('utf-8').split('\n') + for line in lines: + insts = line.strip().split() + if len(insts) < 5: + continue + for inst in insts[1:]: + if len(inst) != 8: + continue + if inst[6] == 'f': + return CODE_MIX + return CODE_VEC diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/const_var.py b/atb_operator/AtbAddSample/AddOp/cmake/util/const_var.py new file mode 100644 index 000000000..8b32c3b91 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/const_var.py @@ -0,0 +1,33 @@ + +#!/usr/bin/env python +# coding=utf-8 +""" +Function: +The replay funtion entry +Copyright Information: +Huawei Technologies Co., Ltd. All Rights Reserved © 2020 +""" + +import os +import stat + + +REPLAY_BATCH = 'batch' +REPLAY_ITERATE = 'iterate' +CFG_IMPL_DIR = 'impl_dir' +CFG_OUT_DIR = 'out_dir' +AUTO_GEN_DIR = 'auto_gen_dir' +WFLAGS = os.O_WRONLY | os.O_CREAT | os.O_TRUNC +WMODES = stat.S_IWUSR | stat.S_IRUSR +SOC_MAP_EXT = {'ascend310p': 'Ascend310P3', 'ascend310b': 'Ascend310B1', + 'ascend910': 'Ascend910A', 'ascend910b': 'Ascend910B1'} +BIN_CMD = 'opc $1 --main_func={fun} --input_param={param} --soc_version={soc} \ +--output=$2 --impl_mode={impl} --simplified_key_mode=0 --op_mode=dynamic\n' +CHK_CMD = ''' +if ! test -f $2/{res_file} ; then + echo "$2/{res_file} not generated!" + exit 1 +fi +''' +ATTR_DEF_VAL = {'str' : '', 'int': 0, 'float': 0.0, 'bool': False, 'list_bool': [], + 'list_int': [], 'list_float': [], 'list_list_int': [[]]} diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/gen_impl_and_mrege_json.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/gen_impl_and_mrege_json.sh new file mode 100644 index 000000000..55e12e5ed --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/gen_impl_and_mrege_json.sh @@ -0,0 +1,57 @@ +#!/usr/bin/bash + +project_path=$1 +build_path=$2 +vendor_name=customize +if [[ ! -d "$project_path" ]]; then + echo "[ERROR] No projcet path is provided" + exit 1 +fi + +if [[ ! -d "$build_path" ]]; then + echo "[ERROR] No build path is provided" + exit 1 +fi + +# copy ai_core operators implements +# tbe_impl_files_num=$(ls $project_path/tbe/impl/* 2> /dev/null | wc -l) +# if [[ "$tbe_impl_files_num" -gt 0 ]];then +# cp -r ${project_path}/tbe/impl/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/ai_core/tbe/customize_impl +# cp -r ${project_path}/tbe/impl/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/vector_core/tbe/customize_impl +# fi + +# copy aicpu kernel so operators +if [[ -d "${project_path}/cpukernel/aicpu_kernel_lib" ]]; then + cp -f ${project_path}/cpukernel/aicpu_kernel_lib/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/cpu/aicpu_kernel/impl + rm -rf ${project_path}/cpukernel/aicpu_kernel_lib +fi + +# merge aicpu.ini and aicore.ini to generate npu_supported_ops.json +# mkdir -p ${build_path}/framework/op_info_cfg +# mkdir -p ${build_path}/framework/op_info_cfg/aicpu_kernel +# mkdir -p ${build_path}/framework/op_info_cfg/ai_core + +# if [[ -d "${project_path}/tbe/op_info_cfg/ai_core" ]]; then +# bash ${project_path}/cmake/util/gen_ops_filter.sh ${project_path}/tbe/op_info_cfg/ai_core ${build_path}/framework/op_info_cfg/ai_core +# fi + +# if [[ -d "${project_path}/cpukernel/op_info_cfg/aicpu_kernel" ]]; then +# bash ${project_path}/cmake/util/gen_ops_filter.sh ${project_path}/cpukernel/op_info_cfg/aicpu_kernel ${build_path}/framework/op_info_cfg/aicpu_kernel +# fi + +# aicpu_filter_file=${build_path}/framework/op_info_cfg/aicpu_kernel/npu_supported_ops.json +# aicore_filter_file=${build_path}/framework/op_info_cfg/ai_core/npu_supported_ops.json +# if [[ -f "${aicpu_filter_file}" ]] && [[ ! -f "${aicore_filter_file}" ]]; then +# cp $aicpu_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow +# fi +# if [[ -f "${aicore_filter_file}" ]] && [[ ! -f "${aicpu_filter_file}" ]]; then +# cp $aicore_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow +# fi + +# if [[ -f "${aicore_filter_file}" ]] && [[ -f "${aicpu_filter_file}" ]]; then +# chmod u+w ${aicpu_filter_file} +# python3 ${project_path}/cmake/util/insert_op_info.py ${aicore_filter_file} ${aicpu_filter_file} +# chmod u-w ${aicpu_filter_file} +# cp $aicpu_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow +# fi + diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/gen_ops_filter.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/gen_ops_filter.sh new file mode 100644 index 000000000..d4c27d17f --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/gen_ops_filter.sh @@ -0,0 +1,62 @@ +#!/bin/bash +# Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +# Description: Generate npu_supported_ops.json +# ============================================================================== + +if [[ -z "$1" ]]; then + echo -e "[ERROR] No source dir provided" + exit 1 +fi + +if [[ -z "$2" ]]; then + echo -e "[ERROR] No destination dir provided" + exit 1 +fi + +src=$1 +dest_file=$2/npu_supported_ops.json + +if [ -f "$dest_file" ];then + chmod u+w $dest_file +fi + +echo $* + +add_ops() { + name=$1 + isHeavy=$2 + file=$3 + grep -w "\"$name\"" ${file} >/dev/null + if [ $? == 0 ];then + return + fi + echo " \"${name}\": {" >> ${file} + echo " \"isGray\": false," >> ${file} + echo " \"isHeavy\": ${isHeavy}" >> ${file} + echo " }," >> ${file} +} + +echo "{" > ${dest_file} +ini_files=$(find ${src} -name "*.ini") +for file in ${ini_files} ; do + name=$(grep '^\[' ${file} | sed 's/\[//g' | sed 's/]//g' | sed 's/\r//g') + grep 'heavyOp.flag' ${file} >/dev/null + if [ $? == 0 ];then + isHeavy=$(grep 'heavyOp.flag' ${file} | awk -F= '{print $2}') + else + isHeavy="false" + fi + for op in ${name} ; do + add_ops ${op} "false" ${dest_file} + done +done +echo "}" >> ${dest_file} +file_count=$(cat ${dest_file} | wc -l) +line=$(($file_count-1)) +sed -i "${line}{s/,//g}" ${dest_file} + +chmod 640 "${dest_file}" +echo -e "[INFO] Succed generated ${dest_file}" + +exit 0 + diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/gen_version_info.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/gen_version_info.sh new file mode 100644 index 000000000..a06cfc78d --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/gen_version_info.sh @@ -0,0 +1,6 @@ +ascend_install_dir=$1 +gen_file_dir=$2 + +# create version.info +compiler_version=$(grep "Version" -w ${ascend_install_dir}/compiler/version.info | awk -F = '{print $2}') +echo "custom_opp_compiler_version=${compiler_version}" > ${gen_file_dir}/version.info \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/insert_op_info.py b/atb_operator/AtbAddSample/AddOp/cmake/util/insert_op_info.py new file mode 100644 index 000000000..28ba08757 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/insert_op_info.py @@ -0,0 +1,36 @@ +# -*- coding: utf-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" +import json +import os +import sys +import stat +import const_var + + +if __name__ == '__main__': + if len(sys.argv) != 3: + print(sys.argv) + print('argv error, inert_op_info.py your_op_file lib_op_file') + sys.exit(2) + + with open(sys.argv[1], 'r') as load_f: + insert_operator = json.load(load_f) + + all_operators = {} + if os.path.exists(sys.argv[2]): + if os.path.getsize(sys.argv[2]) != 0: + with open(sys.argv[2], 'r') as load_f: + all_operators = json.load(load_f) + + for k in insert_operator.keys(): + if k in all_operators.keys(): + print('replace op:[', k, '] success') + else: + print('insert op:[', k, '] success') + all_operators[k] = insert_operator[k] + + with os.fdopen(os.open(sys.argv[2], const_var.WFLAGS, const_var.WMODES), 'w') as json_file: + json_file.write(json.dumps(all_operators, indent=4)) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/insert_simplified_keys.py b/atb_operator/AtbAddSample/AddOp/cmake/util/insert_simplified_keys.py new file mode 100644 index 000000000..ace727b90 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/insert_simplified_keys.py @@ -0,0 +1,248 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import re +import glob +import json +import argparse +import const_var + + +DATA_TPYE_DICT = { + 'float32': 0, + 'float16': 1, + 'int8': 2, + 'int16': 6, + 'uint16': 7, + 'uint8': 4, + 'int32': 3, + 'int64': 9, + 'uint32': 8, + 'uint64': 10, + 'bool': 12, + 'double': 11, + 'string': 13, + 'dual': 14, + 'dual': 15, + 'complex64': 16, + 'complex128': 17, + 'qint8': 18, + 'qint16': 19, + 'qint32': 20, + 'quint8': 21, + 'quint16': 22, + 'resource': 23, + 'string': 24, + 'dual': 25, + 'variant': 26, + 'bf16': 27, + 'bfloat16': 27, + 'undefined': 28, + 'int4': 29, + 'uint1': 30, + 'int2': 31 +} + +FORMAT_DICT = { + 'NCHW': 0, + 'NHWC': 1, + 'ND': 2, + 'NC1HWC0': 3, + 'FRACTAL_Z': 4, + 'NC1C0HWPAD': 5, + 'NHWC1C0': 6, + 'FSR_NCHW': 7, + 'FRACTAL_DECONV': 8, + 'C1HWNC0': 9, + 'FRACTAL_DECONV_TRANSPOSE': 10, + 'FRACTAL_DECONV_SP_STRIDE_TRANS': 11, + 'NC1HWC0_C04': 12, + 'FRACTAL_Z_C04': 13, + 'CHWN': 14, + 'FRACTAL_DECONV_SP_STRIDE8_TRANS': 15, + 'HWCN': 16, + 'NC1KHKWHWC0': 17, + 'BN_WEIGHT': 18, + 'FILTER_HWCK': 19, + 'HASHTABLE_LOOKUP_LOOKUPS': 20, + 'HASHTABLE_LOOKUP_KEYS': 21, + 'HASHTABLE_LOOKUP_VALUE': 22, + 'HASHTABLE_LOOKUP_OUTPUT': 23, + 'HASHTABLE_LOOKUP_HITS': 24, + 'C1HWNCoC0': 25, + 'MD': 26, + 'NDHWC': 27, + 'FRACTAL_ZZ': 28, + 'FRACTAL_NZ': 29, + 'NCDHW': 30, + 'DHWCN': 31, + 'NDC1HWC0': 32, + 'FRACTAL_Z_3D': 33, + 'CN': 34, + 'NC': 35, + 'DHWNC': 36, + 'FRACTAL_Z_3D_TRANSPOSE': 37, + 'FRACTAL_ZN_LSTM': 38, + 'FRACTAL_Z_G': 39, + 'RESERVED': 40, + 'ALL': 41, + 'NULL': 42, + 'ND_RNN_BIAS': 43, + 'FRACTAL_ZN_RNN': 44, + 'NYUV': 45, + 'NYUV_A': 46 +} + + +def load_json(json_file: str): + with open(json_file, encoding='utf-8') as file: + json_content = json.load(file) + return json_content + + +def get_specified_suffix_file(root_dir, suffix): + specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) + all_suffix_files = glob.glob(specified_suffix, recursive=True) + return all_suffix_files + + +def get_deterministic_value(support_info): + deterministic_key = 'deterministic' + if deterministic_key not in support_info: + return 0 + deterministic_value = support_info.get(deterministic_key) + if deterministic_value == 'true': + return 1 + else: + return 0 + + +def get_precision_value(support_info): + precision_key = 'implMode' + precision_value = support_info.get(precision_key) + if precision_value == 'high_performance': + _value = 1 + elif precision_value == 'high_precision': + _value = 2 + else: + _value = 0 + return _value + + +def get_overflow_value(support_info): + return 0 + + +def get_parameters(info): + if info: + if 'dtype' in info: + data_type = info['dtype'] + data_type_value = DATA_TPYE_DICT.get(data_type) + else: + data_type_value = 0 + if 'format' in info: + _format = info['format'] + _format_value = FORMAT_DICT.get(_format) + else: + _format_value = 0 + else: + data_type_value = 0 + _format_value = 0 + return str(data_type_value), str(_format_value) + + +def get_dynamic_parameters(info): + # 动态输入时只需获取第一个参数 + return get_parameters(info[0]) + + +def get_all_parameters(support_info, _type): + result_list = list() + info_lists = support_info.get(_type) + if info_lists: + for _info in info_lists: + # 输入为列表时是动态输入 + if isinstance(_info, (list, tuple)): + data_type_value, _format_value = get_dynamic_parameters(_info) + else: + data_type_value, _format_value = get_parameters(_info) + result_list.append("{},{}".format(data_type_value, _format_value)) + return result_list + + +def get_all_input_parameters(support_info): + result = get_all_parameters(support_info, 'inputs') + return '/'.join(result) + + +def insert_content_into_file(input_file, content): + with open(input_file, 'r+') as file: + lines = file.readlines() + for index, line in enumerate(lines): + match_result = re.search(r'"staticKey":', line) + if match_result: + count = len(line) - len(line.lstrip()) + new_content = "{}{}".format(' ' * count, content) + # 插入到前一行,防止插入最后时还需要考虑是否添加逗号 + lines.insert(index, new_content) + break + file.seek(0) + file.write(''.join(lines)) + + +def insert_simplified_keys(json_file): + contents = load_json(json_file) + # 不存在'binFileName'或者'supportInfo'字段时,非需要替换的解析json文件 + if ('binFileName' not in contents) or ('supportInfo' not in contents): + return + support_info = contents.get('supportInfo') + bin_file_name = contents.get('binFileName') + bin_suffix = contents.get('binFileSuffix') + # 'simplifiedKey'字段已经存在时,直接返回,不重复生成 + if 'simplifiedKey' in support_info: + return + op_type = bin_file_name.split('_')[0] + deterministic = str(get_deterministic_value(support_info)) + precision = str(get_precision_value(support_info)) + overflow = str(get_overflow_value(support_info)) + input_parameters = get_all_input_parameters(support_info) + key = '{}/d={},p={},o={}/{}/'.format( + op_type, + deterministic, + precision, + overflow, + input_parameters) + result = '"simplifiedKey": "' + key + '",\n' + insert_content_into_file(json_file, result) + + +def insert_all_simplified_keys(root_dir): + suffix = 'json' + all_json_files = get_specified_suffix_file(root_dir, suffix) + for _json in all_json_files: + insert_simplified_keys(_json) + + +def args_prase(): + parser = argparse.ArgumentParser() + parser.add_argument('-p', + '--path', + nargs='?', + required=True, + help='Parse the path of the json file.') + return parser.parse_args() + + +def main(): + args = args_prase() + insert_all_simplified_keys(args.path) + + +if __name__ == '__main__': + main() diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/kernel_entry.py b/atb_operator/AtbAddSample/AddOp/cmake/util/kernel_entry.py new file mode 100644 index 000000000..2b77c970d --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/kernel_entry.py @@ -0,0 +1,115 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + + +def gen_fun_def(title, kernel, argn, arg_type, arg_name): + entry = [] + entry.append(title) + entry.append(kernel) + entry.append('(') + args = [] + for i in range(0, argn): + args.append(arg_type + ' ' + arg_name + str(i)) + entry.append(', '.join(args)) + entry.append(')') + return ' '.join(entry) + + +def gen_batch_kernel_body(fname, argn, arg_name): + body = [] + body.append('{') + fun = [] + fun.append(fname) + fun.append('(') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + fun.append(', '.join(args)) + fun.append(');') + body.append(' '.join(fun)) + body.append('}') + return '\n'.join(body) + + +def gen_mc_kernel_body(kn, argn, arg_name, blknum): + body = [] + body.append('{') + body.append(' switch(block_idx) {') + for blk in range(0, blknum): + fun = [] + fun.append('{}_blk{:02d}'.format(kn, blk)) + fun.append('(') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + fun.append(', '.join(args)) + fun.append(')') + body.append(' case {}: {}; break;'.format(blk, ' '.join(fun))) + body.append(' default: break;') + body.append(' }') + body.append('}') + return '\n'.join(body) + + +def gen_proc_body(argn, arg_name): + body = [] + body.append('{') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + body.append('uint64_t __x = (uint64_t)' + ' + (uint64_t)'.join(args) + ';') + body.append('__asm__ ("NOP");') + body.append('__asm__ ("NOP");') + body.append('__asm__ ("NOP");') + body.append('}') + return '\n'.join(body) + + +def batch_code_gen(kn, argn, argt): + codes = [] + kernel_name = kn + proc_name = kernel_name + '_percore' + arg_num = int(argn) + data_type = argt + arg_type = '__gm__ ' + data_type + '* __restrict__' + arg_name = 'arg' + kernel_title = 'extern \"C\" __global__ __aicore__ void' + proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' + codes.append('#ifndef __aicore__') + codes.append('#define __aicore__ [aicore]') + codes.append('#endif') + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') + codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) + codes.append(gen_batch_kernel_body(proc_name, arg_num, arg_name)) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) + codes.append(gen_proc_body(arg_num, arg_name)) + return '\n'.join(codes) + '\n' + + +def mc_code_gen(kn, argn, argt, blknum): + codes = [] + kernel_name = kn + core_num = int(blknum) + arg_num = int(argn) + data_type = argt + arg_type = '__gm__ ' + data_type + '* __restrict__' + arg_name = 'arg' + kernel_title = 'extern \"C\" __global__ __aicore__ void' + proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' + codes.append('#ifndef __aicore__') + codes.append('#define __aicore__ [aicore]') + codes.append('#endif') + for i in range(0, core_num): + proc_name = '{}_blk{:02d}'.format(kernel_name, i) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') + codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) + codes.append(gen_mc_kernel_body(kernel_name, arg_num, arg_name, core_num)) + for i in range(0, core_num): + proc_name = '{}_blk{:02d}'.format(kernel_name, i) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) + codes.append(gen_proc_body(arg_num, arg_name)) + return '\n'.join(codes) + '\n' diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/kernel_impl.temp b/atb_operator/AtbAddSample/AddOp/cmake/util/kernel_impl.temp new file mode 100644 index 000000000..5079a1043 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/kernel_impl.temp @@ -0,0 +1,10 @@ +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#define __ASCENDC_REPLAY_CODE__ +#include "__CCE_FILE__" diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/COPYING b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/COPYING new file mode 100644 index 000000000..d159169d1 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/COPYING @@ -0,0 +1,339 @@ + GNU GENERAL PUBLIC LICENSE + Version 2, June 1991 + + Copyright (C) 1989, 1991 Free Software Foundation, Inc., + 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + Everyone is permitted to copy and distribute verbatim copies + of this license document, but changing it is not allowed. + + Preamble + + The licenses for most software are designed to take away your +freedom to share and change it. By contrast, the GNU General Public +License is intended to guarantee your freedom to share and change free +software--to make sure the software is free for all its users. This +General Public License applies to most of the Free Software +Foundation's software and to any other program whose authors commit to +using it. (Some other Free Software Foundation software is covered by +the GNU Lesser General Public License instead.) You can apply it to +your programs, too. + + When we speak of free software, we are referring to freedom, not +price. Our General Public Licenses are designed to make sure that you +have the freedom to distribute copies of free software (and charge for +this service if you wish), that you receive source code or can get it +if you want it, that you can change the software or use pieces of it +in new free programs; and that you know you can do these things. + + To protect your rights, we need to make restrictions that forbid +anyone to deny you these rights or to ask you to surrender the rights. +These restrictions translate to certain responsibilities for you if you +distribute copies of the software, or if you modify it. + + For example, if you distribute copies of such a program, whether +gratis or for a fee, you must give the recipients all the rights that +you have. You must make sure that they, too, receive or can get the +source code. And you must show them these terms so they know their +rights. + + We protect your rights with two steps: (1) copyright the software, and +(2) offer you this license which gives you legal permission to copy, +distribute and/or modify the software. + + Also, for each author's protection and ours, we want to make certain +that everyone understands that there is no warranty for this free +software. If the software is modified by someone else and passed on, we +want its recipients to know that what they have is not the original, so +that any problems introduced by others will not reflect on the original +authors' reputations. + + Finally, any free program is threatened constantly by software +patents. We wish to avoid the danger that redistributors of a free +program will individually obtain patent licenses, in effect making the +program proprietary. To prevent this, we have made it clear that any +patent must be licensed for everyone's free use or not licensed at all. + + The precise terms and conditions for copying, distribution and +modification follow. + + GNU GENERAL PUBLIC LICENSE + TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION + + 0. This License applies to any program or other work which contains +a notice placed by the copyright holder saying it may be distributed +under the terms of this General Public License. The "Program", below, +refers to any such program or work, and a "work based on the Program" +means either the Program or any derivative work under copyright law: +that is to say, a work containing the Program or a portion of it, +either verbatim or with modifications and/or translated into another +language. (Hereinafter, translation is included without limitation in +the term "modification".) Each licensee is addressed as "you". + +Activities other than copying, distribution and modification are not +covered by this License; they are outside its scope. The act of +running the Program is not restricted, and the output from the Program +is covered only if its contents constitute a work based on the +Program (independent of having been made by running the Program). +Whether that is true depends on what the Program does. + + 1. You may copy and distribute verbatim copies of the Program's +source code as you receive it, in any medium, provided that you +conspicuously and appropriately publish on each copy an appropriate +copyright notice and disclaimer of warranty; keep intact all the +notices that refer to this License and to the absence of any warranty; +and give any other recipients of the Program a copy of this License +along with the Program. + +You may charge a fee for the physical act of transferring a copy, and +you may at your option offer warranty protection in exchange for a fee. + + 2. You may modify your copy or copies of the Program or any portion +of it, thus forming a work based on the Program, and copy and +distribute such modifications or work under the terms of Section 1 +above, provided that you also meet all of these conditions: + + a) You must cause the modified files to carry prominent notices + stating that you changed the files and the date of any change. + + b) You must cause any work that you distribute or publish, that in + whole or in part contains or is derived from the Program or any + part thereof, to be licensed as a whole at no charge to all third + parties under the terms of this License. + + c) If the modified program normally reads commands interactively + when run, you must cause it, when started running for such + interactive use in the most ordinary way, to print or display an + announcement including an appropriate copyright notice and a + notice that there is no warranty (or else, saying that you provide + a warranty) and that users may redistribute the program under + these conditions, and telling the user how to view a copy of this + License. (Exception: if the Program itself is interactive but + does not normally print such an announcement, your work based on + the Program is not required to print an announcement.) + +These requirements apply to the modified work as a whole. If +identifiable sections of that work are not derived from the Program, +and can be reasonably considered independent and separate works in +themselves, then this License, and its terms, do not apply to those +sections when you distribute them as separate works. But when you +distribute the same sections as part of a whole which is a work based +on the Program, the distribution of the whole must be on the terms of +this License, whose permissions for other licensees extend to the +entire whole, and thus to each and every part regardless of who wrote it. + +Thus, it is not the intent of this section to claim rights or contest +your rights to work written entirely by you; rather, the intent is to +exercise the right to control the distribution of derivative or +collective works based on the Program. + +In addition, mere aggregation of another work not based on the Program +with the Program (or with a work based on the Program) on a volume of +a storage or distribution medium does not bring the other work under +the scope of this License. + + 3. You may copy and distribute the Program (or a work based on it, +under Section 2) in object code or executable form under the terms of +Sections 1 and 2 above provided that you also do one of the following: + + a) Accompany it with the complete corresponding machine-readable + source code, which must be distributed under the terms of Sections + 1 and 2 above on a medium customarily used for software interchange; or, + + b) Accompany it with a written offer, valid for at least three + years, to give any third party, for a charge no more than your + cost of physically performing source distribution, a complete + machine-readable copy of the corresponding source code, to be + distributed under the terms of Sections 1 and 2 above on a medium + customarily used for software interchange; or, + + c) Accompany it with the information you received as to the offer + to distribute corresponding source code. (This alternative is + allowed only for noncommercial distribution and only if you + received the program in object code or executable form with such + an offer, in accord with Subsection b above.) + +The source code for a work means the preferred form of the work for +making modifications to it. For an executable work, complete source +code means all the source code for all modules it contains, plus any +associated interface definition files, plus the scripts used to +control compilation and installation of the executable. However, as a +special exception, the source code distributed need not include +anything that is normally distributed (in either source or binary +form) with the major components (compiler, kernel, and so on) of the +operating system on which the executable runs, unless that component +itself accompanies the executable. + +If distribution of executable or object code is made by offering +access to copy from a designated place, then offering equivalent +access to copy the source code from the same place counts as +distribution of the source code, even though third parties are not +compelled to copy the source along with the object code. + + 4. You may not copy, modify, sublicense, or distribute the Program +except as expressly provided under this License. Any attempt +otherwise to copy, modify, sublicense or distribute the Program is +void, and will automatically terminate your rights under this License. +However, parties who have received copies, or rights, from you under +this License will not have their licenses terminated so long as such +parties remain in full compliance. + + 5. You are not required to accept this License, since you have not +signed it. However, nothing else grants you permission to modify or +distribute the Program or its derivative works. These actions are +prohibited by law if you do not accept this License. Therefore, by +modifying or distributing the Program (or any work based on the +Program), you indicate your acceptance of this License to do so, and +all its terms and conditions for copying, distributing or modifying +the Program or works based on it. + + 6. Each time you redistribute the Program (or any work based on the +Program), the recipient automatically receives a license from the +original licensor to copy, distribute or modify the Program subject to +these terms and conditions. You may not impose any further +restrictions on the recipients' exercise of the rights granted herein. +You are not responsible for enforcing compliance by third parties to +this License. + + 7. If, as a consequence of a court judgment or allegation of patent +infringement or for any other reason (not limited to patent issues), +conditions are imposed on you (whether by court order, agreement or +otherwise) that contradict the conditions of this License, they do not +excuse you from the conditions of this License. If you cannot +distribute so as to satisfy simultaneously your obligations under this +License and any other pertinent obligations, then as a consequence you +may not distribute the Program at all. For example, if a patent +license would not permit royalty-free redistribution of the Program by +all those who receive copies directly or indirectly through you, then +the only way you could satisfy both it and this License would be to +refrain entirely from distribution of the Program. + +If any portion of this section is held invalid or unenforceable under +any particular circumstance, the balance of the section is intended to +apply and the section as a whole is intended to apply in other +circumstances. + +It is not the purpose of this section to induce you to infringe any +patents or other property right claims or to contest validity of any +such claims; this section has the sole purpose of protecting the +integrity of the free software distribution system, which is +implemented by public license practices. Many people have made +generous contributions to the wide range of software distributed +through that system in reliance on consistent application of that +system; it is up to the author/donor to decide if he or she is willing +to distribute software through any other system and a licensee cannot +impose that choice. + +This section is intended to make thoroughly clear what is believed to +be a consequence of the rest of this License. + + 8. If the distribution and/or use of the Program is restricted in +certain countries either by patents or by copyrighted interfaces, the +original copyright holder who places the Program under this License +may add an explicit geographical distribution limitation excluding +those countries, so that distribution is permitted only in or among +countries not thus excluded. In such case, this License incorporates +the limitation as if written in the body of this License. + + 9. The Free Software Foundation may publish revised and/or new versions +of the General Public License from time to time. Such new versions will +be similar in spirit to the present version, but may differ in detail to +address new problems or concerns. + +Each version is given a distinguishing version number. If the Program +specifies a version number of this License which applies to it and "any +later version", you have the option of following the terms and conditions +either of that version or of any later version published by the Free +Software Foundation. If the Program does not specify a version number of +this License, you may choose any version ever published by the Free Software +Foundation. + + 10. If you wish to incorporate parts of the Program into other free +programs whose distribution conditions are different, write to the author +to ask for permission. For software which is copyrighted by the Free +Software Foundation, write to the Free Software Foundation; we sometimes +make exceptions for this. Our decision will be guided by the two goals +of preserving the free status of all derivatives of our free software and +of promoting the sharing and reuse of software generally. + + NO WARRANTY + + 11. BECAUSE THE PROGRAM IS LICENSED FREE OF CHARGE, THERE IS NO WARRANTY +FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN +OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES +PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED +OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF +MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS +TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE +PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, +REPAIR OR CORRECTION. + + 12. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING +WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MAY MODIFY AND/OR +REDISTRIBUTE THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, +INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING +OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED +TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY +YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER +PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE +POSSIBILITY OF SUCH DAMAGES. + + END OF TERMS AND CONDITIONS + + How to Apply These Terms to Your New Programs + + If you develop a new program, and you want it to be of the greatest +possible use to the public, the best way to achieve this is to make it +free software which everyone can redistribute and change under these terms. + + To do so, attach the following notices to the program. It is safest +to attach them to the start of each source file to most effectively +convey the exclusion of warranty; and each file should have at least +the "copyright" line and a pointer to where the full notice is found. + + + Copyright (C) + + This program is free software; you can redistribute it and/or modify + it under the terms of the GNU General Public License as published by + the Free Software Foundation; either version 2 of the License, or + (at your option) any later version. + + This program is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + GNU General Public License for more details. + + You should have received a copy of the GNU General Public License along + with this program; if not, write to the Free Software Foundation, Inc., + 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. + +Also add information on how to contact you by electronic and paper mail. + +If the program is interactive, make it output a short notice like this +when it starts in an interactive mode: + + Gnomovision version 69, Copyright (C) year name of author + Gnomovision comes with ABSOLUTELY NO WARRANTY; for details type `show w'. + This is free software, and you are welcome to redistribute it + under certain conditions; type `show c' for details. + +The hypothetical commands `show w' and `show c' should show the appropriate +parts of the General Public License. Of course, the commands you use may +be called something other than `show w' and `show c'; they could even be +mouse-clicks or menu items--whatever suits your program. + +You should also get your employer (if you work as a programmer) or your +school, if any, to sign a "copyright disclaimer" for the program, if +necessary. Here is a sample; alter the names: + + Yoyodyne, Inc., hereby disclaims all copyright interest in the program + `Gnomovision' (which makes passes at compilers) written by James Hacker. + + , 1 April 1989 + Ty Coon, President of Vice + +This General Public License does not permit incorporating your program into +proprietary programs. If your program is a subroutine library, you may +consider it more useful to permit linking proprietary applications with the +library. If this is what you want to do, use the GNU Lesser General +Public License instead of this License. diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/README.md b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/README.md new file mode 100644 index 000000000..b41f01682 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/README.md @@ -0,0 +1,246 @@ +[![License: GPL v2](https://img.shields.io/badge/License-GPL%20v2-blue.svg)](https://www.gnu.org/licenses/old-licenses/gpl-2.0.en.html) +![Build Status](https://github.com/megastep/makeself/workflows/CI/badge.svg) + +# makeself - Make self-extractable archives on Unix + +[makeself.sh][1] is a small shell script that generates a self-extractable +compressed tar archive from a directory. The resulting file appears as a shell script +(many of those have a **.run** suffix), and can be launched as is. The archive +will then uncompress itself to a temporary directory and an optional arbitrary +command will be executed (for example an installation script). This is pretty +similar to archives generated with WinZip Self-Extractor in the Windows world. +Makeself archives also include checksums for integrity self-validation (CRC +and/or MD5/SHA256 checksums). + +The makeself.sh script itself is used only to create the archives from a +directory of files. The resultant archive is actually a compressed (using +gzip, bzip2, or compress) TAR archive, with a small shell script stub at the +beginning. This small stub performs all the steps of extracting the files, +running the embedded command, and removing the temporary files when done. +All the user has to do to install the software contained in such an +archive is to "run" the archive, i.e **sh nice-software.run**. I recommend +using the ".run" (which was introduced by some Makeself archives released by +Loki Software) or ".sh" suffix for such archives not to confuse the users, +so that they will know they are actually shell scripts (with quite a lot of binary data +attached to them though!). + +I am trying to keep the code of this script as portable as possible, i.e it is +not relying on any bash-specific features and only calls commands that are +installed on any functioning UNIX-compatible system. This script as well as +the archives it generates should run on any Unix flavor, with any compatible +Bourne shell, provided of course that the compression programs are available. + +As of version 2.1, Makeself has been rewritten and tested on the following +platforms : + + * Linux (all distributions) + * Sun Solaris (8 and above) + * HP-UX (tested on 11.0 and 11i on HPPA RISC) + * SCO OpenUnix and OpenServer + * IBM AIX 5.1L + * macOS (Darwin) + * SGI IRIX 6.5 + * FreeBSD + * UnicOS / Cray + * Cygwin (Windows) + +If you successfully run Makeself and/or archives created with it on another +system, then please [let me know][2]! + +Examples of publicly available archives made using makeself are : + + * Game patches and installers for [Id Software][3] games like Quake 3 for Linux or Return To Castle Wolfenstein ; + * All game patches released by [Loki Software][4] for the Linux version of popular games ; + * The [nVidia drivers][5] for Linux + * The installer for the Linux version of [Google Earth][6] + * The [VirtualBox][7] installers for Linux + * The [Makeself][1] distribution itself ;-) + * and countless others... + +**Important note for Apache users:** By default, most Web servers will think that Makeself archives are regular text files and thus they may show up as text in a Web browser. The correct way to prevent this is to add a MIME type for this file format, like so (in httpd.conf) : + +`AddType application/x-makeself .run` + +**Important note for certain GNU/Linux distributions:** Archives created with Makeself prior to v2.1.2 were using an old syntax for the _head_ and _tail_ Unix commands that is being progressively obsoleted in their GNU forms. Therefore you may have problems uncompressing some of these archives. A workaround for this is to set the environment variable $_POSIX2_VERSION to enable the old syntax, i.e. : + +`export _POSIX2_VERSION=199209` + +## Usage + +The syntax of makeself is the following: + +``` +makeself.sh [args] archive_dir file_name label startup_script [script_args] +``` + + * _args_ are optional options for Makeself. The available ones are : + + * **`--version`** : Prints the version number on stdout, then exits immediately + * **`--gzip`** : Use gzip for compression (the default on platforms on which gzip is commonly available, like Linux) + * **`--bzip2`** : Use bzip2 instead of gzip for better compression. The bzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. + * **`--pbzip2`** : Use pbzip2 instead of gzip for better and faster compression on machines having multiple CPUs. The pbzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. + * **`--xz`** : Use xz instead of gzip for better compression. The xz command must be available in the command path. It is recommended that the archive prefix be set to something like '.xz.run' for the archive, so that potential users know that they'll need xz to extract it. + * **`--lzo`** : Use lzop instead of gzip for better compression. The lzop command must be available in the command path. It is recommended that the archive prefix be set to something like `.lzo.run` for the archive, so that potential users know that they'll need lzop to extract it. + * **`--lz4`** : Use lz4 instead of gzip for better compression. The lz4 command must be available in the command path. It is recommended that the archive prefix be set to something like '.lz4.run' for the archive, so that potential users know that they'll need lz4 to extract it. + * **`--zstd`** : Use zstd instead of gzip for better compression. The zstd command must be available in the command path. It is recommended that the archive prefix be set to something like '.zstd.run' for the archive, so that potential users know that they'll need zstd to extract it. + * **`--pigz`** : Use pigz for compression. + * **`--base64`** : Encode the archive to ASCII in Base64 format instead of compressing (base64 command required). + * **`--gpg-encrypt`** : Encrypt the archive using `gpg -ac -z $COMPRESS_LEVEL`. This will prompt for a password to encrypt with. Assumes that potential users have `gpg` installed. + * **`--ssl-encrypt`** : Encrypt the archive using `openssl aes-256-cbc -a -salt`. This will prompt for a password to encrypt with. Assumes that the potential users have the OpenSSL tools installed. + * **`--compress`** : Use the UNIX `compress` command to compress the data. This should be the default on all platforms that don't have gzip available. + * **`--nocomp`** : Do not use any compression for the archive, which will then be an uncompressed TAR. + * **`--complevel`** : Specify the compression level for gzip, bzip2, pbzip2, zstd, xz, lzo or lz4. (defaults to 9) + * **`--threads`** : Specify the number of threads to be used by compressors that support parallelization. Omit to use compressor's default. Most useful (and required) for opting into xz's threading, usually with `--threads=0` for all available cores. pbzip2 and pigz are parallel by default, and setting this value allows limiting the number of threads they use. + * **`--notemp`** : The generated archive will not extract the files to a temporary directory, but in a new directory created in the current directory. This is better to distribute software packages that may extract and compile by themselves (i.e. launch the compilation through the embedded script). + * **`--current`** : Files will be extracted to the current directory, instead of in a subdirectory. This option implies `--notemp` above. + * **`--follow`** : Follow the symbolic links inside of the archive directory, i.e. store the files that are being pointed to instead of the links themselves. + * **`--append`** _(new in 2.1.x)_: Append data to an existing archive, instead of creating a new one. In this mode, the settings from the original archive are reused (compression type, label, embedded script), and thus don't need to be specified again on the command line. + * **`--header`** : Makeself uses a separate file to store the header stub, called `makeself-header.sh`. By default, it is assumed that it is stored in the same location as makeself.sh. This option can be used to specify its actual location if it is stored someplace else. + * **`--cleanup`** : Specify a script that is run when execution is interrupted or finishes successfully. The script is executed with the same environment and initial `script_args` as `startup_script`. + * **`--copy`** : Upon extraction, the archive will first extract itself to a temporary directory. The main application of this is to allow self-contained installers stored in a Makeself archive on a CD, when the installer program will later need to unmount the CD and allow a new one to be inserted. This prevents "Filesystem busy" errors for installers that span multiple CDs. + * **`--nox11`** : Disable the automatic spawning of a new terminal in X11. + * **`--nowait`** : When executed from a new X11 terminal, disable the user prompt at the end of the script execution. + * **`--nomd5`** and **`--nocrc`** : Disable the creation of a MD5 / CRC checksum for the archive. This speeds up the extraction process if integrity checking is not necessary. + * **`--sha256`** : Adds a SHA256 checksum for the archive. This is in addition to the MD5 / CRC checksums unless `--nomd5` is also used. + * **`--lsm` _file_** : Provide and LSM file to makeself, that will be embedded in the generated archive. LSM files are describing a software package in a way that is easily parseable. The LSM entry can then be later retrieved using the `--lsm` argument to the archive. An example of a LSM file is provided with Makeself. + * **`--tar-format opt`** : Specify the tar archive format (default is ustar); you may use any value accepted by your tar command (such as posix, v7, etc). + * **`--tar-extra opt`** : Append more options to the tar command line. + + For instance, in order to exclude the `.git` directory from the packaged archive directory using the GNU `tar`, one can use `makeself.sh --tar-extra "--exclude=.git" ...` + + * **`--keep-umask`** : Keep the umask set to shell default, rather than overriding when executing self-extracting archive. + * **`--packaging-date date`** : Use provided string as the packaging date instead of the current date. + * **`--license`** : Append a license file. + * **`--nooverwrite`** : Do not extract the archive if the specified target directory already exists. + * **`--help-header file`** : Add a header to the archive's `--help` output. + * `archive_dir` is the name of the directory that contains the files to be archived + * `file_name` is the name of the archive to be created + * `label` is an arbitrary text string describing the package. It will be displayed while extracting the files. + * `startup_script` is the command to be executed _from within_ the directory of extracted files. Thus, if you wish to execute a program contained in this directory, you must prefix your command with `./`. For example, `./program` will be fine. The `script_args` are additional arguments for this command. + +Here is an example, assuming the user has a package image stored in a **/home/joe/mysoft**, and he wants to generate a self-extracting package named +**mysoft.sh**, which will launch the "setup" script initially stored in /home/joe/mysoft : + +`makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup +` + +Here is also how I created the [makeself.run][9] archive which contains the Makeself distribution : + +`makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" ` + +Archives generated with Makeself can be passed the following arguments: + + * **`--keep`** : Prevent the files to be extracted in a temporary directory that will be removed after the embedded script's execution. The files will then be extracted in the current working directory and will stay here until you remove them. + * **`--verbose`** : Will prompt the user before executing the embedded command + * **`--target dir`** : Allows to extract the archive in an arbitrary place. + * **`--nox11`** : Do not spawn a X11 terminal. + * **`--confirm`** : Prompt the user for confirmation before running the embedded command. + * **`--info`** : Print out general information about the archive (does not extract). + * **`--lsm`** : Print out the LSM entry, if it is present. + * **`--list`** : List the files in the archive. + * **`--check`** : Check the archive for integrity using the embedded checksums. Does not extract the archive. + * **`--nochown`** : By default, a `chown -R` command is run on the target directory after extraction, so that all files belong to the current user. This is mostly needed if you are running as root, as tar will then try to recreate the initial user ownerships. You may disable this behavior with this flag. + * **`--tar`** : Run the tar command on the contents of the archive, using the following arguments as parameter for the command. + * **`--noexec`** : Do not run the embedded script after extraction. + * **`--noexec-cleanup`** : Do not run the embedded cleanup script. + * **`--nodiskspace`** : Do not check for available disk space before attempting to extract. + * **`--cleanup-args`** : Specify arguments to be passed to the cleanup script. Wrap value in quotes to specify multiple arguments. + +Any subsequent arguments to the archive will be passed as additional arguments to the embedded command. You must explicitly use the `--` special command-line construct before any such options to make sure that Makeself will not try to interpret them. + +## Startup Script + +The startup script must be a regular Shell script. + +Within the startup script, you can use the `$USER_PWD` variable to get the path of the folder from which the self-extracting script is executed. This is especially useful to access files that are located in the same folder as the script, as shown in the example below. + +`my-self-extracting-script.sh --fooBarFileParameter foo.bar` + +## Building and Testing + +Clone the git repo and execute `git submodule update --init --recursive` to obtain all submodules. + +* To make a release: `make` +* To run all tests: `make test` + +## Maven Usage + +Makeself is now supported by the following maven plugin [makeself-maven-plugin](https://github.com/hazendaz/makeself-maven-plugin). Please refer to project for usage and report any bugs in regards to maven plugin on that project. + +## License + +Makeself itself is covered by the [GNU General Public License][8] (GPL) version 2 and above. Archives generated by Makeself don't have to be placed under this license (although I encourage it ;-)), since the archive itself is merely data for Makeself. + +## Contributing + +I will gladly consider merging your pull requests on the [GitHub][10] repository. However, please keep the following in mind: + + * One of the main purposes of Makeself is portability. Do not submit patches that will break supported platforms. The more platform-agnostic, the better. + * Please explain clearly what the purpose of the patch is, and how you achieved it. + +## Download + +Get the latest official distribution [here][9] (version 2.4.2). + +The latest development version can be grabbed from [GitHub][10]. Feel free to submit any patches there through the fork and pull request process. + +## Version history + + * **v1.0:** Initial public release + * **v1.1:** The archive can be passed parameters that will be passed on to the embedded script, thanks to John C. Quillan + * **v1.2:** Cosmetic updates, support for bzip2 compression and non-temporary archives. Many ideas thanks to Francois Petitjean. + * **v1.3:** More patches from Bjarni R. Einarsson and Francois Petitjean: Support for no compression (`--nocomp`), script is no longer mandatory, automatic launch in an xterm, optional verbose output, and -target archive option to indicate where to extract the files. + * **v1.4:** Many patches from Francois Petitjean: improved UNIX compatibility, automatic integrity checking, support of LSM files to get info on the package at run time.. + * **v1.5.x:** A lot of bugfixes, and many other patches, including automatic verification through the usage of checksums. Version 1.5.5 was the stable release for a long time, even though the Web page didn't get updated ;-). Makeself was also officially made a part of the [Loki Setup installer][11], and its source is being maintained as part of this package. + * **v2.0:** Complete internal rewrite of Makeself. The command-line parsing was vastly improved, the overall maintenance of the package was greatly improved by separating the stub from makeself.sh. Also Makeself was ported and tested to a variety of Unix platforms. + * **v2.0.1:** First public release of the new 2.0 branch. Prior versions are officially obsoleted. This release introduced the `--copy` argument that was introduced in response to a need for the [UT2K3][12] Linux installer. + * **v2.1.0:** Big change : Makeself can now support multiple embedded tarballs, each stored separately with their own checksums. An existing archive can be updated with the `--append` flag. Checksums are also better managed, and the `--nochown` option for archives appeared. + * **v2.1.1:** Fixes related to the Unix compression (compress command). Some Linux distributions made the insane choice to make it unavailable, even though gzip is capable of uncompressing these files, plus some more bugfixes in the extraction and checksum code. + * **v2.1.2:** Some bug fixes. Use head -n to avoid problems with POSIX conformance. + * **v2.1.3:** Bug fixes with the command line when spawning terminals. Added `--tar`, `--noexec` for archives. Added `--nomd5` and `--nocrc` to avoid creating checksums in archives. The embedded script is now run through "eval". The `--info` output now includes the command used to create the archive. A man page was contributed by Bartosz Fenski. + * **v2.1.4:** Fixed `--info` output. Generate random directory name when extracting files to . to avoid problems. Better handling of errors with wrong permissions for the directory containing the files. Avoid some race conditions, Unset the $CDPATH variable to avoid problems if it is set. Better handling of dot files in the archive directory. + * **v2.1.5:** Made the md5sum detection consistent with the header code. Check for the presence of the archive directory. Added `--encrypt` for symmetric encryption through gpg (Eric Windisch). Added support for the digest command on Solaris 10 for MD5 checksums. Check for available disk space before extracting to the target directory (Andreas Schweitzer). Allow extraction to run asynchronously (patch by Peter Hatch). Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo). + * **v2.1.6:** Replaced one dot per file progress with a realtime progress percentage and a spinning cursor. Added `--noprogress` to prevent showing the progress during the decompression. Added `--target` dir to allow extracting directly to a target directory. (Guy Baconniere) + * **v2.2.0:** First major new release in years! Includes many bugfixes and user contributions. Please look at the [project page on Github][10] for all the details. + * **v2.3.0:** Support for archive encryption via GPG or OpenSSL. Added LZO and LZ4 compression support. Options to set the packaging date and stop the umask from being overriden. Optionally ignore check for available disk space when extracting. New option to check for root permissions before extracting. + * **v2.3.1:** Various compatibility updates. Added unit tests for Travis CI in the GitHub repo. New `--tar-extra`, `--untar-extra`, `--gpg-extra`, `--gpg-asymmetric-encrypt-sign` options. + * **v2.4.0:** Added optional support for SHA256 archive integrity checksums. + * **v2.4.2:** New --cleanup and --cleanup-args arguments for cleanup scripts. Added threading support for supported compressors. Now supports zstd compression. + * **v2.4.3:** Make explicit POSIX tar archives for increased compatibility. + * **v2.4.4:** Fixed various compatibility issues (no longer use POSIX tar archives), Github Actions to check on Solaris and FreeBSD. + * **v2.4.5:** Added `--tar-format` option to set the tar archive format (default is ustar) + +## Links + + * Check out the ["Loki Setup"][11] installer, used to install many Linux games and other applications, and of which I am the co-author. Since the demise of Loki, I am now the official maintainer of the project, and it is now being hosted here on GitHub. + * Bjarni R. Einarsson also wrote the **setup.sh** installer script, inspired by Makeself. [Check it out !][14] + +## Contact + +This script was written by [Stéphane Peter][15] (megastep at megastep.org). Any enhancements and suggestions are welcome. + +Contributions were included from John C. Quillan, Bjarni R. Einarsson, +Francois Petitjean, Ryan C. Gordon, and many contributors on GitHub. If you think I forgot +your name, don't hesitate to contact me. + +This project is now hosted on GitHub. Feel free to submit patches and bug reports on the [project page][10]. + +* * * + +[Stephane Peter][2] + + [1]: http://makeself.io/ + [2]: mailto:megastep@megastep.org + [3]: http://www.idsoftware.com/ + [4]: http://www.lokigames.com/products/myth2/updates.php3 + [5]: http://www.nvidia.com/ + [6]: http://earth.google.com/ + [7]: http://www.virtualbox.org/ + [8]: http://www.gnu.org/copyleft/gpl.html + [9]: https://github.com/megastep/makeself/releases/download/release-2.4.5/makeself-2.4.5.run + [10]: https://github.com/megastep/makeself + [11]: https://github.com/megastep/loki_setup/ + [12]: http://www.unrealtournament2003.com/ + [13]: http://www.icculus.org/ + [14]: http://bre.klaki.net/programs/setup.sh/ + [15]: https://stephanepeter.com/ diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/VERSION b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/VERSION new file mode 100644 index 000000000..59aa62c1f --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/VERSION @@ -0,0 +1 @@ +2.4.5 diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/make-release.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/make-release.sh new file mode 100644 index 000000000..b5692d490 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/make-release.sh @@ -0,0 +1,9 @@ +#!/bin/sh +# +# Create a distributable archive of the current version of Makeself + +VER=`cat VERSION` +mkdir -p /tmp/makeself-$VER release +cp -pPR makeself* test README.md COPYING VERSION .gitmodules /tmp/makeself-$VER/ +./makeself.sh --notemp /tmp/makeself-$VER release/makeself-$VER.run "Makeself v$VER" echo "Makeself has extracted itself" + diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself-header.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself-header.sh new file mode 100644 index 000000000..940903148 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself-header.sh @@ -0,0 +1,660 @@ +cat << EOF > "$archname" +#!/bin/bash +# This script was generated using Makeself $MS_VERSION +# The license covering this archive and its contents, if any, is wholly independent of the Makeself license (GPL) +# 2022.3.19-Modified the MS_Help function and some options +# Huawei Technologies Co., Ltd. + +ORIG_UMASK=\`umask\` + +CRCsum="$CRCsum" +MD5="$MD5sum" +SHA="$SHAsum" +SIGNATURE="$Signature" +TMPROOT=\${TMPDIR:="\$HOME"} +if ! test -d "\$TMPROOT"; then + TMPROOT="\$PWD" +fi +export TMPDIR="\$TMPROOT" +USER_PWD="\$PWD" +if ! test -d "\$USER_PWD"; then + exit 1 +fi +export USER_PWD +ARCHIVE_DIR=\`dirname "\$0"\` +export ARCHIVE_DIR + +name_of_file="\$0 " +pwd_of_file="\$PWD" +label="$LABEL" +script="$SCRIPT" +scriptargs="$SCRIPTARGS" +cleanup_script="${CLEANUP_SCRIPT}" +licensetxt="$LICENSE" +helpheader='$HELPHEADER' +targetdir="$archdirname" +filesizes="$filesizes" +totalsize="$totalsize" +keep="$KEEP" +nooverwrite="$NOOVERWRITE" +quiet="n" +accept="n" +nodiskspace="n" +export_conf="$EXPORT_CONF" +decrypt_cmd="$DECRYPT_CMD" +skip="$SKIP" + +print_cmd_arg="" +if type printf > /dev/null; then + print_cmd="printf" +elif test -x /usr/ucb/echo; then + print_cmd="/usr/ucb/echo" +else + print_cmd="echo" +fi + +if test -d /usr/xpg4/bin; then + PATH=/usr/xpg4/bin:\$PATH + export PATH +fi + +if test -d /usr/sfw/bin; then + PATH=\$PATH:/usr/sfw/bin + export PATH +fi + +unset CDPATH + +MS_Printf() +{ + \$print_cmd \$print_cmd_arg "\$1" +} + +MS_PrintLicense() +{ + PAGER=\${PAGER:=more} + if test x"\$licensetxt" != x; then + PAGER_PATH=\`exec <&- 2>&-; which \$PAGER || command -v \$PAGER || type \$PAGER\` + if test -x "\$PAGER_PATH"; then + echo "\$licensetxt" | \$PAGER + else + echo "\$licensetxt" + fi + if test x"\$accept" != xy; then + while true + do + MS_Printf "Please type y to accept, n otherwise: " + read yn + if test x"\$yn" = xn; then + keep=n + eval \$finish; exit 1 + break; + elif test x"\$yn" = xy; then + break; + fi + done + fi + fi +} + +MS_diskspace() +{ + ( + df -kP "\$1" | tail -1 | awk '{ if (\$4 ~ /%/) {print \$3} else {print \$4} }' + ) +} + +MS_dd() +{ + blocks=\`expr \$3 / 1024\` + bytes=\`expr \$3 % 1024\` + # Test for ibs, obs and conv feature + if dd if=/dev/zero of=/dev/null count=1 ibs=512 obs=512 conv=sync 2> /dev/null; then + dd if="\$1" ibs=\$2 skip=1 obs=1024 conv=sync 2> /dev/null | \\ + { test \$blocks -gt 0 && dd ibs=1024 obs=1024 count=\$blocks ; \\ + test \$bytes -gt 0 && dd ibs=1 obs=1024 count=\$bytes ; } 2> /dev/null + else + dd if="\$1" bs=\$2 skip=1 2> /dev/null + fi +} + +MS_dd_Progress() +{ + if test x"\$noprogress" = xy; then + MS_dd "\$@" + return \$? + fi + file="\$1" + offset=\$2 + length=\$3 + pos=0 + bsize=4194304 + while test \$bsize -gt \$length; do + bsize=\`expr \$bsize / 4\` + done + blocks=\`expr \$length / \$bsize\` + bytes=\`expr \$length % \$bsize\` + ( + dd ibs=\$offset skip=1 2>/dev/null + pos=\`expr \$pos \+ \$bsize\` + MS_Printf " 0%% " 1>&2 + if test \$blocks -gt 0; then + while test \$pos -le \$length; do + dd bs=\$bsize count=1 2>/dev/null + pcent=\`expr \$length / 100\` + pcent=\`expr \$pos / \$pcent\` + if test \$pcent -lt 100; then + MS_Printf "\b\b\b\b\b\b\b" 1>&2 + if test \$pcent -lt 10; then + MS_Printf " \$pcent%% " 1>&2 + else + MS_Printf " \$pcent%% " 1>&2 + fi + fi + pos=\`expr \$pos \+ \$bsize\` + done + fi + if test \$bytes -gt 0; then + dd bs=\$bytes count=1 2>/dev/null + fi + MS_Printf "\b\b\b\b\b\b\b" 1>&2 + MS_Printf " 100%% " 1>&2 + ) < "\$file" +} + +MS_Help() +{ + cat << EOH >&2 +Usage: \$0 [options] +Options: + --help | -h Print this message + --info Print embedded info : title, default target directory, embedded script ... + --list Print the list of files in the archive + --check Checks integrity and version dependency of the archive + --quiet Quiet install mode, skip human-computer interactions + --nox11 Do not spawn an xterm + --noexec Do not run embedded script + --extract= Extract directly to a target directory (absolute or relative) + Usually used with --noexec to just extract files without running + --tar arg1 [arg2 ...] Access the contents of the archive through the tar command +\${helpheader} +EOH +} + +MS_Verify_Sig() +{ + GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` + MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` + test -x "\$GPG_PATH" || GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` + test -x "\$MKTEMP_PATH" || MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` + offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` + temp_sig=\`mktemp -t XXXXX\` + echo \$SIGNATURE | base64 --decode > "\$temp_sig" + gpg_output=\`MS_dd "\$1" \$offset \$totalsize | LC_ALL=C "\$GPG_PATH" --verify "\$temp_sig" - 2>&1\` + gpg_res=\$? + rm -f "\$temp_sig" + if test \$gpg_res -eq 0 && test \`echo \$gpg_output | grep -c Good\` -eq 1; then + if test \`echo \$gpg_output | grep -c \$sig_key\` -eq 1; then + test x"\$quiet" = xn && echo "GPG signature is good" >&2 + else + echo "GPG Signature key does not match" >&2 + exit 2 + fi + else + test x"\$quiet" = xn && echo "GPG signature failed to verify" >&2 + exit 2 + fi +} + +MS_Check() +{ + OLD_PATH="\$PATH" + PATH=\${GUESS_MD5_PATH:-"\$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} + MD5_ARG="" + MD5_PATH=\`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum\` + test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which md5 || command -v md5 || type md5\` + test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which digest || command -v digest || type digest\` + PATH="\$OLD_PATH" + + SHA_PATH=\`exec <&- 2>&-; which shasum || command -v shasum || type shasum\` + test -x "\$SHA_PATH" || SHA_PATH=\`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum\` + + if test x"\$quiet" = xn; then + MS_Printf "Verifying archive integrity..." + fi + offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` + fsize=\`cat "\$1" | wc -c | tr -d " "\` + if test \$totalsize -ne \`expr \$fsize - \$offset\`; then + echo " Unexpected archive size." >&2 + exit 2 + fi + verb=\$2 + i=1 + for s in \$filesizes + do + crc=\`echo \$CRCsum | cut -d" " -f\$i\` + if test -x "\$SHA_PATH"; then + if test x"\`basename \$SHA_PATH\`" = xshasum; then + SHA_ARG="-a 256" + fi + sha=\`echo \$SHA | cut -d" " -f\$i\` + if test x"\$sha" = x0000000000000000000000000000000000000000000000000000000000000000; then + test x"\$verb" = xy && echo " \$1 does not contain an embedded SHA256 checksum." >&2 + else + shasum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$SHA_PATH \$SHA_ARG" | cut -b-64\`; + if test x"\$shasum" != x"\$sha"; then + echo "Error in SHA256 checksums: \$shasum is different from \$sha" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " SHA256 checksums are OK." >&2 + fi + crc="0000000000"; + fi + fi + if test -x "\$MD5_PATH"; then + if test x"\`basename \$MD5_PATH\`" = xdigest; then + MD5_ARG="-a md5" + fi + md5=\`echo \$MD5 | cut -d" " -f\$i\` + if test x"\$md5" = x00000000000000000000000000000000; then + test x"\$verb" = xy && echo " \$1 does not contain an embedded MD5 checksum." >&2 + else + md5sum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$MD5_PATH \$MD5_ARG" | cut -b-32\`; + if test x"\$md5sum" != x"\$md5"; then + echo "Error in MD5 checksums: \$md5sum is different from \$md5" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " MD5 checksums are OK." >&2 + fi + crc="0000000000"; verb=n + fi + fi + if test x"\$crc" = x0000000000; then + test x"\$verb" = xy && echo " \$1 does not contain a CRC checksum." >&2 + else + sum1=\`MS_dd_Progress "\$1" \$offset \$s | CMD_ENV=xpg4 cksum | awk '{print \$1}'\` + if test x"\$sum1" != x"\$crc"; then + echo "Error in checksums: \$sum1 is different from \$crc" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " CRC checksums are OK." >&2 + fi + fi + i=\`expr \$i + 1\` + offset=\`expr \$offset + \$s\` + done + if test x"\$quiet" = xn; then + echo " All good." + fi +} + +MS_Decompress() +{ + if test x"\$decrypt_cmd" != x""; then + { eval "\$decrypt_cmd" || echo " ... Decryption failed." >&2; } | eval "$GUNZIP_CMD" + else + eval "$GUNZIP_CMD" + fi + + if test \$? -ne 0; then + echo " ... Decompression failed." >&2 + fi +} + +UnTAR() +{ + if test x"\$quiet" = xn; then + tar \$1vf - $UNTAR_EXTRA 2>&1 || { echo " ... Extraction failed." >&2; kill -15 \$$; } + else + tar \$1f - $UNTAR_EXTRA 2>&1 || { echo Extraction failed. >&2; kill -15 \$$; } + fi +} + +MS_exec_cleanup() { + if test x"\$cleanup" = xy && test x"\$cleanup_script" != x""; then + cleanup=n + cd "\$tmpdir" + eval "\"\$cleanup_script\" \$scriptargs \$cleanupargs" + fi +} + +MS_cleanup() +{ + echo 'Signal caught, cleaning up' >&2 + MS_exec_cleanup + cd "\$TMPROOT" + rm -rf "\$tmpdir" + eval \$finish; exit 15 +} + +Script_Args_Check() +{ + script_supported_args=\$(echo \${helpheader} | grep -o -E "\-\-[^ ]+" | awk -F"=" {'print \$1'}) + arg_to_test=\$(echo \$1|awk -F"=" {'print \$1'}) + + for arg in \${script_supported_args}; + do + if test x"\$arg_to_test" = x"\$arg" ;then + return + fi + done + + MS_Help + exit 1 +} + +finish=true +xterm_loop= +noprogress=$NOPROGRESS +nox11=$NOX11 +copy=$COPY +ownership=$OWNERSHIP +verbose=n +cleanup=y +cleanupargs= +sig_key= + +initargs="\$@" + +while [ -n "\$*" ] +do + case "\$1" in + -h | --help) + MS_Help + exit 0 + ;; + -q | --quiet) + quiet=y + noprogress=y + shift + ;; + --info) + echo Identification: "\$label" + echo Target directory: "\$targetdir" + echo Uncompressed size: $USIZE KB + echo Compression: $COMPRESS + if test x"$ENCRYPT" != x""; then + echo Encryption: $ENCRYPT + fi + echo Date of packaging: $DATE + echo Built with Makeself version $MS_VERSION + echo Build command was: "$MS_COMMAND" + if test x"\$script" != x; then + echo Script run after extraction: + echo " " \$script \$scriptargs + fi + if test x"$copy" = xcopy; then + echo "Archive will copy itself to a temporary location" + fi + if test x"$NEED_ROOT" = xy; then + echo "Root permissions required for extraction" + fi + if test x"$KEEP" = xy; then + echo "directory \$targetdir is permanent" + else + echo "\$targetdir will be removed after extraction" + fi + exit 0 + ;; + --list) + echo Target directory: \$targetdir + offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + for s in \$filesizes + do + MS_dd "\$0" \$offset \$s | MS_Decompress | UnTAR t + offset=\`expr \$offset + \$s\` + done + exit 0 + ;; + --tar) + offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + arg1="\$2" + shift 2 || { MS_Help; exit 1; } + for s in \$filesizes + do + MS_dd "\$0" \$offset \$s | MS_Decompress | tar "\$arg1" - "\$@" + offset=\`expr \$offset + \$s\` + done + exit 0 + ;; + --check) + MS_Check "\$0" y + scriptargs="\$scriptargs \$1" + shift + ;; + --noexec) + script="" + cleanup_script="" + shift + ;; + --extract=*) + keep=y + targetdir=\`echo \$1 | cut -d"=" -f2 \` + if ! shift; then MS_Help; exit 1; fi + ;; + --nox11) + nox11=y + shift + ;; + --xwin) + if test "$NOWAIT" = n; then + finish="echo Press Return to close this window...; read junk" + fi + xterm_loop=1 + shift + ;; + --phase2) + copy=phase2 + shift + ;; + --repack | --repack-path=*) + Script_Args_Check \$1 + scriptargs="\$scriptargs '\$1'" + shift + if [[ ! "\$1" =~ ^-.* ]]; then + scriptargs="\$scriptargs '\$1'" + shift + fi + ;; + *) + Script_Args_Check \$1 + scriptargs="\$scriptargs '\$1'" + shift + ;; + esac +done + +quiet_para="" +if test x"\$quiet" = xy; then + quiet_para="--quiet " +fi +scriptargs="--\$name_of_file""--\"\$pwd_of_file\""" \$quiet_para""\$scriptargs" + +if test x"\$quiet" = xy -a x"\$verbose" = xy; then + echo Cannot be verbose and quiet at the same time. >&2 + exit 1 +fi + +if test x"$NEED_ROOT" = xy -a \`id -u\` -ne 0; then + echo "Administrative privileges required for this archive (use su or sudo)" >&2 + exit 1 +fi + +if test x"\$copy" \!= xphase2; then + MS_PrintLicense +fi + +case "\$copy" in +copy) + tmpdir="\$TMPROOT"/makeself.\$RANDOM.\`date +"%y%m%d%H%M%S"\`.\$\$ + mkdir "\$tmpdir" || { + echo "Could not create temporary directory \$tmpdir" >&2 + exit 1 + } + SCRIPT_COPY="\$tmpdir/makeself" + echo "Copying to a temporary location..." >&2 + cp "\$0" "\$SCRIPT_COPY" + chmod +x "\$SCRIPT_COPY" + cd "\$TMPROOT" + exec "\$SCRIPT_COPY" --phase2 -- \$initargs + ;; +phase2) + finish="\$finish ; rm -rf \`dirname \$0\`" + ;; +esac + +if test x"\$nox11" = xn; then + if tty -s; then # Do we have a terminal? + : + else + if test x"\$DISPLAY" != x -a x"\$xterm_loop" = x; then # No, but do we have X? + if xset q > /dev/null 2>&1; then # Check for valid DISPLAY variable + GUESS_XTERMS="xterm gnome-terminal rxvt dtterm eterm Eterm xfce4-terminal lxterminal kvt konsole aterm terminology" + for a in \$GUESS_XTERMS; do + if type \$a >/dev/null 2>&1; then + XTERM=\$a + break + fi + done + chmod a+x \$0 || echo Please add execution rights on \$0 + if test \`echo "\$0" | cut -c1\` = "/"; then # Spawn a terminal! + exec \$XTERM -e "\$0 --xwin \$initargs" + else + exec \$XTERM -e "./\$0 --xwin \$initargs" + fi + fi + fi + fi +fi + +if test x"\$targetdir" = x.; then + tmpdir="." +else + if test x"\$keep" = xy; then + if test x"\$nooverwrite" = xy && test -d "\$targetdir"; then + echo "Target directory \$targetdir already exists, aborting." >&2 + exit 1 + fi + if test x"\$quiet" = xn; then + echo "Creating directory \$targetdir" >&2 + fi + tmpdir="\$targetdir" + dashp="-p" + else + tmpdir="\$TMPROOT/selfgz\$\$\$RANDOM" + dashp="" + fi + mkdir \$dashp "\$tmpdir" || { + echo 'Cannot create target directory' \$tmpdir >&2 + echo 'You should try option --extract=' >&2 + eval \$finish + exit 1 + } +fi + +location="\`pwd\`" +if test x"\$SETUP_NOCHECK" != x1; then + MS_Check "\$0" +fi +offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + +if test x"\$verbose" = xy; then + MS_Printf "About to extract $USIZE KB in \$tmpdir ... Proceed ? [Y/n] " + read yn + if test x"\$yn" = xn; then + eval \$finish; exit 1 + fi +fi + +if test x"\$quiet" = xn; then + # Decrypting with openssl will ask for password, + # the prompt needs to start on new line + if test x"$ENCRYPT" = x"openssl"; then + echo "Decrypting and uncompressing \$label..." + else + MS_Printf "Uncompressing \$label" + fi +fi +res=3 +if test x"\$keep" = xn; then + trap MS_cleanup 1 2 3 15 +fi + +if test x"\$nodiskspace" = xn; then + leftspace=\`MS_diskspace "\$tmpdir"\` + if test -n "\$leftspace"; then + if test "\$leftspace" -lt $USIZE; then + echo + echo "Not enough space left in "\`dirname \$tmpdir\`" (\$leftspace KB) to decompress \$0 ($USIZE KB)" >&2 + if test x"\$keep" = xn; then + echo "Consider setting TMPDIR to a directory with more free space." + fi + eval \$finish; exit 1 + fi + fi +fi + +for s in \$filesizes +do + if MS_dd_Progress "\$0" \$offset \$s | MS_Decompress | ( cd "\$tmpdir"; umask \$ORIG_UMASK ; UnTAR xp ) 1>/dev/null; then + if test x"\$ownership" = xy; then + (cd "\$tmpdir"; chown -R \`id -u\` .; chgrp -R \`id -g\` .) + fi + else + echo >&2 + echo "Unable to decompress \$0" >&2 + eval \$finish; exit 1 + fi + offset=\`expr \$offset + \$s\` +done +if test x"\$quiet" = xn; then + echo +fi + +cd "\$tmpdir" +res=0 +if test x"\$script" != x; then + if test x"\$export_conf" = x"y"; then + MS_BUNDLE="\$0" + MS_LABEL="\$label" + MS_SCRIPT="\$script" + MS_SCRIPTARGS="\$scriptargs" + MS_ARCHDIRNAME="\$archdirname" + MS_KEEP="\$KEEP" + MS_NOOVERWRITE="\$NOOVERWRITE" + MS_COMPRESS="\$COMPRESS" + MS_CLEANUP="\$cleanup" + export MS_BUNDLE MS_LABEL MS_SCRIPT MS_SCRIPTARGS + export MS_ARCHDIRNAME MS_KEEP MS_NOOVERWRITE MS_COMPRESS + fi + + if test x"\$verbose" = x"y"; then + yn="x" + while test x"\$yn" != x -a x"\$yn" != xy -a x"\$yn" != xY -a x"\$yn" != xn -a x"\$yn" != xN + do + MS_Printf "OK to execute: \$script \$scriptargs \$* ? [Y/n] " + read yn + if test x"\$yn" = x -o x"\$yn" = xy -o x"\$yn" = xY; then + eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$?; + elif test x"\$yn" = xn -o x"\$yn" = xN; then + echo "Unable to decompress \$script ,because of aborting! ";res=\$? + else + echo "Input value is unacceptable,please try again." + fi + done + else + eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$? + fi + if test "\$res" -ne 0; then + test x"\$verbose" = xy && echo "The program '\$script' returned an error code (\$res)" >&2 + fi +fi + +MS_exec_cleanup + +if test x"\$keep" = xn; then + cd "\$TMPROOT" + rm -rf "\$tmpdir" +fi +eval \$finish; exit \$res +EOF diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.1 b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.1 new file mode 100644 index 000000000..81bf6e4ff --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.1 @@ -0,0 +1,110 @@ +.TH "MAKESELF" "1" "2.4.5" +.SH "NAME" +makeself \- An utility to generate self-extractable archives. +.SH "SYNTAX" +.B makeself [\fIoptions\fP] archive_dir file_name label +.B [\fIstartup_script\fP] [\fIargs\fP] +.SH "DESCRIPTION" +This program is a free (GPL) utility designed to create self-extractable +archives from a directory. +.SH "OPTIONS" +The following options are supported. +.TP 15 +.B -v, --version +Prints out the makeself version number and exits. +.TP +.B -h, --help +Print out help information. +.TP +.B --tar-quietly +Suppress verbose output from the tar command +.TP +.B --quiet +Do not print any messages other than errors +.TP +.B --gzip +Compress using gzip (default if detected). +.TP +.B --bzip2 +Compress using bzip2. +.TP +.B --pbzip2 +Compress using pbzip2. +.TP +.B --xz +Compress using xz. +.TP +.B --lzo +Compress using lzop. +.TP +.B --lz4 +Compress using lz4. +.TP +.B --compress +Compress using the UNIX 'compress' command. +.TP +.B --nocomp +Do not compress the data. +.TP +.B --complevel lvl +Specify the compression level for gzip,bzip2,pbzui2,xz,lzo or lz4 +.TP +.B --notemp +The archive will create archive_dir in the current directory and +uncompress in ./archive_dir. +.TP +.B --copy +Upon extraction, the archive will first copy itself to a temporary directory. +.TP +.B --append +Append more files to an existing makeself archive. The label and startup scripts will then be ignored. +.TP +.B --current +Files will be extracted to the current directory. Both --current and --target dir imply --notemp. +.TP +.B --target dir +Extract directly to a target directory. Directory path can be either absolute or relative. +.TP +.B --header file +Specify location of the header script. +.TP +.B --cleanup file +Specify a cleanup script that executes on interrupt and when finished successfully. +.TP +.B --follow +Follow the symlinks in the archive. +.TP +.B --noprogress +Do not show the progress during the decompression. +.TP +.B --nox11 +Disable automatic spawn of an xterm if running in X11. +.TP +.B --nowait +Do not wait for user input after executing embedded program from an xterm. +.TP +.B --nomd5 +Do not create a MD5 checksum for the archive. +.TP +.B --nocrc +Do not create a CRC32 checksum for the archive. +.TP +.B --lsm file +LSM file describing the package. +.B --packaging-date date +Use provided string as the packaging date instead of the current date. +.SH "EXAMPLES" +Here is an example, assuming the user has a package image stored in a /home/joe/mysoft, +and he wants to generate a self-extracting package named mysoft.sh, which will launch +the "setup" script initially stored in /home/joe/mysoft: +.TP +makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup +.TP +Here is also how I created the makeself.run archive which contains the Makeself distribution: +.TP +makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" +.SH "AUTHORS" +Makeself has been written by Stéphane Peter . +.BR +This man page was originally written by Bartosz Fenski for the +Debian GNU/Linux distribution (but it may be used by others). diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.lsm b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.lsm new file mode 100644 index 000000000..3c4cea8c1 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.lsm @@ -0,0 +1,16 @@ +Begin3 +Title: makeself.sh +Version: 2.4.5 +Description: makeself.sh is a shell script that generates a self-extractable + tar.gz archive from a directory. The resulting file appears as a shell + script, and can be launched as is. The archive will then uncompress + itself to a temporary directory and an arbitrary command will be + executed (for example an installation script). This is pretty similar + to archives generated with WinZip Self-Extractor in the Windows world. +Keywords: Installation archive tar winzip +Author: Stephane Peter (megastep@megastep.org) +Maintained-by: Stephane Peter (megastep@megastep.org) +Original-site: https://makeself.io/ +Platform: Unix +Copying-policy: GPL +End diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.sh new file mode 100644 index 000000000..c8ea56597 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.sh @@ -0,0 +1,822 @@ +#!/bin/sh +# +# Makeself version 2.4.x +# by Stephane Peter +# +# Utility to create self-extracting tar.gz archives. +# The resulting archive is a file holding the tar.gz archive with +# a small Shell script stub that uncompresses the archive to a temporary +# directory and then executes a given script from withing that directory. +# +# Makeself home page: https://makeself.io/ +# +# Version 2.0 is a rewrite of version 1.0 to make the code easier to read and maintain. +# +# Version history : +# - 1.0 : Initial public release +# - 1.1 : The archive can be passed parameters that will be passed on to +# the embedded script, thanks to John C. Quillan +# - 1.2 : Package distribution, bzip2 compression, more command line options, +# support for non-temporary archives. Ideas thanks to Francois Petitjean +# - 1.3 : More patches from Bjarni R. Einarsson and Francois Petitjean: +# Support for no compression (--nocomp), script is no longer mandatory, +# automatic launch in an xterm, optional verbose output, and -target +# archive option to indicate where to extract the files. +# - 1.4 : Improved UNIX compatibility (Francois Petitjean) +# Automatic integrity checking, support of LSM files (Francois Petitjean) +# - 1.5 : Many bugfixes. Optionally disable xterm spawning. +# - 1.5.1 : More bugfixes, added archive options -list and -check. +# - 1.5.2 : Cosmetic changes to inform the user of what's going on with big +# archives (Quake III demo) +# - 1.5.3 : Check for validity of the DISPLAY variable before launching an xterm. +# More verbosity in xterms and check for embedded command's return value. +# Bugfix for Debian 2.0 systems that have a different "print" command. +# - 1.5.4 : Many bugfixes. Print out a message if the extraction failed. +# - 1.5.5 : More bugfixes. Added support for SETUP_NOCHECK environment variable to +# bypass checksum verification of archives. +# - 1.6.0 : Compute MD5 checksums with the md5sum command (patch from Ryan Gordon) +# - 2.0 : Brand new rewrite, cleaner architecture, separated header and UNIX ports. +# - 2.0.1 : Added --copy +# - 2.1.0 : Allow multiple tarballs to be stored in one archive, and incremental updates. +# Added --nochown for archives +# Stopped doing redundant checksums when not necesary +# - 2.1.1 : Work around insane behavior from certain Linux distros with no 'uncompress' command +# Cleaned up the code to handle error codes from compress. Simplified the extraction code. +# - 2.1.2 : Some bug fixes. Use head -n to avoid problems. +# - 2.1.3 : Bug fixes with command line when spawning terminals. +# Added --tar for archives, allowing to give arbitrary arguments to tar on the contents of the archive. +# Added --noexec to prevent execution of embedded scripts. +# Added --nomd5 and --nocrc to avoid creating checksums in archives. +# Added command used to create the archive in --info output. +# Run the embedded script through eval. +# - 2.1.4 : Fixed --info output. +# Generate random directory name when extracting files to . to avoid problems. (Jason Trent) +# Better handling of errors with wrong permissions for the directory containing the files. (Jason Trent) +# Avoid some race conditions (Ludwig Nussel) +# Unset the $CDPATH variable to avoid problems if it is set. (Debian) +# Better handling of dot files in the archive directory. +# - 2.1.5 : Made the md5sum detection consistent with the header code. +# Check for the presence of the archive directory +# Added --encrypt for symmetric encryption through gpg (Eric Windisch) +# Added support for the digest command on Solaris 10 for MD5 checksums +# Check for available disk space before extracting to the target directory (Andreas Schweitzer) +# Allow extraction to run asynchronously (patch by Peter Hatch) +# Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo) +# - 2.1.6 : Replaced one dot per file progress with a realtime progress percentage and a spining cursor (Guy Baconniere) +# Added --noprogress to prevent showing the progress during the decompression (Guy Baconniere) +# Added --target dir to allow extracting directly to a target directory (Guy Baconniere) +# - 2.2.0 : Many bugfixes, updates and contributions from users. Check out the project page on Github for the details. +# - 2.3.0 : Option to specify packaging date to enable byte-for-byte reproducibility. (Marc Pawlowsky) +# - 2.4.0 : Optional support for SHA256 checksums in archives. +# - 2.4.2 : Add support for threads for several compressors. (M. Limber) +# Added zstd support. +# - 2.4.3 : Make explicit POSIX tar archives for increased compatibility. +# - 2.4.5 : Added --tar-format to override ustar tar archive format +# +# (C) 1998-2021 by Stephane Peter +# +# This software is released under the terms of the GNU GPL version 2 and above +# Please read the license at http://www.gnu.org/copyleft/gpl.html +# Self-extracting archives created with this script are explictly NOT released under the term of the GPL +# + +MS_VERSION=2.4.5 +MS_COMMAND="$0" +unset CDPATH + +for f in ${1+"$@"}; do + MS_COMMAND="$MS_COMMAND \\\\ + \\\"$f\\\"" +done + +# For Solaris systems +if test -d /usr/xpg4/bin; then + PATH=/usr/xpg4/bin:$PATH + export PATH +fi + +# Procedures + +MS_Usage() +{ + echo "Usage: $0 [args] archive_dir file_name label startup_script [script_args]" + echo "args can be one or more of the following :" + echo " --version | -v : Print out Makeself version number and exit" + echo " --help | -h : Print out this help message" + echo " --tar-quietly : Suppress verbose output from the tar command" + echo " --quiet | -q : Do not print any messages other than errors." + echo " --gzip : Compress using gzip (default if detected)" + echo " --pigz : Compress with pigz" + echo " --zstd : Compress with zstd" + echo " --bzip2 : Compress using bzip2 instead of gzip" + echo " --pbzip2 : Compress using pbzip2 instead of gzip" + echo " --xz : Compress using xz instead of gzip" + echo " --lzo : Compress using lzop instead of gzip" + echo " --lz4 : Compress using lz4 instead of gzip" + echo " --compress : Compress using the UNIX 'compress' command" + echo " --complevel lvl : Compression level for gzip pigz zstd xz lzo lz4 bzip2 and pbzip2 (default 9)" + echo " --threads thds : Number of threads to be used by compressors that support parallelization." + echo " Omit to use compressor's default. Most useful (and required) for opting" + echo " into xz's threading, usually with '--threads=0' for all available cores." + echo " pbzip2 and pigz are parallel by default, and setting this value allows" + echo " limiting the number of threads they use." + echo " --base64 : Instead of compressing, encode the data using base64" + echo " --gpg-encrypt : Instead of compressing, encrypt the data using GPG" + echo " --gpg-asymmetric-encrypt-sign" + echo " : Instead of compressing, asymmetrically encrypt and sign the data using GPG" + echo " --gpg-extra opt : Append more options to the gpg command line" + echo " --ssl-encrypt : Instead of compressing, encrypt the data using OpenSSL" + echo " --ssl-passwd pass : Use the given password to encrypt the data using OpenSSL" + echo " --ssl-pass-src src : Use the given src as the source of password to encrypt the data" + echo " using OpenSSL. See \"PASS PHRASE ARGUMENTS\" in man openssl." + echo " If this option is not supplied, the user will be asked to enter" + echo " encryption password on the current terminal." + echo " --ssl-no-md : Do not use \"-md\" option not supported by older OpenSSL." + echo " --nochown : Do not give the target folder to the current user (default)" + echo " --chown : Give the target folder to the current user recursively" + echo " --nocomp : Do not compress the data" + echo " --notemp : The archive will create archive_dir in the" + echo " current directory and uncompress in ./archive_dir" + echo " --needroot : Check that the root user is extracting the archive before proceeding" + echo " --copy : Upon extraction, the archive will first copy itself to" + echo " a temporary directory" + echo " --append : Append more files to an existing Makeself archive" + echo " The label and startup scripts will then be ignored" + echo " --target dir : Extract directly to a target directory" + echo " directory path can be either absolute or relative" + echo " --nooverwrite : Do not extract the archive if the specified target directory exists" + echo " --current : Files will be extracted to the current directory" + echo " Both --current and --target imply --notemp" + echo " --tar-format opt : Specify a tar archive format (default is ustar)" + echo " --tar-extra opt : Append more options to the tar command line" + echo " --untar-extra opt : Append more options to the during the extraction of the tar archive" + echo " --nomd5 : Don't calculate an MD5 for archive" + echo " --nocrc : Don't calculate a CRC for archive" + echo " --sha256 : Compute a SHA256 checksum for the archive" + echo " --header file : Specify location of the header script" + echo " --cleanup file : Specify a cleanup script that executes on interrupt and when finished successfully." + echo " --follow : Follow the symlinks in the archive" + echo " --noprogress : Do not show the progress during the decompression" + echo " --nox11 : Disable automatic spawn of a xterm" + echo " --nowait : Do not wait for user input after executing embedded" + echo " program from an xterm" + echo " --sign passphrase : Signature private key to sign the package with" + echo " --lsm file : LSM file describing the package" + echo " --license file : Append a license file" + echo " --help-header file : Add a header to the archive's --help output" + echo " --packaging-date date" + echo " : Use provided string as the packaging date" + echo " instead of the current date." + echo + echo " --keep-umask : Keep the umask set to shell default, rather than overriding when executing self-extracting archive." + echo " --export-conf : Export configuration variables to startup_script" + echo + echo "Do not forget to give a fully qualified startup script name" + echo "(i.e. with a ./ prefix if inside the archive)." + exit 1 +} + +# Default settings +if type gzip >/dev/null 2>&1; then + COMPRESS=gzip +elif type compress >/dev/null 2>&1; then + COMPRESS=compress +else + echo "ERROR: missing commands: gzip, compress" >&2 + MS_Usage +fi +ENCRYPT=n +PASSWD="" +PASSWD_SRC="" +OPENSSL_NO_MD=n +COMPRESS_LEVEL=9 +DEFAULT_THREADS=123456 # Sentinel value +THREADS=$DEFAULT_THREADS +KEEP=n +CURRENT=n +NOX11=n +NOWAIT=n +APPEND=n +TAR_QUIETLY=n +KEEP_UMASK=n +QUIET=n +NOPROGRESS=n +COPY=none +NEED_ROOT=n +TAR_ARGS=rvf +TAR_FORMAT=ustar +TAR_EXTRA="" +GPG_EXTRA="" +DU_ARGS=-ks +HEADER=`dirname "$0"`/makeself-header.sh +SIGNATURE="" +TARGETDIR="" +NOOVERWRITE=n +DATE=`LC_ALL=C date` +EXPORT_CONF=n +SHA256=n +OWNERSHIP=n +SIGN=n +GPG_PASSPHRASE="" + +# LSM file stuff +LSM_CMD="echo No LSM. >> \"\$archname\"" + +while true +do + case "$1" in + --version | -v) + echo Makeself version $MS_VERSION + exit 0 + ;; + --pbzip2) + COMPRESS=pbzip2 + shift + ;; + --bzip2) + COMPRESS=bzip2 + shift + ;; + --gzip) + COMPRESS=gzip + shift + ;; + --pigz) + COMPRESS=pigz + shift + ;; + --zstd) + COMPRESS=zstd + shift + ;; + --xz) + COMPRESS=xz + shift + ;; + --lzo) + COMPRESS=lzo + shift + ;; + --lz4) + COMPRESS=lz4 + shift + ;; + --compress) + COMPRESS=compress + shift + ;; + --base64) + COMPRESS=base64 + shift + ;; + --gpg-encrypt) + COMPRESS=gpg + shift + ;; + --gpg-asymmetric-encrypt-sign) + COMPRESS=gpg-asymmetric + shift + ;; + --gpg-extra) + GPG_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-encrypt) + ENCRYPT=openssl + shift + ;; + --ssl-passwd) + PASSWD=$2 + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-pass-src) + PASSWD_SRC=$2 + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-no-md) + OPENSSL_NO_MD=y + shift + ;; + --nocomp) + COMPRESS=none + shift + ;; + --complevel) + COMPRESS_LEVEL="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --threads) + THREADS="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --nochown) + OWNERSHIP=n + shift + ;; + --chown) + OWNERSHIP=y + shift + ;; + --notemp) + KEEP=y + shift + ;; + --copy) + COPY=copy + shift + ;; + --current) + CURRENT=y + KEEP=y + shift + ;; + --tar-format) + TAR_FORMAT="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --tar-extra) + TAR_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --untar-extra) + UNTAR_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --target) + TARGETDIR="$2" + KEEP=y + shift 2 || { MS_Usage; exit 1; } + ;; + --sign) + SIGN=y + GPG_PASSPHRASE="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --nooverwrite) + NOOVERWRITE=y + shift + ;; + --needroot) + NEED_ROOT=y + shift + ;; + --header) + HEADER="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --cleanup) + CLEANUP_SCRIPT="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --license) + # We need to escape all characters having a special meaning in double quotes + LICENSE=$(sed 's/\\/\\\\/g; s/"/\\\"/g; s/`/\\\`/g; s/\$/\\\$/g' "$2") + shift 2 || { MS_Usage; exit 1; } + ;; + --follow) + TAR_ARGS=rvhf + DU_ARGS=-ksL + shift + ;; + --noprogress) + NOPROGRESS=y + shift + ;; + --nox11) + NOX11=y + shift + ;; + --nowait) + NOWAIT=y + shift + ;; + --nomd5) + NOMD5=y + shift + ;; + --sha256) + SHA256=y + shift + ;; + --nocrc) + NOCRC=y + shift + ;; + --append) + APPEND=y + shift + ;; + --lsm) + LSM_CMD="cat \"$2\" >> \"\$archname\"" + shift 2 || { MS_Usage; exit 1; } + ;; + --packaging-date) + DATE="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --help-header) + HELPHEADER=`sed -e "s/'/'\\\\\''/g" $2` + shift 2 || { MS_Usage; exit 1; } + [ -n "$HELPHEADER" ] && HELPHEADER="$HELPHEADER +" + ;; + --tar-quietly) + TAR_QUIETLY=y + shift + ;; + --keep-umask) + KEEP_UMASK=y + shift + ;; + --export-conf) + EXPORT_CONF=y + shift + ;; + -q | --quiet) + QUIET=y + shift + ;; + -h | --help) + MS_Usage + ;; + -*) + echo Unrecognized flag : "$1" + MS_Usage + ;; + *) + break + ;; + esac +done + +if test $# -lt 1; then + MS_Usage +else + if test -d "$1"; then + archdir="$1" + else + echo "Directory $1 does not exist." >&2 + exit 1 + fi +fi +archname="$2" + +if test "$QUIET" = "y" || test "$TAR_QUIETLY" = "y"; then + if test "$TAR_ARGS" = "rvf"; then + TAR_ARGS="rf" + elif test "$TAR_ARGS" = "rvhf"; then + TAR_ARGS="rhf" + fi +fi + +if test "$APPEND" = y; then + if test $# -lt 2; then + MS_Usage + fi + + # Gather the info from the original archive + OLDENV=`sh "$archname" --dumpconf` + if test $? -ne 0; then + echo "Unable to update archive: $archname" >&2 + exit 1 + else + eval "$OLDENV" + OLDSKIP=`expr $SKIP + 1` + fi +else + if test "$KEEP" = n -a $# = 3; then + echo "ERROR: Making a temporary archive with no embedded command does not make sense!" >&2 + echo >&2 + MS_Usage + fi + # We don't want to create an absolute directory unless a target directory is defined + if test "$CURRENT" = y; then + archdirname="." + elif test x"$TARGETDIR" != x; then + archdirname="$TARGETDIR" + else + archdirname=`basename "$1"` + fi + + if test $# -lt 3; then + MS_Usage + fi + + LABEL="$3" + SCRIPT="$4" + test "x$SCRIPT" = x || shift 1 + shift 3 + SCRIPTARGS="$*" +fi + +if test "$KEEP" = n -a "$CURRENT" = y; then + echo "ERROR: It is A VERY DANGEROUS IDEA to try to combine --notemp and --current." >&2 + exit 1 +fi + +case $COMPRESS in +gzip) + GZIP_CMD="gzip -c$COMPRESS_LEVEL" + GUNZIP_CMD="gzip -cd" + ;; +pigz) + GZIP_CMD="pigz -$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD --processes $THREADS" + fi + GUNZIP_CMD="gzip -cd" + ;; +zstd) + GZIP_CMD="zstd -$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD --threads=$THREADS" + fi + GUNZIP_CMD="zstd -cd" + ;; +pbzip2) + GZIP_CMD="pbzip2 -c$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD -p$THREADS" + fi + GUNZIP_CMD="bzip2 -d" + ;; +bzip2) + GZIP_CMD="bzip2 -$COMPRESS_LEVEL" + GUNZIP_CMD="bzip2 -d" + ;; +xz) + GZIP_CMD="xz -c$COMPRESS_LEVEL" + # Must opt-in by specifying a value since not all versions of xz support threads + if test $THREADS -ne $DEFAULT_THREADS; then + GZIP_CMD="$GZIP_CMD --threads=$THREADS" + fi + GUNZIP_CMD="xz -d" + ;; +lzo) + GZIP_CMD="lzop -c$COMPRESS_LEVEL" + GUNZIP_CMD="lzop -d" + ;; +lz4) + GZIP_CMD="lz4 -c$COMPRESS_LEVEL" + GUNZIP_CMD="lz4 -d" + ;; +base64) + GZIP_CMD="base64" + GUNZIP_CMD="base64 --decode -i -" + ;; +gpg) + GZIP_CMD="gpg $GPG_EXTRA -ac -z$COMPRESS_LEVEL" + GUNZIP_CMD="gpg -d" + ENCRYPT="gpg" + ;; +gpg-asymmetric) + GZIP_CMD="gpg $GPG_EXTRA -z$COMPRESS_LEVEL -es" + GUNZIP_CMD="gpg --yes -d" + ENCRYPT="gpg" + ;; +compress) + GZIP_CMD="compress -fc" + GUNZIP_CMD="(type compress >/dev/null 2>&1 && compress -fcd || gzip -cd)" + ;; +none) + GZIP_CMD="cat" + GUNZIP_CMD="cat" + ;; +esac + +if test x"$ENCRYPT" = x"openssl"; then + if test x"$APPEND" = x"y"; then + echo "Appending to existing archive is not compatible with OpenSSL encryption." >&2 + fi + + ENCRYPT_CMD="openssl enc -aes-256-cbc -salt" + DECRYPT_CMD="openssl enc -aes-256-cbc -d" + + if test x"$OPENSSL_NO_MD" != x"y"; then + ENCRYPT_CMD="$ENCRYPT_CMD -md sha256" + DECRYPT_CMD="$DECRYPT_CMD -md sha256" + fi + + if test -n "$PASSWD_SRC"; then + ENCRYPT_CMD="$ENCRYPT_CMD -pass $PASSWD_SRC" + elif test -n "$PASSWD"; then + ENCRYPT_CMD="$ENCRYPT_CMD -pass pass:$PASSWD" + fi +fi + +tmpfile="${TMPDIR:-/tmp}/mkself$$" + +if test -f "$HEADER"; then + oldarchname="$archname" + archname="$tmpfile" + # Generate a fake header to count its lines + SKIP=0 + . "$HEADER" + SKIP=`cat "$tmpfile" |wc -l` + # Get rid of any spaces + SKIP=`expr $SKIP` + rm -f "$tmpfile" + if test "$QUIET" = "n"; then + echo "Header is $SKIP lines long" >&2 + fi + archname="$oldarchname" +else + echo "Unable to open header file: $HEADER" >&2 + exit 1 +fi + +if test "$QUIET" = "n"; then + echo +fi + +if test "$APPEND" = n; then + if test -f "$archname"; then + echo "WARNING: Overwriting existing file: $archname" >&2 + fi +fi + +USIZE=`du $DU_ARGS "$archdir" | awk '{print $1}'` + +if test "." = "$archdirname"; then + if test "$KEEP" = n; then + archdirname="makeself-$$-`date +%Y%m%d%H%M%S`" + fi +fi + +test -d "$archdir" || { echo "Error: $archdir does not exist."; rm -f "$tmpfile"; exit 1; } +if test "$QUIET" = "n"; then + echo "About to compress $USIZE KB of data..." + echo "Adding files to archive named \"$archname\"..." +fi + +# See if we have GNU tar +TAR=`exec <&- 2>&-; which gtar || command -v gtar || type gtar` +test -x "$TAR" || TAR=tar + +tmparch="${TMPDIR:-/tmp}/mkself$$.tar" +( + if test "$APPEND" = "y"; then + tail -n "+$OLDSKIP" "$archname" | eval "$GUNZIP_CMD" > "$tmparch" + fi + cd "$archdir" + # "Determining if a directory is empty" + # https://www.etalabs.net/sh_tricks.html + find . \ + \( \ + ! -type d \ + -o \ + \( -links 2 -exec sh -c ' + is_empty () ( + cd "$1" + set -- .[!.]* ; test -f "$1" && return 1 + set -- ..?* ; test -f "$1" && return 1 + set -- * ; test -f "$1" && return 1 + return 0 + ) + is_empty "$0"' {} \; \ + \) \ + \) -print \ + | LC_ALL=C sort \ + | sed 's/./\\&/g' \ + | xargs $TAR $TAR_EXTRA --format $TAR_FORMAT -$TAR_ARGS "$tmparch" +) || { + echo "ERROR: failed to create temporary archive: $tmparch" + rm -f "$tmparch" "$tmpfile" + exit 1 +} + +USIZE=`du $DU_ARGS "$tmparch" | awk '{print $1}'` + +eval "$GZIP_CMD" <"$tmparch" >"$tmpfile" || { + echo "ERROR: failed to create temporary file: $tmpfile" + rm -f "$tmparch" "$tmpfile" + exit 1 +} +rm -f "$tmparch" + +if test x"$ENCRYPT" = x"openssl"; then + echo "About to encrypt archive \"$archname\"..." + { eval "$ENCRYPT_CMD -in $tmpfile -out ${tmpfile}.enc" && mv -f ${tmpfile}.enc $tmpfile; } || \ + { echo Aborting: could not encrypt temporary file: "$tmpfile".; rm -f "$tmpfile"; exit 1; } +fi + +fsize=`cat "$tmpfile" | wc -c | tr -d " "` + +# Compute the checksums + +shasum=0000000000000000000000000000000000000000000000000000000000000000 +md5sum=00000000000000000000000000000000 +crcsum=0000000000 + +if test "$NOCRC" = y; then + if test "$QUIET" = "n"; then + echo "skipping crc at user request" + fi +else + crcsum=`CMD_ENV=xpg4 cksum < "$tmpfile" | sed -e 's/ /Z/' -e 's/ /Z/' | cut -dZ -f1` + if test "$QUIET" = "n"; then + echo "CRC: $crcsum" + fi +fi + +if test "$SHA256" = y; then + SHA_PATH=`exec <&- 2>&-; which shasum || command -v shasum || type shasum` + if test -x "$SHA_PATH"; then + shasum=`eval "$SHA_PATH -a 256" < "$tmpfile" | cut -b-64` + else + SHA_PATH=`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum` + shasum=`eval "$SHA_PATH" < "$tmpfile" | cut -b-64` + fi + if test "$QUIET" = "n"; then + if test -x "$SHA_PATH"; then + echo "SHA256: $shasum" + else + echo "SHA256: none, SHA command not found" + fi + fi +fi +if test "$NOMD5" = y; then + if test "$QUIET" = "n"; then + echo "Skipping md5sum at user request" + fi +else + # Try to locate a MD5 binary + OLD_PATH=$PATH + PATH=${GUESS_MD5_PATH:-"$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} + MD5_ARG="" + MD5_PATH=`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum` + test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which md5 || command -v md5 || type md5` + test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which digest || command -v digest || type digest` + PATH=$OLD_PATH + if test -x "$MD5_PATH"; then + if test `basename ${MD5_PATH}`x = digestx; then + MD5_ARG="-a md5" + fi + md5sum=`eval "$MD5_PATH $MD5_ARG" < "$tmpfile" | cut -b-32` + if test "$QUIET" = "n"; then + echo "MD5: $md5sum" + fi + else + if test "$QUIET" = "n"; then + echo "MD5: none, MD5 command not found" + fi + fi +fi +if test "$SIGN" = y; then + GPG_PATH=`exec <&- 2>&-; which gpg || command -v gpg || type gpg` + if test -x "$GPG_PATH"; then + SIGNATURE=`$GPG_PATH --pinentry-mode=loopback --batch --yes --passphrase "$GPG_PASSPHRASE" --output - --detach-sig $tmpfile | base64 | tr -d \\\\n` + if test "$QUIET" = "n"; then + echo "Signature: $SIGNATURE" + fi + else + echo "Missing gpg command" >&2 + fi +fi + +totalsize=0 +for size in $fsize; +do + totalsize=`expr $totalsize + $size` +done + +if test "$APPEND" = y; then + mv "$archname" "$archname".bak || exit + + # Prepare entry for new archive + filesizes="$fsize" + CRCsum="$crcsum" + MD5sum="$md5sum" + SHAsum="$shasum" + Signature="$SIGNATURE" + # Generate the header + . "$HEADER" + # Append the new data + cat "$tmpfile" >> "$archname" + + chmod +x "$archname" + rm -f "$archname".bak + if test "$QUIET" = "n"; then + echo "Self-extractable archive \"$archname\" successfully updated." + fi +else + filesizes="$fsize" + CRCsum="$crcsum" + MD5sum="$md5sum" + SHAsum="$shasum" + Signature="$SIGNATURE" + + # Generate the header + . "$HEADER" + + # Append the compressed tar data after the stub + if test "$QUIET" = "n"; then + echo + fi + cat "$tmpfile" >> "$archname" + chmod +x "$archname" + if test "$QUIET" = "n"; then + echo Self-extractable archive \"$archname\" successfully created. + fi +fi +rm -f "$tmpfile" diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/run-tests.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/run-tests.sh new file mode 100644 index 000000000..31ee16511 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/run-tests.sh @@ -0,0 +1,8 @@ +#!/bin/sh +# Run every available test - Bash needed +cd test +for test in *test; +do + echo "Running test $test ..." + bash $test || { echo "*** ERROR: Test '$test' failed!"; exit 1; } +done diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/merge_aicpu_info_json.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/merge_aicpu_info_json.sh new file mode 100644 index 000000000..a977bd51d --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/merge_aicpu_info_json.sh @@ -0,0 +1,31 @@ +#!/bin/bash + +project_path=$1 +build_path=$2 +vendor_name=customize +echo $@ +if [[ ! -d "$project_path" ]]; then + echo "[ERROR] No projcet path is provided" + exit 1 +fi + +if [[ ! -d "$build_path" ]]; then + echo "[ERROR] No build path is provided" + exit 1 +fi + +if [[ ! -d "$ASCEND_OPP_PATH" ]]; then + echo "[ERROR] No opp install path is provided" + exit 1 +fi +custom_exist_info_json=$ASCEND_OPP_PATH/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json +custom_new_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json +temp_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/temp_cust_aicpu_kernel.json + +if [[ -f "$custom_exist_info_json" ]] && [[ -f "$custom_new_info_json" ]]; then + cp -f $custom_exist_info_json $temp_info_json + chmod +w $temp_info_json + python3 ${project_path}/cmake/util/insert_op_info.py ${custom_new_info_json} ${temp_info_json} + cp -f $temp_info_json $custom_new_info_json + rm -f $temp_info_json +fi diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/opdesc_parser.py b/atb_operator/AtbAddSample/AddOp/cmake/util/opdesc_parser.py new file mode 100644 index 000000000..c38e79edc --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/opdesc_parser.py @@ -0,0 +1,260 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os + + +OP_ALL = '__ALLOP__' +SOC_ALL = '__ALLSOC__' +SOC_TO_SHORT_SOC_MAP = { + "ascend910a": "ascend910", + "ascend910proa": "ascend910", + "ascend910b": "ascend910", + "ascend910prob": "ascend910", + "ascend910premiuma": "ascend910", + "ascend910b1": "ascend910b", + "ascend910b2": "ascend910b", + "ascend910b3": "ascend910b", + "ascend910b4": "ascend910b", + "ascend310p1": "ascend310p", + "ascend310p3": "ascend310p", + "ascend310p3vir01": "ascend310p", + "ascend310p3vir02": "ascend310p", + "ascend310p3vir04": "ascend310p", + "ascend310p3vir08": "ascend310p", + "ascend310b1": "ascend310b", + "bs9sx1aa": "bs9sx1a" +} + + +class OpDesc: + def __init__(self: any, op_type: str): + self.op_type = op_type + self.attr_list = [] + self.attr_val = {} + self.input_name = [] + self.input_type = [] + self.input_dtype = [] + self.input_fmt = [] + self.output_name = [] + self.output_type = [] + self.output_dtype = [] + self.output_fmt = [] + self.op_fmt_sel = False + self.op_chk_support = False + self.op_intf = '' + self.kern_name = '' + self.op_file = '' + self.op_replay_flag = False + self.op_replay_batch = False + self.input_idx = -1 + self.output_idx = -1 + self.max_block_dim = 32 + self.max_shape_size = 268435456 + self.dynamic_shape = False + self.op_range_limit = '' + self.custom_compile_options = {} + self.custom_all_compile_options = {} + + @staticmethod + def _parse_digit(conf: str) -> int: + return int(conf.split('=')[1]) + + @staticmethod + def _parse_flag(conf: str) -> bool: + if 'true' == conf.split('=')[1]: + return True + return False + + @staticmethod + def _parse_str(conf: str) -> str: + return conf.split('=')[1] + + @staticmethod + def _parse_list(conf: str) -> list: + return conf.split('=')[1].split(',') + + def parse_input(self: any, conf: str): + if conf.startswith('input{}.name'.format(int(self.input_idx) + 1)): + self.input_idx += 1 + self.input_name.append(self._parse_str(conf)) + elif conf.startswith('input{}.paramType'.format(int(self.input_idx))): + self.input_type.append(self._parse_str(conf)) + elif conf.startswith('input{}.dtype'.format(int(self.input_idx))): + self.input_dtype.append(self._parse_str(conf)) + elif conf.startswith('input{}.format'.format(int(self.input_idx))): + self.input_fmt.append(self._parse_str(conf)) + else: + return + + def parse_output(self: any, conf: str): + if conf.startswith('output{}.name'.format(int(self.output_idx) + 1)): + self.output_idx += 1 + self.output_name.append(self._parse_str(conf)) + elif conf.startswith('output{}.paramType'.format(int(self.output_idx))): + self.output_type.append(self._parse_str(conf)) + elif conf.startswith('output{}.dtype'.format(int(self.output_idx))): + self.output_dtype.append(self._parse_str(conf)) + elif conf.startswith('output{}.format'.format(int(self.output_idx))): + self.output_fmt.append(self._parse_str(conf)) + else: + return + + def parse_op_format(self: any, conf: str): + self.op_fmt_sel = self._parse_flag(conf) + + def parse_check_support(self: any, conf: str): + self.op_chk_support = self._parse_flag(conf) + + def parse_range_limit(self: any, conf: str): + self.op_range_limit = self._parse_str(conf) + + def parse_kern_name(self: any, conf: str): + self.kern_name = self._parse_str(conf) + + def parse_op_intf(self: any, conf: str): + self.op_intf = self._parse_str(conf) + + def parse_op_file(self: any, conf: str): + self.op_file = self._parse_str(conf) + + def parse_dynamic_shape(self: any, conf: str): + self.dynamic_shape = self._parse_flag(conf) + + def parse_attr_list(self: any, conf: str): + self.attr_list = self._parse_list(conf) + + def parse_attr_val(self: any, conf: str): + for attr in self.attr_list: + if self.attr_val.get(attr) is None: + self.attr_val[attr] = {} + if conf.startswith('attr_{}.type'.format(attr)): + self.attr_val.get(attr)['type'] = self._parse_str(conf) + elif conf.startswith('attr_{}.paramType'.format(attr)): + self.attr_val.get(attr)['paramType'] = self._parse_str(conf) + elif conf.startswith('attr_{}.defaultValue'.format(attr)): + self.attr_val.get(attr)['defaultValue'] = self._parse_str(conf) + + def parse_replay_val(self: any, batch_list: list, iterator_list: list): + if self.op_type in batch_list: + self.op_replay_flag = True + self.op_replay_batch = True + elif self.op_type in iterator_list: + self.op_replay_flag = True + self.op_replay_batch = False + + +def _is_op_type_in_opdesc(op_descs: list, op_type: str): + for op in op_descs: + if op_type == op.op_type: + return True + return False + + +def _set_all_options_to_opdescs(op_descs, soc_ver_compile_options): + for op in op_descs: + op.custom_all_compile_options = soc_ver_compile_options + + +def _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options): + for op in op_descs: + if op.op_type != op_type: + continue + op.custom_compile_options = soc_ver_compile_options + + +def _trans_soc_ver_to_short(soc_ver: str): + low_soc_ver = soc_ver.lower() + if low_soc_ver not in SOC_TO_SHORT_SOC_MAP: + print(f'WARNING: caution: {soc_ver} will trans into ascend910, if not your intention,' + f'use ascend910b1~4 instead') + return SOC_TO_SHORT_SOC_MAP[low_soc_ver] + + +def _get_op_custom_options(op_descs: list, auto_gen_dir: str): + if auto_gen_dir is None: + return {} + file = os.path.join(auto_gen_dir, "custom_compile_options.ini") + if not os.path.exists(file): + print(f'WARNING: cannot find {auto_gen_dir}/custom_compile_options.ini') + return {} + with open (file, 'r') as fd: + lines = fd.readlines() + for line in lines: + param_list = str.split(line.rstrip('\n'), ',') + if len(param_list) != 3: + raise Exception(f'ERROR: custom compile option {param_list} len is not 3') + op_type = param_list[0] + if op_type.upper() == 'ALL': + op_type = OP_ALL + if op_type != OP_ALL and _is_op_type_in_opdesc(op_descs, op_type) == False: + print(f'WARNING: op: {op_type} are not exists in this project') + continue + soc_ver_compile_options = {} + soc_ver = param_list[1] + options_str = param_list[2] + options = str.split(options_str, ';') + if soc_ver == '': + soc_ver_compile_options[SOC_ALL] = options + else: + soc_ver_list = str.split(soc_ver, ';') + for ver in soc_ver_list: + short_ver = _trans_soc_ver_to_short(ver) + soc_ver_compile_options[short_ver] = options + if op_type == OP_ALL: + _set_all_options_to_opdescs(op_descs, soc_ver_compile_options) + else: + _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options) + + +def get_op_desc(file: str, batch_list: list, iterator_list: list, builder: any, + op_type: list, auto_gen_dir: str = None) -> list: + op_descs = [] + op_match = False + with open (file, 'r') as fd: + lines = fd.readlines() + for line in lines: + line = line.strip() + if line.startswith('['): + name = line[1:-1] + if op_type is None or name in op_type: + op_match = True + op_desc = builder(name) + op_desc.parse_replay_val(batch_list, iterator_list) + op_descs.append(op_desc) + else: + op_match = False + if op_type is not None and len(op_descs) == len(op_type): + return op_descs + continue + if not op_match: + continue + if line.startswith('input'): + op_desc.parse_input(line) + elif line.startswith('output'): + op_desc.parse_output(line) + elif line.startswith('dynamicFormat.flag'): + op_desc.parse_op_format(line) + elif line.startswith('needCheckSupport.flag'): + op_desc.parse_check_support(line) + elif line.startswith('rangeLimit.value'): + op_desc.parse_range_limit(line) + elif line.startswith('opInterface.value'): + op_desc.parse_op_intf(line) + elif line.startswith('kernel.name'): + op_desc.parse_kern_name(line) + elif line.startswith('opFile.value'): + op_desc.parse_op_file(line) + elif line.startswith('dynamicShapeSupport.flag'): + op_desc.parse_dynamic_shape(line) + elif line.startswith('attr.list'): + op_desc.parse_attr_list(line) + elif line.startswith('attr_'): + op_desc.parse_attr_val(line) + _get_op_custom_options(op_descs, auto_gen_dir) + return op_descs diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/parse_ini_to_json.py b/atb_operator/AtbAddSample/AddOp/cmake/util/parse_ini_to_json.py new file mode 100644 index 000000000..721465fee --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/parse_ini_to_json.py @@ -0,0 +1,338 @@ +# Copyright 2020-2021 Huawei Technologies Co., Ltd +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +""" +parser ini to json +""" + +import json +import os +import stat +import sys + + +ATTR_TYPE_LIST = ["int", "float", "bool", "str", "listInt", "listFloat", "listBool", "listStr", "listListInt", + "type", "listType", "tensor", "listTensor"] +ATTR_PARAMTYPE_LIST = ["optional", "required"] +BOOL_FLAG_KEY = ["dynamicFormat", "dynamicShapeSupport", "dynamicRankSupport", "precision_reduce", "heavyOp", + "needCheckSupport"] +BOOL_LIST = ["true", "false"] +DTYPE_LIST = ["float16", "float", "float32", "int8", "int16", "int32", "uint8", "uint16", "uint32", "bool", + "int64", "uint64", "qint8", "qint16", "qint32", "quint8", "quint16", "double", "complex64", + "complex128", "string", "resource", "dual", "dual_sub_int8", "dual_sub_uint8", "string_ref", + "int4", "bfloat16", "uint1"] +FORMAT_LIST = ["NCHW", "NHWC", "ND", "NC1HWC0", "FRACTAL_Z", "NC1C0HWPAD", "NHWC1C0", "FSR_NCHW", "FRACTAL_DECONV", + "C1HWNC0", "FRACTAL_DECONV_TRANSPOSE", "FRACTAL_DECONV_SP_STRIDE_TRANS", "NC1HWC0_C04", + "FRACTAL_Z_C04", "CHWN", "FRACTAL_DECONV_SP_STRIDE8_TRANS", "HWCN", "NC1KHKWHWC0", "BN_WEIGHT", + "FILTER_HWCK", "HASHTABLE_LOOKUP_LOOKUPS", "HASHTABLE_LOOKUP_KEYS", "HASHTABLE_LOOKUP_VALUE", + "HASHTABLE_LOOKUP_OUTPUT", "HASHTABLE_LOOKUP_HITS", "C1HWNCoC0", "MD", "NDHWC", "FRACTAL_ZZ", + "FRACTAL_NZ", "NCDHW", "DHWCN", "NDC1HWC0", "FRACTAL_Z_3D", "CN", "NC", "DHWNC", + "FRACTAL_Z_3D_TRANSPOSE", "FRACTAL_ZN_LSTM", "FRACTAL_ZN_RNN", "FRACTAL_Z_G", "NULL"] + + +def parse_ini_files(ini_files): + """ + parse ini files to json + Parameters: + ---------------- + ini_files:input file list + return:ops_info + ---------------- + """ + tbe_ops_info = {} + for ini_file in ini_files: + check_file_size(ini_file) + parse_ini_to_obj(ini_file, tbe_ops_info) + return tbe_ops_info + + +def check_file_size(input_file): + try: + file_size = os.path.getsize(input_file) + except OSError as os_error: + print('[ERROR] Failed to open "%s". %s' % (input_file, str(os_error))) + raise OSError from os_error + if file_size > 10*1024*1024: + print('[WARN] The size of %s exceeds 10MB, it may take more time to run, please wait.' % input_file) + + +def parse_ini_to_obj(ini_file, tbe_ops_info): + """ + parse ini file to json obj + Parameters: + ---------------- + ini_file:ini file path + tbe_ops_info:ops_info + ---------------- + """ + with open(ini_file) as ini_file: + lines = ini_file.readlines() + op_dict = {} + op_name = "" + find_op_type = False + for line in lines: + line = line.rstrip() + if line == "": + continue + if line.startswith("["): + if line.endswith("]"): + op_name = line[1:-1] + op_dict = {} + tbe_ops_info[op_name] = op_dict + find_op_type = True + elif "=" in line: + key1 = line[:line.index("=")] + key2 = line[line.index("=")+1:] + key1_0, key1_1 = key1.split(".") + if key1_0 not in op_dict: + op_dict[key1_0] = {} + if key1_1 in op_dict.get(key1_0): + raise RuntimeError("Op:" + op_name + " " + key1_0 + " " + + key1_1 + " is repeated!") + dic_key = op_dict.get(key1_0) + dic_key[key1_1] = key2 + else: + continue + if not find_op_type: + raise RuntimeError("Not find OpType in .ini file.") + + +def check_output_exist(op_dict, is_valid): + """ + Function Description: + Check output is exist + Parameter: op_dict + Parameter: is_valid + """ + if "output0" in op_dict: + output0_dict = op_dict.get("output0") + if output0_dict.get("name", None) is None: + is_valid = False + print("output0.name is required in .ini file!") + else: + is_valid = False + print("output0 is required in .ini file!") + return is_valid + + +def check_attr_dict(attr_dict, is_valid, attr): + """ + Function Description: + Check attr_dict + Parameter: attr_dict + Parameter: is_valid + Parameter: attr + """ + attr_type = attr_dict.get("type") + value = attr_dict.get("value") + param_type = attr_dict.get("paramType") + if attr_type is None or value is None: + is_valid = False + print("If attr.list is exist, {0}.type and {0}.value is required".format(attr)) + if param_type and param_type not in ATTR_PARAMTYPE_LIST: + is_valid = False + print("{0}.paramType only support {1}.".format(attr, ATTR_PARAMTYPE_LIST)) + if attr_type and attr_type not in ATTR_TYPE_LIST: + is_valid = False + print("{0}.type only support {1}.".format(attr, ATTR_TYPE_LIST)) + return is_valid + + +def check_attr(op_dict, is_valid): + """ + Function Description: + Check attr + Parameter: op_dict + Parameter: is_valid + """ + if "attr" in op_dict: + attr_dict = op_dict.get("attr") + attr_list_str = attr_dict.get("list", None) + if attr_list_str is None: + is_valid = False + print("attr.list is required in .ini file!") + else: + attr_list = attr_list_str.split(",") + for attr_name in attr_list: + attr = "attr_" + attr_name.strip() + attr_dict = op_dict.get(attr) + if attr_dict: + is_valid = check_attr_dict(attr_dict, is_valid, attr) + else: + is_valid = False + print("%s is required in .ini file, when attr.list is %s!" % (attr, attr_list_str)) + return is_valid + + +def check_bool_flag(op_dict, is_valid): + """ + Function Description: + check_bool_flag + Parameter: op_dict + Parameter: is_valid + """ + for key in BOOL_FLAG_KEY: + if key in op_dict: + op_bool_key = op_dict.get(key) + if op_bool_key.get("flag").strip() not in BOOL_LIST: + is_valid = False + print("{0}.flag only support {1}.".format(key, BOOL_LIST)) + return is_valid + + +def check_type_format(op_info, is_valid, op_info_key): + """ + Function Description: + Check type and format + Parameter: op_info + Parameter: is_valid + Parameter: op_info_key + """ + op_info_dtype_str = op_info.get("dtype") + op_info_dtype_num = 0 + op_info_format_num = 0 + if op_info_dtype_str: + op_info_dtype = op_info_dtype_str.split(",") + op_info_dtype_num = len(op_info_dtype) + for dtype in op_info_dtype: + if dtype.strip() not in DTYPE_LIST: + is_valid = False + print("{0}.dtype not support {1}.".format(op_info_key, dtype)) + op_info_format_str = op_info.get("format") + if op_info_format_str: + op_info_format = op_info_format_str.split(",") + op_info_format_num = len(op_info_format) + for op_format in op_info_format: + if op_format.strip() not in FORMAT_LIST: + is_valid = False + print("{0}.format not support {1}.".format(op_info_key, op_format)) + if op_info_dtype_num > 0 and op_info_format_num > 0: + if op_info_dtype_num != op_info_format_num: + is_valid = False + print("The number of {0}.dtype not match the number of {0}.format.".format(op_info_key)) + return is_valid + + +def check_op_info(tbe_ops): + """ + Function Description: + Check info. + Parameter: tbe_ops + Return Value: is_valid + """ + print("\n\n==============check valid for ops info start==============") + required_op_input_info_keys = ["paramType", "name"] + required_op_output_info_keys = ["paramType", "name"] + param_type_valid_value = ["dynamic", "optional", "required"] + is_valid = True + for op_key in tbe_ops: + op_dict = tbe_ops[op_key] + for op_info_key in op_dict: + if op_info_key.startswith("input"): + op_input_info = op_dict[op_info_key] + missing_keys = [] + for required_op_input_info_key in required_op_input_info_keys: + if required_op_input_info_key not in op_input_info: + missing_keys.append(required_op_input_info_key) + if len(missing_keys) > 0: + print("op: " + op_key + " " + op_info_key + " missing: " + + ",".join(missing_keys)) + is_valid = False + else: + if not op_input_info["paramType"] in param_type_valid_value: + print("op: " + op_key + " " + op_info_key + \ + " paramType not valid, valid key:[dynamic, " + "optional, required]") + is_valid = False + is_valid = check_type_format(op_input_info, is_valid, op_info_key) + if op_info_key.startswith("output"): + op_input_info = op_dict[op_info_key] + missing_keys = [] + for required_op_input_info_key in required_op_output_info_keys: + if required_op_input_info_key not in op_input_info: + missing_keys.append(required_op_input_info_key) + if len(missing_keys) > 0: + print("op: " + op_key + " " + op_info_key + " missing: " + + ",".join(missing_keys)) + is_valid = False + else: + if not op_input_info["paramType"] in param_type_valid_value: + print("op: " + op_key + " " + op_info_key + + " paramType not valid, valid key:[dynamic, " + "optional, required]") + is_valid = False + is_valid = check_type_format(op_input_info, is_valid, op_info_key) + is_valid = check_attr(op_dict, is_valid) + is_valid = check_bool_flag(op_dict, is_valid) + print("==============check valid for ops info end================\n\n") + return is_valid + + +def write_json_file(tbe_ops_info, json_file_path): + """ + Save info to json file + Parameters: + ---------------- + tbe_ops_info: ops_info + json_file_path: json file path + ---------------- + """ + json_file_real_path = os.path.realpath(json_file_path) + wr_flag = os.O_WRONLY | os.O_CREAT + wr_mode = stat.S_IWUSR | stat.S_IRUSR + with os.fdopen(os.open(json_file_real_path, wr_flag, wr_mode), 'w') as file_path: + # Only the owner and group have rights + os.chmod(json_file_real_path, stat.S_IWGRP + stat.S_IWUSR + stat.S_IRGRP + + stat.S_IRUSR) + json.dump(tbe_ops_info, file_path, sort_keys=True, indent=4, + separators=(',', ':')) + print("Compile op info cfg successfully.") + + +def parse_ini_to_json(ini_file_paths, outfile_path): + """ + parse ini files to json file + Parameters: + ---------------- + ini_file_paths: list of ini file path + outfile_path: output file path + ---------------- + """ + tbe_ops_info = parse_ini_files(ini_file_paths) + if not check_op_info(tbe_ops_info): + print("Compile op info cfg failed.") + return False + write_json_file(tbe_ops_info, outfile_path) + return True + + +if __name__ == '__main__': + args = sys.argv + + OUTPUT_FILE_PATH = "tbe_ops_info.json" + ini_file_path_list = [] + + for arg in args: + if arg.endswith("ini"): + ini_file_path_list.append(arg) + OUTPUT_FILE_PATH = arg.replace(".ini", ".json") + if arg.endswith("json"): + OUTPUT_FILE_PATH = arg + + if len(ini_file_path_list) == 0: + ini_file_path_list.append("tbe_ops_info.ini") + + if not parse_ini_to_json(ini_file_path_list, OUTPUT_FILE_PATH): + sys.exit(1) + sys.exit(0) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/preset_parse.py b/atb_operator/AtbAddSample/AddOp/cmake/util/preset_parse.py new file mode 100644 index 000000000..8f1124b1d --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/preset_parse.py @@ -0,0 +1,23 @@ +import json +import sys +import os + + +def get_config_opts(file): + src_dir = os.path.abspath(os.path.dirname(file)) + opts = '' + with open(file, 'r') as fd: + config = json.load(fd) + for conf in config: + if conf == 'configurePresets': + for node in config[conf]: + macros = node.get('cacheVariables') + if macros is not None: + for key in macros: + opts += '-D{}={} '.format(key, macros[key]['value']) + opts = opts.replace('${sourceDir}', src_dir) + print(opts) + + +if __name__ == "__main__": + get_config_opts(sys.argv[1]) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/replay_codegen.py b/atb_operator/AtbAddSample/AddOp/cmake/util/replay_codegen.py new file mode 100644 index 000000000..1baa364ef --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/replay_codegen.py @@ -0,0 +1,105 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import os +import stat +import collections +import kernel_entry as keb +from tiling_data_def_build import gen_tiling +import code_channel_infer +import const_var + +PYF_PATH = os.path.dirname(__file__) + +ReplayCodeGenParams = collections.namedtuple('ReplayCodeGenParams',\ +['op_type', 'impl', 'tiling_file', 'kernel', 'entry', 'argn', 'op_replay_batch', 'max_block_dim', 'max_shape_size']) + + +class ReplayCodeGen: + def __init__(self, replayCodeGenParams): + self.op_type = replayCodeGenParams.op_type + self.impl = replayCodeGenParams.impl + self.tiling_file = replayCodeGenParams.tiling_file + self.tiling_data_file = '' + self.kernel = replayCodeGenParams.kernel + self.entry = replayCodeGenParams.entry + self.argn = replayCodeGenParams.argn + self.batch = False + self.outdir = '' + self.data_type = 'uint8_t' + self.blknum = 32 + self.op_replay_batch = replayCodeGenParams.op_replay_batch + self.max_block_dim = replayCodeGenParams.max_block_dim + self.max_shape_size = replayCodeGenParams.max_shape_size + + def set_batch(self, is_batch): + self.batch = is_batch + + def set_outdir(self, outdir): + self.outdir = outdir + + def gen_replay(self, ops_product: str): + kerentry = os.path.join(self.outdir, self.kernel + '_entry.cce') + kerimpl = os.path.join(self.outdir, self.kernel + '_impl.cpp') + replayimpl = os.path.join(self.outdir, self.kernel + '_replay.cpp') + if self.batch: + reptmp = os.path.join(PYF_PATH, 'batch_replay_impl.temp') + else: + reptmp = os.path.join(PYF_PATH, 'replay_impl.temp') + kertmp = os.path.join(PYF_PATH, 'kernel_impl.temp') + self._gen_kentry(kerentry) + self._gen_kimpl_code(kerimpl, kertmp) + self._gen_tiling_data_header() + self._gen_replay_code(replayimpl, reptmp, ops_product) + + def _gen_tiling_data_header(self): + self.tiling_data_file = os.path.join(self.outdir, self.kernel + '_tiling_data.h') + gen_tiling(self.tiling_file, self.tiling_data_file) + + def _gen_kimpl_code(self, src, tmpfile): + with open(tmpfile, 'r') as fd: + temp = fd.read() + temp = temp.replace('__CCE_FILE__', self.impl) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(temp) + + def _gen_replay_code(self, src, tmpfile, ops_product: str): + with open(tmpfile, 'r') as fd: + temp = fd.read() + temp = temp.replace('__ARG_NUM__', str(self.argn)) + argdef = [] + kargs = [] + for i in range(0, self.argn): + argdef.append('{} *'.format(self.data_type)) + kargs.append('({} *)GetArg({})'.format(self.data_type, i)) + temp = temp.replace('__ARGS_DEF__', ', '.join(argdef)) + temp = temp.replace('__KERNEL_ARGS__', ', '.join(kargs)) + temp = temp.replace('__KERNEL_FUN__', self.entry) + core_type_infer = 'core_type' + code_channel = code_channel_infer.infer_code_channel(code_channel_infer.InfoCodeChanelParams(self.impl,\ + self.tiling_data_file, self.kernel, self.outdir, ops_product, None)) + if code_channel == code_channel_infer.CODE_VEC: + core_type_infer = '0' + elif code_channel == code_channel_infer.CODE_CUBE: + core_type_infer = '1' + temp = temp.replace('__CORE_TYPE__', core_type_infer) + # regist function + temp = temp.replace('__OPS_PRODUCT__', ops_product) + temp = temp.replace('__OPTYPE__', self.op_type) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(temp) + + def _gen_kentry(self, src): + kf = '' + pre_alloc_str = 'A' * 256 + if self.batch: + kf += keb.batch_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str), self.argn, self.data_type) + else: + kf += keb.mc_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str),\ + self.argn, self.data_type, self.blknum) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(kf) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/replay_impl.temp b/atb_operator/AtbAddSample/AddOp/cmake/util/replay_impl.temp new file mode 100644 index 000000000..1d30dd865 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/replay_impl.temp @@ -0,0 +1,120 @@ +#include +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#include "register/op_check.h" +#define __ASCENDC_REPLAY_CODE__ +using namespace std; +using namespace optiling; +using namespace AscendCReplay; + +extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); +extern "C" int elf_append(char *elf, uint32_t elfSize, char *jit, int kernum, int blknum[], char *atext[], + int alen[], int atlen, const char* kernelname[]); + +#define KERNEL_N 1 +#define ARG_N (__ARG_NUM__) +#define MAX_L (1024 * 1024 * 100) +#define MAX_E (1024 * 1024) + +int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) +{ + // gen type 1 : direct call codes 0: load .o file + if (param.gentype < 0 || param.gentype > 1) { + printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); + return 0; + } else if (param.gentype == 1 && param.objptr == nullptr) { + printf("Error: call replay with direct call mode, but code obj addr is null\n"); + return 0; + } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { + printf("Error: call replay with object file mode, but object file path is null\n"); + return 0; + } + // core_type 0:MIX 1:CUBE 2:VEC + if (core_type < 0 || core_type > 2) { + printf("Error: call replay core type is %d !\n", core_type); + return 0; + } + g_coreType = __CORE_TYPE__; + g_taskRation = param.task_ration; + g_tilingKey = param.tiling_key; + + unsigned char *buf, *jit; + char *kernel[KERNEL_N * 32]; + int len[KERNEL_N * 32]; + int blknum[KERNEL_N]; + int max; + block_num = param.block_dim; + g_ubBase = block_num; + uint8_t *code = (uint8_t *)malloc(MAX_L); + uint8_t *pos = code; + struct timespec tp1, tp2; + + clock_gettime(CLOCK_MONOTONIC, &tp1); + if (block_num > 32) { + printf("Error: block_num > 32\n"); + return 0; + } + //__OP_FOPEN__ + for (int i = 0; i < KERNEL_N; i++) { + for (int j = 0; j < ARG_N; j++) + AddArg(j, ARG_STEP * (j + 1)); + for (block_idx = 0; block_idx < block_num; block_idx++) { + //__OP_SET_KERNEL__ + int code_idx = i * block_num + block_idx; +#ifdef FP_CEILING + SetCtrlFloatEnable(); +#else + SetCtrlFloatDisable(); +#endif + CodeInit(pos, false); + __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); + CodeEnd(); + kernel[code_idx] = (char *)pos; + len[code_idx] = CodeLen(); + pos += len[code_idx]; + printf("kernel %d core %ld code generated len %d\n", i, block_idx, len[code_idx]); + } + blknum[i] = block_num; + } + //__OP_FCLOSE__ + clock_gettime(CLOCK_MONOTONIC, &tp2); + buf = (unsigned char *)malloc(MAX_E); + int fd = open(param.entry_file, O_RDONLY); + if (fd < 0) { + printf("[error]: cannot find entry.o : %s\n", param.entry_file); + return 0; + } + uint32_t bufSize = read(fd, buf, MAX_E); + if (bufSize <= 0) { + printf("[error]: entry.o : %s is too small ! \n", param.entry_file); + } + close(fd); + jit = (unsigned char *)malloc(MAX_L); + printf("total code generated %ld\n", pos - code); + int sz = elf_append((char *)buf, bufSize, (char *)jit, KERNEL_N, blknum, kernel, len, pos - code, ¶m.kernel_name); + if (tp1.tv_sec != tp2.tv_sec) { + printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); + } else { + printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); + } + printf("new elf size %d\n", sz); + if (param.gentype == 0) { + fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); + (void)write(fd, jit, sz); + close(fd); + free(jit); + } else if (param.gentype == 1) { + *param.objptr = (char*)jit; + } + free(buf); + free(code); + return sz; +} + +REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/tiling_data_def_build.py b/atb_operator/AtbAddSample/AddOp/cmake/util/tiling_data_def_build.py new file mode 100644 index 000000000..a96304261 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/cmake/util/tiling_data_def_build.py @@ -0,0 +1,84 @@ +#!/usr/bin/env python +# coding=utf-8 +""" +Function: +The replay funtion entry +Copyright Information: +Huawei Technologies Co., Ltd. All Rights Reserved © 2020 +""" + +import sys +import os +import stat +import re +import const_var + + +def gen_tiling(tiling_header_file: str, tiling_file_out: str): + if not os.path.exists(tiling_header_file): + print("warning: no userdef tiling header file: ", tiling_header_file) + return + print("generate tiling def header file: ", tiling_file_out) + tmp_name = os.path.splitext(os.path.basename(tiling_header_file))[0].upper() + tiling_source = '#ifndef __{}_H__\n'.format(tmp_name) + tiling_source += '#define __{}_H__\n\n'.format(tmp_name) + tiling_source += '#include \n' + tiling_source += '#include \n\n' + tiling_source += '#include "kernel_tiling/kernel_tiling.h"\n\n' + end_source = "" + pattern = re.compile(r'[(](.*)[)]', re.S) + with open(tiling_header_file, 'r') as fd: + lines = fd.readlines() + for line in lines: + line = line.strip() + if (line.startswith('BEGIN_TILING_DATA_DEF')): + tiling_source += '#pragma pack(1)\n' + tiling_source += 'struct ' + struct_def = re.findall(pattern, line)[0] + tiling_source += struct_def + ' {\n' + elif (line.startswith('TILING_DATA_FIELD_DEF_ARR')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {}[{}] = {{}};\n'.format(fds[0].strip(), fds[2].strip(), fds[1].strip()) + elif (line.startswith('TILING_DATA_FIELD_DEF_STRUCT')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {};\n'.format(fds[0].strip(), fds[1].strip()) + elif (line.startswith('TILING_DATA_FIELD_DEF')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {} = 0;\n'.format(fds[0].strip(), fds[1].strip()) + elif (line.startswith('END_TILING_DATA_DEF')): + tiling_source += '};\n' + tiling_source += '#pragma pack()\n\n' + tiling_source += '#ifdef __NPU_TILING__\n' + tiling_source += \ + 'inline [aicore] void Init{stru}(const __gm__ uint8_t* tiling, {stru}* const_data)\n'\ + .format(stru=struct_def) + tiling_source += '{\n' + tiling_source += ' const __gm__ uint32_t *src = (const __gm__ uint32_t *)tiling;\n' + tiling_source += ' uint32_t *dst = (uint32_t *)const_data;\n' + tiling_source += ' for (auto i = 0; i < sizeof({}) / 4; i++) *(dst + i) = *(src + i);\n'\ + .format(struct_def) + tiling_source += '}\n' + tiling_source += '#else\n' + tiling_source += 'inline void Init{stru}(uint8_t* tiling, {stru}* const_data)\n'.format(stru=struct_def) + tiling_source += '{\n' + tiling_source += ' memcpy(const_data, tiling, sizeof({}));\n'.format(struct_def) + tiling_source += '}\n' + tiling_source += '#endif\n\n' + end_source = ''' +#define GET_TILING_DATA(tiling_data, tiling_arg) \\ +{stru} tiling_data; \\ +Init{stru}(tiling_arg, &tiling_data)\n +'''.format(stru=struct_def) + tiling_source += end_source + tiling_source += '#endif' + with os.fdopen(os.open(tiling_file_out, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(tiling_source) + + +if __name__ == '__main__': + if len(sys.argv) <= 2: + raise RuntimeError('arguments must greater than 2') + gen_tiling(sys.argv[1], sys.argv[2]) diff --git a/atb_operator/AtbAddSample/AddOp/framework/CMakeLists.txt b/atb_operator/AtbAddSample/AddOp/framework/CMakeLists.txt new file mode 100644 index 000000000..b6be9b492 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/framework/CMakeLists.txt @@ -0,0 +1,11 @@ +if(NOT EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/mindspore") + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/caffe_plugin") + add_subdirectory(caffe_plugin) + endif() + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/tf_plugin") + add_subdirectory(tf_plugin) + endif() + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/onnx_plugin") + add_subdirectory(onnx_plugin) + endif() +endif() diff --git a/atb_operator/AtbAddSample/AddOp/framework/tf_plugin/CMakeLists.txt b/atb_operator/AtbAddSample/AddOp/framework/tf_plugin/CMakeLists.txt new file mode 100644 index 000000000..a6aba5c20 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/framework/tf_plugin/CMakeLists.txt @@ -0,0 +1,14 @@ + +aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} plugin_srcs) +add_library(cust_tf_parsers SHARED ${plugin_srcs}) +target_compile_definitions(cust_tf_parsers PRIVATE google=ascend_private) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_tf_parsers PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_tf_parsers PRIVATE intf_pub graph) +install(TARGETS cust_tf_parsers + LIBRARY DESTINATION packages/vendors/${vendor_name}/framework/tensorflow +) diff --git a/atb_operator/AtbAddSample/AddOp/framework/tf_plugin/tensorflow_add_custom_plugin.cc b/atb_operator/AtbAddSample/AddOp/framework/tf_plugin/tensorflow_add_custom_plugin.cc new file mode 100644 index 000000000..2cd837ce5 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/framework/tf_plugin/tensorflow_add_custom_plugin.cc @@ -0,0 +1,23 @@ +/* Copyright (C) 2020-2021. Huawei Technologies Co., Ltd. All +rights reserved. + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the Apache License Version 2.0. + * You may not use this file except in compliance with the License. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * Apache License for more details at + * http://www.apache.org/licenses/LICENSE-2.0 + */ + +#include "register/register.h" + +namespace domi { +// register op info to GE +REGISTER_CUSTOM_OP("AddCustom") + .FrameworkType(TENSORFLOW) // type: CAFFE, TENSORFLOW + .OriginOpType("Add") // name in tf module + .ParseParamsByOperatorFn(AutoMappingByOpFn); +} // namespace domi diff --git a/atb_operator/AtbAddSample/AddOp/op_host/CMakeLists.txt b/atb_operator/AtbAddSample/AddOp/op_host/CMakeLists.txt new file mode 100644 index 000000000..40dd51cfa --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/op_host/CMakeLists.txt @@ -0,0 +1,82 @@ + +aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} ops_srcs) + +opbuild(OPS_SRC ${ops_srcs} + OUT_DIR ${ASCEND_AUTOGEN_PATH} +) + +add_library(cust_op_proto SHARED ${ops_srcs} ${ASCEND_AUTOGEN_PATH}/op_proto.cc) +target_compile_definitions(cust_op_proto PRIVATE OP_PROTO_LIB) +target_compile_options(cust_op_proto PRIVATE + -fvisibility=hidden +) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_op_proto PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_op_proto PRIVATE + intf_pub + exe_graph + register + tiling_api + -Wl,--whole-archive + rt2_registry + -Wl,--no-whole-archive +) +set_target_properties(cust_op_proto PROPERTIES OUTPUT_NAME + cust_opsproto_rt2.0 +) +add_library(cust_optiling SHARED ${ops_srcs}) +target_compile_definitions(cust_optiling PRIVATE OP_TILING_LIB) +target_compile_options(cust_optiling PRIVATE + -fvisibility=hidden +) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_optiling PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_optiling PRIVATE + intf_pub + exe_graph + register + tiling_api + -Wl,--whole-archive + rt2_registry + -Wl,--no-whole-archive +) +set_target_properties(cust_optiling PROPERTIES OUTPUT_NAME + cust_opmaster_rt2.0 +) + +file(GLOB aclnn_src ${ASCEND_AUTOGEN_PATH}/aclnn_*.cpp) +file(GLOB aclnn_inc ${ASCEND_AUTOGEN_PATH}/aclnn_*.h) +add_library(cust_opapi SHARED ${aclnn_src}) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_opapi PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_opapi PRIVATE intf_pub ascendcl nnopbase) + +add_custom_target(optiling_compat ALL + COMMAND ln -sf lib/linux/${CMAKE_SYSTEM_PROCESSOR}/$ + ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so +) + +install(TARGETS cust_op_proto + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_proto/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) +install(FILES ${ASCEND_AUTOGEN_PATH}/op_proto.h + DESTINATION packages/vendors/${vendor_name}/op_proto/inc) +install(TARGETS cust_optiling + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) +install(FILES ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so + DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling) +install(TARGETS cust_opapi + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_api/lib) +install(FILES ${aclnn_inc} + DESTINATION packages/vendors/${vendor_name}/op_api/include) diff --git a/atb_operator/AtbAddSample/AddOp/op_host/add_custom.cpp b/atb_operator/AtbAddSample/AddOp/op_host/add_custom.cpp new file mode 100644 index 000000000..b61a67f11 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/op_host/add_custom.cpp @@ -0,0 +1,117 @@ +#include "add_custom_tiling.h" +#include "register/op_def_registry.h" +#include "graph/utils/type_utils.h" +#include "tiling/platform/platform_ascendc.h" + +namespace optiling { +const uint32_t BLOCK_SIZE = 32; +const uint32_t BUFFER_NUM = 2; +static ge::graphStatus TilingFunc(gert::TilingContext* context) +{ + TilingData tiling; + uint64_t ubSize; + auto ascendcPlatform = platform_ascendc::PlatformAscendC(context->GetPlatformInfo()); + ascendcPlatform.GetCoreMemSize(platform_ascendc::CoreMemType::UB, ubSize); + auto coreNum = ascendcPlatform.GetCoreNum(); + + // Based on the input length and the number of inputs, the number of bytes of the input data type is obtained + uint32_t inputNum = context->GetInputShape(0)->GetStorageShape().GetShapeSize(); + uint32_t typeLength = 0; + ge::TypeUtils::GetDataTypeLength(context->GetInputDesc(0)->GetDataType(), typeLength); + uint32_t inputLength = inputNum * typeLength; + uint32_t inputBytes = inputLength / inputNum; + + // There are a total of 3 shared UB spaces in the input and output. If it's int8, there are 2 more TBUFs + uint32_t ubDataNumber = (inputBytes == 1) ? 5 : 3; + // The number of 32B data blocks that can be used for each data. DOUBLE BUFFER is already counted here + uint32_t tileBlockNum = (ubSize / BLOCK_SIZE / BUFFER_NUM) / ubDataNumber; + uint32_t tileDataNum = (tileBlockNum * BLOCK_SIZE) / inputBytes; + + // Input data for 32B alignment + uint32_t inputLengthAlgin32 = (((inputLength + BLOCK_SIZE - 1) / BLOCK_SIZE) * BLOCK_SIZE); + // There is at least 32B of data on each core, satisfying several settings for several cores. The maximum number of audits is the actual number of audits + coreNum = (coreNum < inputLengthAlgin32 / BLOCK_SIZE) ? coreNum : inputLengthAlgin32 / BLOCK_SIZE; + coreNum = (coreNum >= 1) ? coreNum : 1; + uint32_t everyCoreInputBlockNum = inputLengthAlgin32 / BLOCK_SIZE / coreNum; + uint32_t tailBlockNum = (inputLengthAlgin32 / BLOCK_SIZE) % coreNum; + + // Small chunks are calculated and sliced several times using the number of data on each core + uint32_t smallCoreDataNum = everyCoreInputBlockNum * BLOCK_SIZE / inputBytes; + uint32_t smallTileNum = everyCoreInputBlockNum / tileBlockNum; + uint32_t finalSmallTileNum = (everyCoreInputBlockNum % tileBlockNum) == 0 ? smallTileNum : smallTileNum + 1; + // Tail block calculation for small chunks of data + uint32_t smallTailDataNum = smallCoreDataNum - (tileDataNum * smallTileNum); + smallTailDataNum = smallTailDataNum == 0 ? tileDataNum : smallTailDataNum; + + // The total length of a large block of data is 32B larger than that of a small block of data + everyCoreInputBlockNum += 1; + uint32_t bigCoreDataNum = everyCoreInputBlockNum * BLOCK_SIZE / inputBytes; + uint32_t bigTileNum = everyCoreInputBlockNum / tileBlockNum; + uint32_t finalBigTileNum = (everyCoreInputBlockNum % tileBlockNum) == 0 ? bigTileNum : bigTileNum + 1; + uint32_t bigTailDataNum = bigCoreDataNum - tileDataNum * bigTileNum; + bigTailDataNum = bigTailDataNum == 0 ? tileDataNum : bigTailDataNum; + + tiling.set_smallCoreDataNum(smallCoreDataNum); + tiling.set_bigCoreDataNum(bigCoreDataNum); + tiling.set_tileDataNum(tileDataNum); + tiling.set_smallTailDataNum(smallTailDataNum); + tiling.set_bigTailDataNum(bigTailDataNum); + tiling.set_finalSmallTileNum(finalSmallTileNum); + tiling.set_finalBigTileNum(finalBigTileNum); + tiling.set_tailBlockNum(tailBlockNum); + + context->SetBlockDim(coreNum); + tiling.SaveToBuffer(context->GetRawTilingData()->GetData(), context->GetRawTilingData()->GetCapacity()); + context->GetRawTilingData()->SetDataSize(tiling.GetDataSize()); + size_t *currentWorkspace = context->GetWorkspaceSizes(1); + currentWorkspace[0] = 0; + return ge::GRAPH_SUCCESS; +} +} + +namespace ge { +static ge::graphStatus InferShape(gert::InferShapeContext* context) +{ + const gert::Shape* x1_shape = context->GetInputShape(0); + gert::Shape* y_shape = context->GetOutputShape(0); + *y_shape = *x1_shape; + return GRAPH_SUCCESS; +} +static graphStatus InferDataType(gert::InferDataTypeContext* context) +{ + const auto inputDataType = context->GetInputDataType(0); + context->SetOutputDataType(0, inputDataType); + return ge::GRAPH_SUCCESS; +} +} + +namespace ops { +class AddCustom : public OpDef { +public: + explicit AddCustom(const char* name) : OpDef(name) + { + this->Input("x") + .ParamType(REQUIRED) + .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) + .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) + .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); + this->Input("y") + .ParamType(REQUIRED) + .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) + .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) + .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); + this->Output("z") + .ParamType(REQUIRED) + .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) + .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) + .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); + + this->SetInferShape(ge::InferShape).SetInferDataType(ge::InferDataType); + this->AICore() + .SetTiling(optiling::TilingFunc) + .AddConfig("ascend310b") + .AddConfig("ascend910b"); + } +}; +OP_ADD(AddCustom); +} diff --git a/atb_operator/AtbAddSample/AddOp/op_host/add_custom_tiling.h b/atb_operator/AtbAddSample/AddOp/op_host/add_custom_tiling.h new file mode 100644 index 000000000..28fd00a37 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/op_host/add_custom_tiling.h @@ -0,0 +1,22 @@ +/* + * Copyright (c) Huawei Technologies Co., Ltd. 2022-2023. All rights reserved. + */ +#ifndef ADD_CUSTOM_TILING_H +#define ADD_CUSTOM_TILING_H +#include "register/tilingdata_base.h" + +namespace optiling { +BEGIN_TILING_DATA_DEF(TilingData) + TILING_DATA_FIELD_DEF(uint32_t, smallCoreDataNum); + TILING_DATA_FIELD_DEF(uint32_t, bigCoreDataNum); + TILING_DATA_FIELD_DEF(uint32_t, finalBigTileNum); + TILING_DATA_FIELD_DEF(uint32_t, finalSmallTileNum); + TILING_DATA_FIELD_DEF(uint32_t, tileDataNum); + TILING_DATA_FIELD_DEF(uint32_t, smallTailDataNum); + TILING_DATA_FIELD_DEF(uint32_t, bigTailDataNum); + TILING_DATA_FIELD_DEF(uint32_t, tailBlockNum); +END_TILING_DATA_DEF; + +REGISTER_TILING_DATA_CLASS(AddCustom, TilingData) +} +#endif // ADD_CUSTOM_TILING_H \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AddOp/op_kernel/CMakeLists.txt b/atb_operator/AtbAddSample/AddOp/op_kernel/CMakeLists.txt new file mode 100644 index 000000000..0d31a444c --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/op_kernel/CMakeLists.txt @@ -0,0 +1,61 @@ +# set custom compile options +if ("${CMAKE_BUILD_TYPE}x" STREQUAL "Debugx") + add_ops_compile_options(ALL OPTIONS -g -O0) +endif() + +foreach(compute_unit ${ASCEND_COMPUTE_UNIT}) + + # generate aic-${compute_unit}-ops-info.json + add_ops_info_target(TARGET ops_info_gen_${compute_unit} + OUTPUT ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core/${compute_unit}/aic-${compute_unit}-ops-info.json + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/config/${compute_unit} + ) + + # generate ascendc impl py once + if (NOT TARGET ascendc_impl_gen) + add_ops_impl_target(TARGET ascendc_impl_gen + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl + ) + endif() + + # dynamic shape binary compile + if (${ENABLE_BINARY_PACKAGE}) + add_bin_compile_target(TARGET ascendc_bin_${compute_unit} + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} + ADP_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/dynamic + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/binary/${compute_unit} + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/kernel + COMPUTE_UNIT ${compute_unit} + ) + add_dependencies(ascendc_bin_${compute_unit} ascendc_impl_gen) + endif() + +endforeach() + +# generate npu_supported_ops.json +add_npu_support_target(TARGET npu_supported_ops + OPS_INFO_DIR ${ASCEND_AUTOGEN_PATH} + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core + INSTALL_DIR packages/vendors/${vendor_name}/framework/${ASCEND_FRAMEWORK_TYPE} +) + +if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) + add_subdirectory(testcases) +endif() + +# install kernel file +if (${ENABLE_SOURCE_PACKAGE}) + file(GLOB KERNEL_FILES + ${CMAKE_CURRENT_SOURCE_DIR}/*.cpp + ${CMAKE_CURRENT_SOURCE_DIR}/*.h + ${CMAKE_CURRENT_SOURCE_DIR}/*.py + ) + install(FILES ${KERNEL_FILES} + DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl/dynamic + ) +endif() diff --git a/atb_operator/AtbAddSample/AddOp/op_kernel/add_custom.cpp b/atb_operator/AtbAddSample/AddOp/op_kernel/add_custom.cpp new file mode 100644 index 000000000..37c80a825 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/op_kernel/add_custom.cpp @@ -0,0 +1,134 @@ +/* + * Copyright (c) Huawei Technologies Co., Ltd. 2022-2023. All rights reserved. + * + * Function : z = x + y + * This sample is a very basic sample that implements vector add on Ascend plaform. + */ +#include "kernel_operator.h" +// tensor num for each queue +constexpr int32_t BUFFER_NUM = 2; + +template class KernelAdd { + using T = TYPE_X; +public: + __aicore__ inline KernelAdd() {} + __aicore__ inline void Init(GM_ADDR x, GM_ADDR y, GM_ADDR z, uint32_t smallCoreDataNum, + uint32_t bigCoreDataNum, uint32_t finalBigTileNum, + uint32_t finalSmallTileNum, uint32_t tileDataNum, + uint32_t smallTailDataNum, uint32_t bigTailDataNum, + uint32_t tailBlockNum) + { + ASSERT(AscendC::GetBlockNum() != 0 && "block dim can not be zero!"); + uint32_t coreNum = AscendC::GetBlockIdx(); + uint32_t globalBufferIndex = bigCoreDataNum * AscendC::GetBlockIdx(); + this->tileDataNum = tileDataNum; + if (coreNum < tailBlockNum) { + this->coreDataNum = bigCoreDataNum; + this->tileNum = finalBigTileNum; + this->tailDataNum = bigTailDataNum; + } + else { + this->coreDataNum = smallCoreDataNum; + this->tileNum = finalSmallTileNum; + this->tailDataNum = smallTailDataNum; + globalBufferIndex -= (bigCoreDataNum - smallCoreDataNum) * (AscendC::GetBlockIdx() - tailBlockNum); + } + xGm.SetGlobalBuffer((__gm__ TYPE_X*)x + globalBufferIndex, this->coreDataNum); + yGm.SetGlobalBuffer((__gm__ TYPE_Y*)y + globalBufferIndex, this->coreDataNum); + zGm.SetGlobalBuffer((__gm__ TYPE_Z*)z + globalBufferIndex, this->coreDataNum); + pipe.InitBuffer(inQueueX, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_X)); + pipe.InitBuffer(inQueueY, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_Y)); + pipe.InitBuffer(outQueueZ, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_Z)); + pipe.InitBuffer(tmp1, this->tileDataNum * sizeof(half)); + pipe.InitBuffer(tmp2, this->tileDataNum * sizeof(half)); + } + __aicore__ inline void Process() + { + int32_t loopCount = this->tileNum; + this->processDataNum = this->tileDataNum; + for (int32_t i = 0; i < loopCount; i++) { + if (i == this->tileNum - 1) { + this->processDataNum = this->tailDataNum; + } + CopyIn(i); + Compute(i); + CopyOut(i); + } + } + +private: + __aicore__ inline void CopyIn(int32_t progress) + { + AscendC::LocalTensor xLocal = inQueueX.AllocTensor(); + AscendC::LocalTensor yLocal = inQueueY.AllocTensor(); + AscendC::DataCopy(xLocal, xGm[progress * this->tileDataNum], this->processDataNum); + AscendC::DataCopy(yLocal, yGm[progress * this->tileDataNum], this->processDataNum); + inQueueX.EnQue(xLocal); + inQueueY.EnQue(yLocal); + } + __aicore__ inline void Compute(int32_t progress) + { + AscendC::LocalTensor xLocal = inQueueX.DeQue(); + AscendC::LocalTensor yLocal = inQueueY.DeQue(); + AscendC::LocalTensor zLocal = outQueueZ.AllocTensor(); + if constexpr (std::is_same_v) { + auto p1 = tmp1.Get(); + auto p2 = tmp2.Get(); + AscendC::Cast(p1, xLocal, AscendC::RoundMode::CAST_NONE, this->processDataNum); + AscendC::Cast(p2, yLocal, AscendC::RoundMode::CAST_NONE, this->processDataNum); + AscendC::Add(p2, p1, p2, this->processDataNum); + AscendC::Cast(p1.ReinterpretCast(), p2, AscendC::RoundMode::CAST_RINT, this->processDataNum); + AscendC::ShiftLeft(p1.ReinterpretCast(), p1.ReinterpretCast(), int16_t(8), this->processDataNum); + AscendC::ShiftRight(p1.ReinterpretCast(), p1.ReinterpretCast(), int16_t(8), this->processDataNum); + AscendC::Cast(p2, p1.ReinterpretCast(), AscendC::RoundMode::CAST_NONE, this->processDataNum); + AscendC::Cast(zLocal, p2, AscendC::RoundMode::CAST_NONE, this->processDataNum); + } + else { + AscendC::Add(zLocal, xLocal, yLocal, this->processDataNum); + } + outQueueZ.EnQue(zLocal); + inQueueX.FreeTensor(xLocal); + inQueueY.FreeTensor(yLocal); + } + __aicore__ inline void CopyOut(int32_t progress) + { + AscendC::LocalTensor zLocal = outQueueZ.DeQue(); + AscendC::DataCopy(zGm[progress * this->tileDataNum], zLocal, this->processDataNum); + outQueueZ.FreeTensor(zLocal); + } + +private: + AscendC::TPipe pipe; + AscendC::TQue inQueueX, inQueueY; + AscendC::TQue outQueueZ; + AscendC::TBuf tmp1, tmp2; + AscendC::GlobalTensor xGm; + AscendC::GlobalTensor yGm; + AscendC::GlobalTensor zGm; + uint32_t coreDataNum; + uint32_t tileNum; + uint32_t tileDataNum; + uint32_t tailDataNum; + uint32_t processDataNum; +}; + +extern "C" __global__ __aicore__ void add_custom(GM_ADDR x, GM_ADDR y, GM_ADDR z, GM_ADDR workspace, GM_ADDR tiling) +{ + GET_TILING_DATA(tiling_data, tiling); + KernelAdd op; + op.Init(x, y, z, tiling_data.smallCoreDataNum, + tiling_data.bigCoreDataNum, tiling_data.finalBigTileNum, + tiling_data.finalSmallTileNum, tiling_data.tileDataNum, + tiling_data.smallTailDataNum, tiling_data.bigTailDataNum, + tiling_data.tailBlockNum); + op.Process(); +} + +#ifndef ASCENDC_CPU_DEBUG +// call of kernel function +void add_custom_do(uint32_t blockDim, void* l2ctrl, void* stream, uint8_t* x, uint8_t* y, uint8_t* z, + uint8_t* workspace, uint8_t* tiling) +{ + add_custom<<>>(x, y, z, workspace, tiling); +} +#endif diff --git a/atb_operator/AtbAddSample/AddOp/scripts/install.sh b/atb_operator/AtbAddSample/AddOp/scripts/install.sh new file mode 100644 index 000000000..9d45a8052 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/scripts/install.sh @@ -0,0 +1,318 @@ +#!/bin/bash +vendor_name=customize +targetdir=/usr/local/Ascend/opp +target_custom=0 + +sourcedir=$PWD/packages +vendordir=vendors/$vendor_name + +QUIET="y" + +while true +do + case $1 in + --quiet) + QUIET="y" + shift + ;; + --install-path=*) + INSTALL_PATH=$(echo $1 | cut -d"=" -f2-) + INSTALL_PATH=${INSTALL_PATH%*/} + shift + ;; + --*) + shift + ;; + *) + break + ;; + esac +done + +log() { + cur_date=`date +"%Y-%m-%d %H:%M:%S"` + echo "[runtime] [$cur_date] "$1 +} + +if [ -n "${INSTALL_PATH}" ]; then + if [[ ! "${INSTALL_PATH}" = /* ]]; then + log "[ERROR] use absolute path for --install-path argument" + exit 1 + fi + if [ ! -d ${INSTALL_PATH} ]; then + mkdir ${INSTALL_PATH} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${INSTALL_PATH} failed" + exit 1 + fi + fi + targetdir=${INSTALL_PATH} +elif [ -n "${ASCEND_CUSTOM_OPP_PATH}" ]; then + if [ ! -d ${ASCEND_CUSTOM_OPP_PATH} ]; then + mkdir -p ${ASCEND_CUSTOM_OPP_PATH} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${ASCEND_CUSTOM_OPP_PATH} failed" + fi + fi + targetdir=${ASCEND_CUSTOM_OPP_PATH} +else + if [ "x${ASCEND_OPP_PATH}" == "x" ]; then + log "[ERROR] env ASCEND_OPP_PATH no exist" + exit 1 + fi + targetdir="${ASCEND_OPP_PATH}" +fi + +if [ ! -d $targetdir ];then + log "[ERROR] $targetdir no exist" + exit 1 +fi + +upgrade() +{ + if [ ! -d ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 files" + return 0 + fi + + if [ ! -d ${targetdir}/$vendordir/$1 ];then + log "[INFO] create ${targetdir}/$vendordir/$1." + mkdir -p ${targetdir}/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/$1 failed" + return 1 + fi + else + has_same_file=-1 + for file_a in ${sourcedir}/$vendordir/$1/*; do + file_b=${file_a##*/}; + if [ "ls ${targetdir}/$vendordir/$1" = "" ]; then + log "[INFO] ${targetdir}/$vendordir/$1 is empty !!" + return 1 + fi + grep -q $file_b <<<`ls ${targetdir}/$vendordir/$1`; + if [[ $? -eq 0 ]]; then + echo -n "${file_b} " + has_same_file=0 + fi + done + if [ 0 -eq $has_same_file ]; then + if test $QUIET = "n"; then + echo "[INFO]: has old version in ${targetdir}/$vendordir/$1, \ + you want to Overlay Installation , please enter:[o]; \ + or replace directory installation , please enter: [r]; \ + or not install , please enter:[n]." + + while true + do + read orn + if [ "$orn" = n ]; then + return 0 + elif [ "$orn" = m ]; then + break; + elif [ "$0rn" = r ]; then + [ -n "${targetdir}/$vendordir/$1/" ] && rm -rf "${targetdir}/$vendordir/$1"/* + break; + else + echo "[ERROR] input error, please input again!" + fi + done + fi + fi + log "[INFO] replace or merge old ops $1 files .g....." + fi + + log "copy new ops $1 files ......" + if [ -d ${targetdir}/$vendordir/$1/ ]; then + chmod -R +w "$targetdir/$vendordir/$1/" >/dev/null 2>&1 + fi + cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 files failed" + return 1 + fi + + return 0 +} +upgrade_proto() +{ + if [ ! -f ${sourcedir}/$vendordir/custom.proto ]; then + log "[INFO] no need to upgrade custom.proto files" + return 0 + fi + if [ ! -d ${targetdir}/$vendordir/framework/caffe ];then + log "[INFO] create ${targetdir}/$vendordir/framework/caffe." + mkdir -p ${targetdir}/$vendordir/framework/caffe + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/framework/caffe failed" + return 1 + fi + else + if [ -f ${targetdir}/$vendordir/framework/caffe/custom.proto ]; then + # 有老版本,判断是否要覆盖式安装 + if test $QUIET = "n"; then + echo "[INFO] ${targetdir}/$vendordir/framework/caffe has old version"\ + "custom.proto file. Do you want to replace? [y/n] " + + while true + do + read yn + if [ "$yn" = n ]; then + return 0 + elif [ "$yn" = y ]; then + break; + else + echo "[ERROR] input error, please input again!" + fi + done + fi + fi + log "[INFO] replace old caffe.proto files ......" + fi + chmod -R +w "$targetdir/$vendordir/framework/caffe/" >/dev/null 2>&1 + cp -rf ${sourcedir}/$vendordir/custom.proto ${targetdir}/$vendordir/framework/caffe/ + if [ $? -ne 0 ];then + log "[ERROR] copy new custom.proto failed" + return 1 + fi + log "[INFO] copy custom.proto success" + + return 0 +} + +upgrade_file() +{ + if [ ! -e ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 file" + return 0 + fi + + log "copy new $1 files ......" + cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 file failed" + return 1 + fi + + return 0 +} + +delete_optiling_file() +{ + if [ ! -d ${targetdir}/vendors ];then + log "[INFO] $1 not exist, no need to uninstall" + return 0 + fi + sys_info=$(uname -m) + if [ ! -d ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/lib/linux/${sys_info} ];then + rm -rf ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/liboptiling.so + fi + return 0 +} + +log "[INFO] copy uninstall sh success" + +if [ ! -d ${targetdir}/vendors ];then + log "[INFO] create ${targetdir}/vendors." + mkdir -p ${targetdir}/vendors + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/vendors failed" + return 1 + fi +fi +chmod u+w ${targetdir}/vendors + +echo "[ops_custom]upgrade framework" +upgrade framework +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op proto" +upgrade op_proto +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade version.info" +upgrade_file version.info +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op impl" +delete_optiling_file op_impl +upgrade op_impl +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op api" +upgrade op_api +if [ $? -ne 0 ];then + exit 1 +fi + +upgrade_proto +if [ $? -ne 0 ];then + exit 1 +fi + +# set the set_env.bash +if [ -n "${INSTALL_PATH}" ] && [ -d ${INSTALL_PATH} ]; then + _ASCEND_CUSTOM_OPP_PATH=${targetdir}/${vendordir} + bin_path="${_ASCEND_CUSTOM_OPP_PATH}/bin" + set_env_variable="#!/bin/bash\nexport ASCEND_CUSTOM_OPP_PATH=${_ASCEND_CUSTOM_OPP_PATH}:\${ASCEND_CUSTOM_OPP_PATH}" + if [ ! -d ${bin_path} ]; then + mkdir -p ${bin_path} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${bin_path} failed" + exit 1 + fi + fi + echo -e ${set_env_variable} > ${bin_path}/set_env.bash + if [ $? -ne 0 ]; then + log "[ERROR] write ASCEND_CUSTOM_OPP_PATH to set_env.bash failed" + exit 1 + else + log "[INFO] using requirements: when custom module install finished or before you run the custom module, \ + execute the command [ source ${bin_path}/set_env.bash ] to set the environment path" + fi +else + config_file=${targetdir}/vendors/config.ini + if [ ! -f ${config_file} ]; then + touch ${config_file} + chmod 640 ${config_file} + echo "load_priority=$vendor_name" > ${config_file} + if [ $? -ne 0 ];then + echo "echo load_priority failed" + exit 1 + fi + else + found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" + found_vendor=$(echo $found_vendors | sed "s/\<$vendor_name\>//g" | tr ',' ' ') + vendor=$(echo $found_vendor | tr -s ' ' ',') + if [ "$vendor" != "" ]; then + sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" + fi + fi +fi + +chmod u-w ${targetdir}/vendors + +if [ -d ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/ ]; then + chmod -R 440 ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/* >/dev/null 2>&1 +fi +if [ -f ${targetdir}/ascend_install.info ]; then + chmod -R 440 ${targetdir}/ascend_install.info +fi +if [ -f ${targetdir}/scene.info ]; then + chmod -R 440 ${targetdir}/scene.info +fi +if [ -f ${targetdir}/version.info ]; then + chmod -R 440 ${targetdir}/version.info +fi + +echo "SUCCESS" +exit 0 + diff --git a/atb_operator/AtbAddSample/AddOp/scripts/upgrade.sh b/atb_operator/AtbAddSample/AddOp/scripts/upgrade.sh new file mode 100644 index 000000000..34801efc4 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOp/scripts/upgrade.sh @@ -0,0 +1,151 @@ +#!/bin/bash +vendor_name=customize +targetdir=/usr/local/Ascend/opp +target_custom=0 + +sourcedir=$PWD/packages +vendordir=vendors/$vendor_name + +log() { + cur_date=`date +"%Y-%m-%d %H:%M:%S"` + echo "[runtime] [$cur_date] "$1 +} + +if [[ "x${ASCEND_OPP_PATH}" == "x" ]];then + log "[ERROR] env ASCEND_OPP_PATH no exist" + exit 1 +fi + +targetdir=${ASCEND_OPP_PATH} + +if [ ! -d $targetdir ];then + log "[ERROR] $targetdir no exist" + exit 1 +fi + +upgrade() +{ + if [ ! -d ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 files" + return 0 + fi + + if [ ! -d ${targetdir}/$vendordir/$1 ];then + log "[INFO] create ${targetdir}/$vendordir/$1." + mkdir -p ${targetdir}/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/$1 failed" + return 1 + fi + else + vendor_installed_dir=$(ls "$targetdir/vendors" 2> /dev/null) + for i in $vendor_installed_dir;do + vendor_installed_file=$(ls "$vendor_installed_dir/$vendor_name/$i" 2> /dev/null) + if [ "$i" = "$vendor_name" ] && [ "$vendor_installed_file" != "" ]; then + echo "[INFO]: $vendor_name custom opp package has been installed on the path $vendor_installed_dir, \ + you want to Overlay Installation , please enter:[o]; \ + or replace directory installation , please enter: [r]; \ + or not install , please enter:[n]." + fi + while true + do + read mrn + if [ "$mrn" = m ]; then + break + elif [ "$mrn" = r ]; then + [ -n "$vendor_installed_file"] && rm -rf "$vendor_installed_file" + break + elif [ "$mrn" = n ]; then + return 0 + else + echo "[WARNING]: Input error, please input m or r or n to choose!" + fi + done + done + log "[INFO] replace old ops $1 files ......" + fi + + log "copy new ops $1 files ......" + cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 files failed" + return 1 + fi + + return 0 +} + +upgrade_file() +{ + if [ ! -e ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 file" + return 0 + fi + + log "copy new $1 files ......" + cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 file failed" + return 1 + fi + + return 0 +} + +log "[INFO] copy uninstall sh success" + +echo "[ops_custom]upgrade framework" +upgrade framework +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op proto" +upgrade op_proto +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op impl" +upgrade op_impl +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op api" +upgrade op_api +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade version.info" +upgrade_file version.info +if [ $? -ne 0 ];then + exit 1 +fi + +config_file=${targetdir}/vendors/config.ini +found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" +found_vendor=$(echo $found_vendors | sed "s/\<$vendor_name\>//g" | tr ',' ' ') +vendor=$(echo $found_vendor | tr -s ' ' ',') +if [ "$vendor" != "" ]; then + sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" +fi + +changemode() +{ + if [ -d ${targetdir} ];then + chmod -R 550 ${targetdir}>/dev/null 2>&1 + fi + + return 0 +} +echo "[ops_custom]changemode..." +#changemode +if [ $? -ne 0 ];then + exit 1 +fi + +echo "SUCCESS" +exit 0 + diff --git a/atb_operator/AtbAddSample/readme.md b/atb_operator/AtbAddSample/readme.md new file mode 100644 index 000000000..e69de29bb diff --git a/atb_operator/AtbEyeSample/EyeOp/CMakeLists.txt b/atb_operator/AtbEyeSample/EyeOp/CMakeLists.txt new file mode 100755 index 000000000..584132d80 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/CMakeLists.txt @@ -0,0 +1,69 @@ +cmake_minimum_required(VERSION 3.16.0) +project(opp) +if(ENABLE_CROSS_COMPILE) + if(${CMAKE_SYSTEM_PROCESSOR} STREQUAL x86_64) + set(CROSS_COMPILE_PLATFORM aarch64) + else() + set(CROSS_COMPILE_PLATFORM x86_64) + endif() + set(PLATFORM ${CMAKE_SYSTEM_PROCESSOR}) + set(CMAKE_COMPILE_COMPILER_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/linux/${CROSS_COMPILE_PLATFORM}/) + set(CMAKE_COMPILE_RUNTIME_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/${CROSS_COMPILE_PLATFORM}/) + set(CMAKE_SYSTEM_PROCESSOR ${CROSS_COMPILE_PLATFORM}) + set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) + set(CMAKE_CXX_COMPILER ${CMAKE_CROSS_PLATFORM_COMPILER}) +else() + set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) +endif() + +include(cmake/config.cmake) +include(cmake/func.cmake) +include(cmake/intf.cmake) + +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/framework) + add_subdirectory(framework) +endif() +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_host) + add_subdirectory(op_host) +endif() +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_kernel) + add_subdirectory(op_kernel) +endif() +if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) + add_subdirectory(testcases) +endif() + +# modify vendor_name in install.sh and upgrade.sh +add_custom_command(OUTPUT ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh + COMMAND mkdir -p ${CMAKE_BINARY_DIR}/scripts + COMMAND cp -r ${CMAKE_SOURCE_DIR}/scripts/* ${CMAKE_BINARY_DIR}/scripts/ + COMMAND sed -i "s/vendor_name=customize/vendor_name=${vendor_name}/g" ${CMAKE_BINARY_DIR}/scripts/* +) +add_custom_target(modify_vendor ALL DEPENDS ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh) +install(DIRECTORY ${CMAKE_BINARY_DIR}/scripts/ DESTINATION . FILE_PERMISSIONS OWNER_EXECUTE OWNER_READ GROUP_READ) + +install(FILES ${CMAKE_SOURCE_DIR}/custom.proto DESTINATION packages OPTIONAL) + +get_system_info(SYSTEM_INFO) + +# gen version.info +add_custom_target(gen_version_info ALL + COMMAND bash ${CMAKE_CURRENT_SOURCE_DIR}/cmake/util/gen_version_info.sh ${ASCEND_CANN_PACKAGE_PATH} ${CMAKE_CURRENT_BINARY_DIR} +) + +install(FILES ${CMAKE_CURRENT_BINARY_DIR}/version.info + DESTINATION packages/vendors/${vendor_name}/) + +# CPack config +set(CPACK_PACKAGE_NAME ${CMAKE_PROJECT_NAME}) +set(CPACK_PACKAGE_VERSION ${CMAKE_PROJECT_VERSION}) +set(CPACK_PACKAGE_DESCRIPTION "CPack opp project") +set(CPACK_PACKAGE_DESCRIPTION_SUMMARY "CPack opp project") +set(CPACK_PACKAGE_DIRECTORY ${CMAKE_INSTALL_PREFIX}) +set(CPACK_PACKAGE_FILE_NAME "custom_opp_${SYSTEM_INFO}.run") +set(CPACK_GENERATOR External) +set(CPACK_CMAKE_GENERATOR "Unix Makefiles") +set(CPACK_EXTERNAL_ENABLE_STAGING TRUE) +set(CPACK_EXTERNAL_PACKAGE_SCRIPT ${CMAKE_SOURCE_DIR}/cmake/makeself.cmake) +set(CPACK_EXTERNAL_BUILT_PACKAGES ${CPACK_PACKAGE_DIRECTORY}/_CPack_Packages/Linux/External/${CPACK_PACKAGE_FILE_NAME}/${CPACK_PACKAGE_FILE_NAME}) +include(CPack) diff --git a/atb_operator/AtbEyeSample/EyeOp/CMakePresets.json b/atb_operator/AtbEyeSample/EyeOp/CMakePresets.json new file mode 100755 index 000000000..d32bf6d9a --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/CMakePresets.json @@ -0,0 +1,63 @@ +{ + "version": 1, + "cmakeMinimumRequired": { + "major": 3, + "minor": 19, + "patch": 0 + }, + "configurePresets": [ + { + "name": "default", + "displayName": "Default Config", + "description": "Default build using Unix Makefiles generator", + "generator": "Unix Makefiles", + "binaryDir": "${sourceDir}/build_out", + "cacheVariables": { + "CMAKE_BUILD_TYPE": { + "type": "STRING", + "value": "Release" + }, + "ENABLE_SOURCE_PACKAGE": { + "type": "BOOL", + "value": "True" + }, + "ENABLE_BINARY_PACKAGE": { + "type": "BOOL", + "value": "True" + }, + "ASCEND_COMPUTE_UNIT": { + "type": "STRING", + "value": "ascend310b;ascend910b" + }, + "ENABLE_TEST": { + "type": "BOOL", + "value": "True" + }, + "vendor_name": { + "type": "STRING", + "value": "customize" + }, + "ASCEND_CANN_PACKAGE_PATH": { + "type": "PATH", + "value": "/usr/local/Ascend/ascend-toolkit/latest" + }, + "ASCEND_PYTHON_EXECUTABLE": { + "type": "STRING", + "value": "python3" + }, + "CMAKE_INSTALL_PREFIX": { + "type": "PATH", + "value": "${sourceDir}/build_out" + }, + "ENABLE_CROSS_COMPILE": { + "type": "BOOL", + "value": "False" + }, + "CMAKE_CROSS_PLATFORM_COMPILER": { + "type": "PATH", + "value": "/usr/bin/aarch64-linux-gnu-g++" + } + } + } + ] +} \ No newline at end of file diff --git a/atb_operator/AtbEyeSample/EyeOp/README.md b/atb_operator/AtbEyeSample/EyeOp/README.md new file mode 100644 index 000000000..5b2d1c7e6 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/README.md @@ -0,0 +1,216 @@ +# 概述 +本样例基于EyeOp算子工程,介绍了单算子工程及aclnn接口说明。 + +## 算子工程介绍 +### 算子描述 +创建batch_shape个二维矩阵 $m\times n$ ,对角元素全为1,其它元素都为0。 + +### 算子规格描述 + + + + + + + + + + + + + + + + + + + + + + + + + + +
算子类型(OpType)Eye
nameshapedata typeformat默认值
算子输入y-float32, float16, int32ND\
算子输出y-float32, float16, int32ND\
attr属性num_rows\int\\
num_columns\int\0
batch_shape\list_int\{1}
dtype\int\0
核函数名eye
+ + +### 算子工程文件介绍 +其中,算子工程目录EyeOp包含算子实现的模板文件、编译脚本等,如下所示: +``` +├── EyeOp //Add自定义算子工程 +│ ├── cmake +│ ├── framework // 算子插件实现文件目录,单算子模型文件的生成不依赖算子适配插件,无需关注 +│ ├── op_host // host侧实现文件 +│ ├── op_kernel // kernel侧实现文件 +│ ├── scripts // 自定义算子工程打包相关脚本所在目录 +│ ├── build.sh // 编译入口脚本 +│ ├── CMakeLists.txt // 算子工程的CMakeLists.txt +│ └── CMakePresets.json // 编译配置项 +``` + +### 编译运行样例算子 +针对自定义算子工程,编译运行包含如下步骤: +- 编译自定义算子工程生成算子安装包; +- 安装自定义算子到算子库中; +- 调用执行自定义算子; + +详细操作如下所示。 +#### 1. 获取源码包 +编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 +#### 2. 编译算子工程 + 编译自定义算子工程,构建生成自定义算子包。 + + - 执行如下命令,切换到算子工程EyeOp目录。 + + ```bash + cd ${git_clone_path}/samples/atb_operator/AtbEyeSample/EyeOp + ``` + + - 修改CMakePresets.json中ASCEND_CANN_PACKAGE_PATH为CANN软件包安装后的实际路径。 + + + ```json + { + …… + "configurePresets": [ + { + …… + "ASCEND_CANN_PACKAGE_PATH": { + "type": "PATH", + "value": "/usr/local/Ascend/ascend-toolkit/latest" //请替换为CANN软件包安装后的实际路径。eg:/home/HwHiAiUser/Ascend/ascend-toolkit/latest + }, + …… + } + ] + } + ``` + - 在算子工程EyeOp目录下执行如下命令,进行算子工程编译。 + + ```bash + ./build.sh + ``` +编译成功后,会在当前目录下创建build_out目录,并在build_out目录下生成自定义算子安装包custom_opp_\_\.run,例如“custom_opp_ubuntu_x86_64.run”。 + +备注:如果要使用dump调试功能,需要移除op_host内和CMakeLists.txt内的Atlas 200/500 A2 推理产品的配置。 + +#### 3. 部署算子包 + +执行如下命令,在自定义算子安装包所在路径下,安装自定义算子包。 + ```bash + cd build_out + ./custom_opp__.run + ``` +命令执行成功后,自定义算子包中的相关文件将部署至当前环境的OPP算子库的vendors/customize目录中。 + +#### 4. 配置环境变量 + + 请根据当前环境上CANN开发套件包的[安装方式](https://hiascend.com/document/redirect/CannCommunityInstSoftware),选择对应配置环境变量的命令。 + - 默认路径,root用户安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=/usr/local/Ascend/ascend-toolkit/latest + ``` + - 默认路径,非root用户安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=$HOME/Ascend/ascend-toolkit/latest + ``` + - 指定路径install_path,安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=${install_path}/ascend-toolkit/latest + ``` + +## aclnn接口说明 + +自定义算子编译部署后,会自动生成单算子API,可以直接在应用程序中调用。算子API的形式一般定义为“两段式接口”,形如: + ```cpp + aclnnStatus aclnnEyeGetWorkspaceSize(const aclTensor *x, int num_rows, int num_columns, int *batch_shape, int dtype, uint64_t workspaceSize, aclOpExecutor **executor); + aclnnStatus aclnnEye(void *workspace, int64_t workspaceSize, aclOpExecutor **executor, aclrtStream stream); + ``` +其中aclnnEyeGetWorkspaceSize为第一段接口,主要用于计算本次API调用计算过程中需要多少的workspace内存。获取到本次API计算需要的workspace大小之后,按照workspaceSize大小申请Device侧内存,然后调用第二段接口aclnnEye执行计算。具体参考[AscendCL单算子调用](https://hiascend.com/document/redirect/CannCommunityAscendCInVorkSingleOp)>单算子API执行 章节。 + +### 功能描述 +* 算子功能:创建batch_shape个二维矩阵 $m\times n$ ,对角元素全为1,其它元素都为0。 + +#### 参数说明 +##### aclnnEyeGetWorkspaceSize: + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
接口参数描述
aclnnEyeGetWorkspaceSizex输入输出tensor,支持flaot/half/int32类型,ND排布格式
num_rows必选属性,生成矩阵的行数,支持int32类型数值
num_columns可选属性,生成矩阵的列数,支持int32类型数值,默认与num_rows一致
batch_shape可选属性,数组内元素乘积为生成二维矩阵的个数,支持int数组,默认为{}
dtype可选属性,输出tensor的数据类型,当前输入输出同地址,同数据类型,该参数无作用,支持int32类型数值, 默认为0
workspaceSize返回需要在Device侧申请的workspace大小
executor返回op执行器,包含了算子计算流程
+ + +| 返回值 | 描述 | +| ---- | ---- | +| ACL_SUCCESS | 成功 | +| ACL_ERROR | 失败 | + +##### aclnnEye + + + + + + + + + + + + + + + + + + + + + + + +
接口参数描述
aclnnEyeworkspace输入Device侧申请的workspace内存
workspaceSizeworkspace大小
executor输入op执行器,包含了算子计算流程
stream输入执行流
+ +| 返回值 | 描述 | +| ---- | ---- | +| ACL_SUCCESS | 成功 | +| ACL_ERROR | 失败 | + + +## 更新说明 +| 时间 | 更新事项 | +|----|------| +| 2025/03/27 | 新增README | \ No newline at end of file diff --git a/atb_operator/AtbEyeSample/EyeOp/build.sh b/atb_operator/AtbEyeSample/EyeOp/build.sh new file mode 100755 index 000000000..27d1968d1 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/build.sh @@ -0,0 +1,80 @@ +#!/bin/bash +script_path=$(realpath $(dirname $0)) + + +mkdir -p build_out +rm -rf build_out/* +cd build_out + +opts=$(python3 $script_path/cmake/util/preset_parse.py $script_path/CMakePresets.json) +ENABLE_CROSS="-DENABLE_CROSS_COMPILE=True" +ENABLE_BINARY="-DENABLE_BINARY_PACKAGE=True" +cmake_version=$(cmake --version | grep "cmake version" | awk '{print $3}') + +cmake_run_package() +{ + target=$1 + cmake --build . --target $target -j16 + if [ $? -ne 0 ]; then exit 1; fi + + if [ $target = "package" ]; then + if test -d ./op_kernel/binary ; then + ./cust*.run + if [ $? -ne 0 ]; then exit 1; fi + cmake --build . --target binary -j16 + if [ $? -ne 0 ]; then + echo "[ERROR] Kernel compile failed, the run package will not be generated." + rm -rf ./cust*.run && rm -rf ./cust*.run.json && exit 1; + fi + cmake --build . --target $target -j16 + fi + fi +} + +if [[ $opts =~ $ENABLE_CROSS ]] && [[ $opts =~ $ENABLE_BINARY ]] +then + target=package + if [ "$1"x != ""x ]; then target=$1; fi + if [ "$cmake_version" \< "3.19.0" ] ; then + cmake .. $opts -DENABLE_CROSS_COMPILE=0 + else + cmake .. --preset=default -DENABLE_CROSS_COMPILE=0 + fi + cmake_run_package $target + cp -r kernel ../ + rm -rf * + if [ "$cmake_version" \< "3.19.0" ] ; then + cmake .. $opts + else + cmake .. --preset=default + fi + + cmake --build . --target $target -j16 + if [ $? -ne 0 ]; then + echo "[ERROR] Kernel compile failed, the run package will not be generated." + rm -rf ./cust*.run && rm -rf ./cust*.run.json && exit 1; + fi + if [ $target = "package" ]; then + if test -d ./op_kernel/binary ; then + ./cust*.run + fi + fi + rm -rf ../kernel + +else + target=package + if [ "$1"x != ""x ]; then target=$1; fi + if [ "$cmake_version" \< "3.19.0" ] ; then + cmake .. $opts + else + cmake .. --preset=default + fi + cmake_run_package $target +fi + + +# for debug +# cd build_out +# make +# cpack +# verbose append -v \ No newline at end of file diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/config.cmake b/atb_operator/AtbEyeSample/EyeOp/cmake/config.cmake new file mode 100755 index 000000000..886119daa --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/config.cmake @@ -0,0 +1,25 @@ + +set(CMAKE_CXX_FLAGS_DEBUG "") +set(CMAKE_CXX_FLAGS_RELEASE "") + +if (NOT DEFINED vendor_name) + set(vendor_name customize CACHE STRING "") +endif() +if (NOT DEFINED ASCEND_CANN_PACKAGE_PATH) + set(ASCEND_CANN_PACKAGE_PATH /usr/local/Ascend/latest CACHE PATH "") +endif() +if (NOT DEFINED ASCEND_PYTHON_EXECUTABLE) + set(ASCEND_PYTHON_EXECUTABLE python3 CACHE STRING "") +endif() +if (NOT DEFINED ASCEND_COMPUTE_UNIT) + message(FATAL_ERROR "ASCEND_COMPUTE_UNIT not set in CMakePreset.json ! +") +endif() +set(ASCEND_TENSOR_COMPILER_PATH ${ASCEND_CANN_PACKAGE_PATH}/compiler) +set(ASCEND_CCEC_COMPILER_PATH ${ASCEND_TENSOR_COMPILER_PATH}/ccec_compiler/bin) +set(ASCEND_AUTOGEN_PATH ${CMAKE_BINARY_DIR}/autogen) +set(ASCEND_FRAMEWORK_TYPE tensorflow) +file(MAKE_DIRECTORY ${ASCEND_AUTOGEN_PATH}) +set(CUSTOM_COMPILE_OPTIONS "custom_compile_options.ini") +execute_process(COMMAND rm -rf ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} + COMMAND touch ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS}) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/func.cmake b/atb_operator/AtbEyeSample/EyeOp/cmake/func.cmake new file mode 100755 index 000000000..4179dfd25 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/func.cmake @@ -0,0 +1,192 @@ + +function(get_system_info SYSTEM_INFO) + if (UNIX) + execute_process(COMMAND grep -i ^id= /etc/os-release OUTPUT_VARIABLE TEMP) + string(REGEX REPLACE "\n|id=|ID=|\"" "" SYSTEM_NAME ${TEMP}) + set(${SYSTEM_INFO} ${SYSTEM_NAME}_${CMAKE_SYSTEM_PROCESSOR} PARENT_SCOPE) + elseif (WIN32) + message(STATUS "System is Windows. Only for pre-build.") + else () + message(FATAL_ERROR "${CMAKE_SYSTEM_NAME} not support.") + endif () +endfunction() + +function(opbuild) + message(STATUS "Opbuild generating sources") + cmake_parse_arguments(OPBUILD "" "OUT_DIR;PROJECT_NAME;ACCESS_PREFIX" "OPS_SRC" ${ARGN}) + execute_process(COMMAND ${CMAKE_COMPILE} -g -fPIC -shared -std=c++11 ${OPBUILD_OPS_SRC} -D_GLIBCXX_USE_CXX11_ABI=0 + -I ${ASCEND_CANN_PACKAGE_PATH}/include -L ${ASCEND_CANN_PACKAGE_PATH}/lib64 -lexe_graph -lregister -ltiling_api + -o ${OPBUILD_OUT_DIR}/libascend_all_ops.so + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("build ops lib info: ${EXEC_INFO}") + message("build ops lib error: ${EXEC_ERROR}") + message(FATAL_ERROR "opbuild run failed!") + endif() + set(proj_env "") + set(prefix_env "") + if (NOT "${OPBUILD_PROJECT_NAME}x" STREQUAL "x") + set(proj_env "OPS_PROJECT_NAME=${OPBUILD_PROJECT_NAME}") + endif() + if (NOT "${OPBUILD_ACCESS_PREFIX}x" STREQUAL "x") + set(prefix_env "OPS_DIRECT_ACCESS_PREFIX=${OPBUILD_ACCESS_PREFIX}") + endif() + execute_process(COMMAND ${proj_env} ${prefix_env} ${ASCEND_CANN_PACKAGE_PATH}/toolkit/tools/opbuild/op_build + ${OPBUILD_OUT_DIR}/libascend_all_ops.so ${OPBUILD_OUT_DIR} + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("opbuild ops info: ${EXEC_INFO}") + message("opbuild ops error: ${EXEC_ERROR}") + endif() + message(STATUS "Opbuild generating sources - done") +endfunction() + +function(add_ops_info_target) + cmake_parse_arguments(OPINFO "" "TARGET;OPS_INFO;OUTPUT;INSTALL_DIR" "" ${ARGN}) + get_filename_component(opinfo_file_path "${OPINFO_OUTPUT}" DIRECTORY) + add_custom_command(OUTPUT ${OPINFO_OUTPUT} + COMMAND mkdir -p ${opinfo_file_path} + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/parse_ini_to_json.py + ${OPINFO_OPS_INFO} ${OPINFO_OUTPUT} + ) + add_custom_target(${OPINFO_TARGET} ALL + DEPENDS ${OPINFO_OUTPUT} + ) + install(FILES ${OPINFO_OUTPUT} + DESTINATION ${OPINFO_INSTALL_DIR} + ) +endfunction() + +function(add_ops_compile_options OP_TYPE) + cmake_parse_arguments(OP_COMPILE "" "OP_TYPE" "COMPUTE_UNIT;OPTIONS" ${ARGN}) + file(APPEND ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} + "${OP_TYPE},${OP_COMPILE_COMPUTE_UNIT},${OP_COMPILE_OPTIONS}\n") +endfunction() + +function(add_ops_impl_target) + cmake_parse_arguments(OPIMPL "" "TARGET;OPS_INFO;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) + add_custom_command(OUTPUT ${OPIMPL_OUT_DIR}/.impl_timestamp + COMMAND mkdir -m 700 -p ${OPIMPL_OUT_DIR}/dynamic + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py + ${OPIMPL_OPS_INFO} + \"${OPIMPL_OPS_BATCH}\" \"${OPIMPL_OPS_ITERATE}\" + ${OPIMPL_IMPL_DIR} + ${OPIMPL_OUT_DIR}/dynamic + ${ASCEND_AUTOGEN_PATH} + + COMMAND rm -rf ${OPIMPL_OUT_DIR}/.impl_timestamp + COMMAND touch ${OPIMPL_OUT_DIR}/.impl_timestamp + DEPENDS ${OPIMPL_OPS_INFO} + ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py + ) + add_custom_target(${OPIMPL_TARGET} ALL + DEPENDS ${OPIMPL_OUT_DIR}/.impl_timestamp) + if (${ENABLE_SOURCE_PACKAGE}) + install(DIRECTORY ${OPIMPL_OUT_DIR}/dynamic + DESTINATION ${OPIMPL_INSTALL_DIR} + ) + endif() +endfunction() + +function(add_npu_support_target) + cmake_parse_arguments(NPUSUP "" "TARGET;OPS_INFO_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) + get_filename_component(npu_sup_file_path "${NPUSUP_OUT_DIR}" DIRECTORY) + add_custom_command(OUTPUT ${NPUSUP_OUT_DIR}/npu_supported_ops.json + COMMAND mkdir -p ${NPUSUP_OUT_DIR} + COMMAND ${CMAKE_SOURCE_DIR}/cmake/util/gen_ops_filter.sh + ${NPUSUP_OPS_INFO_DIR} + ${NPUSUP_OUT_DIR} + ) + add_custom_target(npu_supported_ops ALL + DEPENDS ${NPUSUP_OUT_DIR}/npu_supported_ops.json + ) + install(FILES ${NPUSUP_OUT_DIR}/npu_supported_ops.json + DESTINATION ${NPUSUP_INSTALL_DIR} + ) +endfunction() + +function(add_bin_compile_target) + cmake_parse_arguments(BINCMP "" "TARGET;OPS_INFO;COMPUTE_UNIT;IMPL_DIR;ADP_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/src) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/gen) + execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_bin_param_build.py + ${BINCMP_OPS_INFO} ${BINCMP_OUT_DIR}/gen ${BINCMP_COMPUTE_UNIT} + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("ops binary compile scripts gen info: ${EXEC_INFO}") + message("ops binary compile scripts gen error: ${EXEC_ERROR}") + message(FATAL_ERROR "ops binary compile scripts gen failed!") + endif() + if (NOT TARGET binary) + add_custom_target(binary) + endif() + add_custom_target(${BINCMP_TARGET} + COMMAND cp -r ${BINCMP_IMPL_DIR}/*.* ${BINCMP_OUT_DIR}/src + ) + add_custom_target(${BINCMP_TARGET}_gen_ops_config + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/insert_simplified_keys.py -p ${BINCMP_OUT_DIR}/bin + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_ops_config.py -p ${BINCMP_OUT_DIR}/bin + -s ${BINCMP_COMPUTE_UNIT} + ) + add_dependencies(binary ${BINCMP_TARGET}_gen_ops_config) + file(GLOB bin_scripts ${BINCMP_OUT_DIR}/gen/*.sh) + foreach(bin_script ${bin_scripts}) + get_filename_component(bin_file ${bin_script} NAME_WE) + string(REPLACE "-" ";" bin_sep ${bin_file}) + list(GET bin_sep 0 op_type) + list(GET bin_sep 1 op_file) + list(GET bin_sep 2 op_index) + if (NOT TARGET ${BINCMP_TARGET}_${op_file}_copy) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file}) + add_custom_target(${BINCMP_TARGET}_${op_file}_copy + COMMAND cp ${BINCMP_ADP_DIR}/${op_file}.py ${BINCMP_OUT_DIR}/src/${op_type}.py + ) + install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} + DESTINATION ${BINCMP_INSTALL_DIR}/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) + install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json + DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT}/ OPTIONAL + ) + endif() + add_custom_target(${BINCMP_TARGET}_${op_file}_${op_index} + COMMAND export HI_PYTHON=${ASCEND_PYTHON_EXECUTABLE} && bash ${bin_script} ${BINCMP_OUT_DIR}/src/${op_type}.py ${BINCMP_OUT_DIR}/bin/${op_file} && echo $(MAKE) + WORKING_DIRECTORY ${BINCMP_OUT_DIR} + ) + add_dependencies(${BINCMP_TARGET}_${op_file}_${op_index} ${BINCMP_TARGET} ${BINCMP_TARGET}_${op_file}_copy) + add_dependencies(${BINCMP_TARGET}_gen_ops_config ${BINCMP_TARGET}_${op_file}_${op_index}) + endforeach() + install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json + DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) + + install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} + DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/../build_out/kernel/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) + install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json + DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/../build_out/kernel/config/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) + install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json + DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/../build_out/kernel/config/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) + +endfunction() + +function(add_cross_compile_target) + cmake_parse_arguments(CROSSMP "" "TARGET;OUT_DIR;INSTALL_DIR" "" ${ARGN}) + add_custom_target(${CROSSMP_TARGET} ALL + DEPENDS ${CROSSMP_OUT_DIR} + ) + install(DIRECTORY ${CROSSMP_OUT_DIR} + DESTINATION ${CROSSMP_INSTALL_DIR} + ) +endfunction() diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/intf.cmake b/atb_operator/AtbEyeSample/EyeOp/cmake/intf.cmake new file mode 100755 index 000000000..2f362c396 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/intf.cmake @@ -0,0 +1,26 @@ + +add_library(intf_pub INTERFACE) +target_compile_options(intf_pub INTERFACE + -fPIC + -fvisibility=hidden + -fvisibility-inlines-hidden + $<$:-O2> + $<$:-O0 -g> + $<$:-std=c++11> + $<$,$>:-ftrapv -fstack-check> + $<$:-pthread -Wfloat-equal -Wshadow -Wformat=2 -Wno-deprecated -Wextra> + $,-fstack-protector-strong,-fstack-protector-all> +) +target_compile_definitions(intf_pub INTERFACE + _GLIBCXX_USE_CXX11_ABI=0 + $<$:_FORTIFY_SOURCE=2> +) +target_include_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/include) +target_link_options(intf_pub INTERFACE + $<$,EXECUTABLE>:-pie> + $<$:-s> + -Wl,-z,relro + -Wl,-z,now + -Wl,-z,noexecstack +) +target_link_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/lib64) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/makeself.cmake b/atb_operator/AtbEyeSample/EyeOp/cmake/makeself.cmake new file mode 100755 index 000000000..48c565bfb --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/makeself.cmake @@ -0,0 +1,17 @@ +execute_process(COMMAND chmod +x ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh) +execute_process(COMMAND ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh + --header ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself-header.sh + --help-header ./help.info + --gzip --complevel 4 --nomd5 --sha256 + ./ ${CPACK_PACKAGE_FILE_NAME} "version:1.0" ./install.sh + WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} + RESULT_VARIABLE EXEC_RESULT + ERROR_VARIABLE EXEC_ERROR +) +if (NOT "${EXEC_RESULT}x" STREQUAL "0x") + message(FATAL_ERROR "CPack Command error: ${EXEC_RESULT}\n${EXEC_ERROR}") +endif() +execute_process(COMMAND cp ${CPACK_EXTERNAL_BUILT_PACKAGES} ${CPACK_PACKAGE_DIRECTORY}/ + COMMAND echo "Copy ${CPACK_EXTERNAL_BUILT_PACKAGES} to ${CPACK_PACKAGE_DIRECTORY}/" + WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} +) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/__init__.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/__init__.py new file mode 100755 index 000000000..c4ddc893a --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/__init__.py @@ -0,0 +1,8 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- + +import sys +import os + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) +sys.path.append(PYF_PATH) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_bin_param_build.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_bin_param_build.py new file mode 100755 index 000000000..63b15c900 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_bin_param_build.py @@ -0,0 +1,127 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import json +import hashlib +import const_var +import opdesc_parser + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + + +class BinParamBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + super().__init__(op_type) + self.soc = '' + self.out_path = '' + + def set_soc_version(self: any, soc: str): + self.soc = soc + + def set_out_path(self: any, out_path: str): + self.out_path = out_path + + def gen_input_json(self: any): + key_map = {} + count = len(self.input_dtype[0].split(',')) + for i in range(0, count): + inputs = [] + outputs = [] + attrs = [] + op_node = {} + for idx in range(0, len(self.input_name)): + idtypes = self.input_dtype[idx].split(',') + ifmts = self.input_fmt[idx].split(',') + itype = self.input_type[idx] + para = {} + para['name'] = self.input_name[idx][:-5] + para['index'] = idx + para['dtype'] = idtypes[i] + para['format'] = ifmts[i] + para['paramType'] = itype + para['shape'] = [-2] + if itype == 'dynamic': + inputs.append([para]) + else: + inputs.append(para) + for idx in range(0, len(self.output_name)): + odtypes = self.output_dtype[idx].split(',') + ofmts = self.output_fmt[idx].split(',') + otype = self.output_type[idx] + para = {} + para['name'] = self.output_name[idx][:-5] + para['index'] = idx + para['dtype'] = odtypes[i] + para['format'] = ofmts[i] + para['paramType'] = otype + para['shape'] = [-2] + if otype == 'dynamic': + outputs.append([para]) + else: + outputs.append(para) + for attr in self.attr_list: + att = {} + att['name'] = attr + atype = self.attr_val.get(attr).get('type').lower() + att['dtype'] = atype + att['value'] = const_var.ATTR_DEF_VAL.get(atype) + attrs.append(att) + op_node['bin_filename'] = '' + op_node['inputs'] = inputs + op_node['outputs'] = outputs + if len(attrs) > 0: + op_node['attrs'] = attrs + param = {} + param['op_type'] = self.op_type + param['op_list'] = [op_node] + objstr = json.dumps(param, indent=' ') + md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() + while key_map.get(md5sum) is not None: + objstr += '1' + md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() + key_map[md5sum] = md5sum + bin_file = self.op_type + '_' + md5sum + op_node['bin_filename'] = bin_file + param_file = os.path.join(self.out_path, bin_file + '_param.json') + param_file = os.path.realpath(param_file) + with os.fdopen(os.open(param_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + json.dump(param, fd, indent=' ') + self._write_buld_cmd(param_file, bin_file, i) + + def _write_buld_cmd(self: any, param_file: str, bin_file: str, index: int): + hard_soc = const_var.SOC_MAP_EXT.get(self.soc) + if not hard_soc: + hard_soc = self.soc.capitalize() + name_com = [self.op_type, self.op_file, str(index)] + compile_file = os.path.join(self.out_path, '-'.join(name_com) + '.sh') + compile_file = os.path.realpath(compile_file) + with os.fdopen(os.open(compile_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + fd.write('#!/bin/bash\n') + fd.write('echo "[{}] Generating {} ..."\n'.format(hard_soc, bin_file)) + cmd = const_var.BIN_CMD.format(fun=self.op_intf, soc=hard_soc, param=param_file, impl='""') + fd.write(cmd) + chk = const_var.CHK_CMD.format(res_file=bin_file + '.json') + fd.write(chk) + chk = const_var.CHK_CMD.format(res_file=bin_file + '.o') + fd.write(chk) + fd.write('echo "[{}] Generating {} Done"\n'.format(hard_soc, bin_file)) + + +def gen_bin_param_file(cfgfile: str, out_dir: str, soc: str): + op_descs = opdesc_parser.get_op_desc(cfgfile, [], [], BinParamBuilder, None) + for op_desc in op_descs: + op_desc.set_soc_version(soc) + op_desc.set_out_path(out_dir) + op_desc.gen_input_json() + + +if __name__ == '__main__': + if len(sys.argv) <= 3: + raise RuntimeError('arguments must greater than 3') + gen_bin_param_file(sys.argv[1], sys.argv[2], sys.argv[3]) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_impl_build.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_impl_build.py new file mode 100755 index 000000000..647801c07 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_impl_build.py @@ -0,0 +1,501 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import re +import opdesc_parser +import const_var + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + +IMPL_HEAD = ''' +import os, sys +import ctypes +import json +import shutil +from tbe.common.platform import get_soc_spec +from tbe.common.utils import para_check +from tbe.tikcpp import compile_op, replay_op, check_op_cap, generalize_op_params, get_code_channel, OpInfo +from tbe.common.buildcfg import get_default_build_config +from impl.util.platform_adapter import tbe_register +from tbe.common.buildcfg import get_current_build_config +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + +DTYPE_MAP = {{"float32": ["DT_FLOAT", "float"], + "float16": ["DT_FLOAT16", "half"], + "int8": ["DT_INT8", "int8_t"], + "int16": ["DT_INT16", "int16_t"], + "int32": ["DT_INT32", "int32_t"], + "int64": ["DT_INT64", "int64_t"], + "uint1": ["DT_UINT1", "uint8_t"], + "uint8": ["DT_UINT8", "uint8_t"], + "uint16": ["DT_UINT16", "uint16_t"], + "uint32": ["DT_UINT32", "uint32_t"], + "uint64": ["DT_UINT64", "uint64_t"], + "bool": ["DT_BOOL", "bool"], + "double": ["DT_DOUBLE", "double"], + "dual": ["DT_DUAL", "unknown"], + "dual_sub_int8": ["DT_DUAL_SUB_INT8", "unknown"], + "dual_sub_uint8": ["DT_DUAL_SUB_UINT8", "unknown"], + "string": ["DT_STRING", "unknown"], + "complex64": ["DT_COMPLEX64", "unknown"], + "complex128": ["DT_COMPLEX128", "unknown"], + "qint8": ["DT_QINT8", "unknown"], + "qint16": ["DT_QINT16", "unknown"], + "qint32": ["DT_QINT32", "unknown"], + "quint8": ["DT_QUINT8", "unknown"], + "quint16": ["DT_QUINT16", "unknown"], + "resource": ["DT_RESOURCE", "unknown"], + "string_ref": ["DT_STRING_REF", "unknown"], + "int4": ["DT_INT4", "int4b_t"], + "bfloat16": ["DT_BF16", "bfloat16_t"]}} + +def add_dtype_fmt_option_single(x, x_n, is_ref: bool = False): + options = [] + x_fmt = x.get("format") + x_dtype = x.get("dtype") + x_n_in_kernel = x_n + '_REF' if is_ref else x_n + options.append("-DDTYPE_{{n}}={{t}}".format(n=x_n_in_kernel, t=DTYPE_MAP.get(x_dtype)[1])) + options.append("-DORIG_DTYPE_{{n}}={{ot}}".format(n=x_n_in_kernel, ot=DTYPE_MAP.get(x_dtype)[0])) + options.append("-DFORMAT_{{n}}=FORMAT_{{f}}".format(n=x_n_in_kernel, f=x_fmt)) + return options + +def get_dtype_fmt_options(__inputs__, __outputs__): + options = [] + input_names = {} + output_names = {} + unique_param_name_set = set() + for idx, x in enumerate(__inputs__): + if x is None: + continue + x_n = input_names[idx].upper() + unique_param_name_set.add(x_n) + options += add_dtype_fmt_option_single(x, x_n) + + for idx, x in enumerate(__outputs__): + if x is None: + continue + x_n = output_names[idx].upper() + if x_n in unique_param_name_set: + options += add_dtype_fmt_option_single(x, x_n, True) + else: + options += add_dtype_fmt_option_single(x, x_n) + return options + +def load_dso(so_path): + try: + ctypes.CDLL(so_path) + except OSError as error : + print(error) + raise RuntimeError("cannot open %s" %(so_path)) + else: + print("load so succ ", so_path) + +def get_shortsoc_compile_option(compile_option_list: list, shortsoc:str): + compile_options = [] + if shortsoc in compile_option_list: + compile_options = compile_option_list[shortsoc] + elif '__ALLSOC__' in compile_option_list: + compile_options = compile_option_list['__ALLSOC__'] + return compile_options + +def get_kernel_source(src_file, dir_snake, dir_ex): + src_ex = os.path.join(PYF_PATH, "..", "ascendc", dir_ex, src_file) + if os.path.exists(src_ex): + return src_ex + src = os.path.join(PYF_PATH, "..", "ascendc", dir_snake, src_file) + if os.path.exists(src): + return src + src = os.path.join(PYF_PATH, src_file) + if os.path.exists(src): + return src + return src_ex + +''' + +IMPL_API = ''' +@tbe_register.register_operator("{}", trans_bool_to_s8=False) +@para_check.check_op_params({}) +def {}({}, kernel_name="{}", impl_mode=""): + if get_current_build_config("enable_op_prebuild"): + return + __inputs__, __outputs__, __attrs__ = _build_args({}) + options = get_dtype_fmt_options(__inputs__, __outputs__) + options += ["-x", "cce"] + bisheng = shutil.which("bisheng") + if bisheng != None: + bisheng_path = os.path.dirname(bisheng) + tikcpp_path = os.path.realpath(os.path.join(bisheng_path, "..", "..", "tikcpp")) + else: + tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") + options.append("-I" + tikcpp_path) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) + options.append("-I" + os.path.join(PYF_PATH, "..", "ascendc", "common")) + if impl_mode == "high_performance": + options.append("-DHIGH_PERFORMANCE=1") + elif impl_mode == "high_precision": + options.append("-DHIGH_PRECISION=1") + if get_default_build_config("enable_deterministic_mode") == 1: + options.append("-DDETEMINISTIC_MODE=1") + + custom_compile_options = {}, + custom_all_compile_options = {}, + soc_version = get_soc_spec("SOC_VERSION") + soc_short = get_soc_spec("SHORT_SOC_VERSION").lower() + custom_compile_options_soc = get_shortsoc_compile_option(custom_compile_options[0], soc_short) + custom_all_compile_options_soc = get_shortsoc_compile_option(custom_all_compile_options[0], soc_short) + options += custom_all_compile_options_soc + options += custom_compile_options_soc + + origin_func_name = "{}" + ascendc_src_dir_ex = "{}" + ascendc_src_dir = "{}" + ascendc_src_file = "{}" + src = get_kernel_source(ascendc_src_file, ascendc_src_dir, ascendc_src_dir_ex) +''' + +REPLAY_OP_API = ''' + print("start replay Acend C Operator {}, kernel name is {}") + tikreplay_codegen_path = tikcpp_path + "/tikreplaylib/lib" + tikreplay_stub_path = tikcpp_path + "/tikreplaylib/lib/" + soc_version + print("start load libtikreplaylib_codegen.so and libtikreplaylib_stub.so") + codegen_so_path = tikreplay_codegen_path + "/libtikreplaylib_codegen.so" + replaystub_so_path = tikreplay_stub_path + "/libtikreplaylib_stub.so" + if PYF_PATH.endswith("dynamic"): + op_replay_path = os.path.join(PYF_PATH, "..", "..", "op_replay") + else: + op_replay_path = os.path.join(PYF_PATH, "..", "op_replay") + replayapi_so_path = os.path.join(op_replay_path, "libreplay_{}_" + soc_short + ".so") + load_dso(codegen_so_path) + load_dso(replaystub_so_path) + load_dso(replayapi_so_path) + op_type = "{}" + entry_obj = os.path.join(op_replay_path, "{}_entry_" + soc_short + ".o") + code_channel = get_code_channel(src, kernel_name, op_type, options) + op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ + attrs = __attrs__, impl_mode = impl_mode) + res, msg = replay_op(op_info, entry_obj, code_channel, src, options) + if not res: + print("call replay op failed for %s and get into call compile op" %(msg)) + compile_op(src, origin_func_name, op_info, options, code_channel, '{}') +''' + +COMPILE_OP_API = ''' + print("start compile Ascend C operator {}. kernel name is " + kernel_name) + op_type = "{}" + code_channel = get_code_channel(src, kernel_name, op_type, options) + op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ + attrs = __attrs__, impl_mode = impl_mode, origin_inputs=[{}], origin_outputs = [{}], param_type_dynamic = {},\\ + mc2_ctx = None, param_type_list = {}) + compile_op(src, origin_func_name, op_info, options, code_channel, '{}') +''' + +SUP_API = ''' +def {}({}, impl_mode=""): + __inputs__, __outputs__, __attrs__ = _build_args({}) + ret_str = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) + ret_dict = json.loads(ret_str) + err_code = ret_dict.get("ret_code") + sup = "Unknown" + reason = "Unknown reason" + if err_code is not None: + if err_code is 0: + sup = "True" + reason = "" + elif err_code is 1: + sup = "False" + reason = ret_dict.get("reason") + else: + sup = "Unknown" + reason = ret_dict.get("reason") + return sup, reason +''' +CAP_API = ''' +def {}({}, impl_mode=""): + __inputs__, __outputs__, __attrs__ = _build_args({}) + result = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) + return result.decode("utf-8") +''' +GLZ_API = ''' +@tbe_register.register_param_generalization("{}") +def {}_generalization({}, generalize_config=None): + __inputs__, __outputs__, __attrs__ = _build_args({}) + ret_str = generalize_op_params("{}", __inputs__, __outputs__, __attrs__, generalize_config) + return [json.loads(ret_str)] +''' + +ATTR_DEFAULT = {'bool': 'False', 'int': '0', 'float': '0.0', 'list_int': '[]', + 'list_float': '[]', 'list_bool': '[]', 'list_list_int': '[[]]', 'str': ''} + + +def _get_snake_str(s, i): + if s[i - 1] != '_': + if not s[i - 1].isupper(): + return "_" + elif s[i - 1].isupper() and (i + 1) < len(s) and s[i + 1].islower(): + return "_" + return "" + return "" + + +def optype_snake(origin_str): + temp_str = origin_str[0].lower() + origin_str[1:] + new_str = re.sub(r'([A-Z])', r'_\1', temp_str).lower() + return new_str + + +def optype_snake_ex(s): + snake_case = "" + for i, c in enumerate(s): + if i == 0: + snake_case += c.lower() + elif c.isupper(): + snake_case += _get_snake_str(s, i) + snake_case += c.lower() + else: + snake_case += c + return snake_case + + +class AdpBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + self.argsdefv = [] + self.op_compile_option:str = '{}' + super().__init__(op_type) + + + def write_adapt(self: any, impl_path, path: str, op_compile_option_all:list = None): + self._build_paradefault() + if impl_path != "": + src_file = os.path.join(impl_path, self.op_file + '.cpp') + if not os.path.exists(src_file): + return + out_path = os.path.abspath(path) + if self.dynamic_shape and not out_path.endswith('dynamic'): + out_path = os.path.join(path, 'dynamic') + os.makedirs(out_path, mode=0o700, exist_ok=True) + adpfile = os.path.join(out_path, self.op_file + '.py') + self._gen_op_compile_option(op_compile_option_all) + with os.fdopen(os.open(adpfile, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + self._write_head(fd) + self._write_argparse(fd) + self._write_impl(fd) + if self.op_chk_support: + self._write_cap('check_supported', fd) + self._write_cap('get_op_support_info', fd) + if self.op_fmt_sel: + self._write_cap('op_select_format', fd) + self._write_cap('get_op_specific_info', fd) + if self.op_range_limit == 'limited' or self.op_range_limit == 'dynamic': + self._write_glz(fd) + + + def _gen_op_compile_option(self:any, op_compile_option_all:list =None): + if op_compile_option_all is not None: + if self.op_type in op_compile_option_all: + self.op_compile_option = op_compile_option_all[self.op_type] + elif "__all__" in op_compile_option_all: + self.op_compile_option = op_compile_option_all["__all__"] + + + def _ip_argpack(self: any, default: bool = True) -> list: + args = [] + for i in range(len(self.input_name)): + arg = self.input_name[i] + if default and self.argsdefv[i] is not None: + arg += '=' + self.argsdefv[i] + args.append(arg) + return args + + def _op_argpack(self: any, default: bool = True) -> list: + args = [] + argidx = len(self.input_name) + for i in range(len(self.output_name)): + arg = self.output_name[i] + if default and self.argsdefv[i + argidx] is not None: + arg += '=' + self.argsdefv[i + argidx] + args.append(arg) + return args + + def _attr_argpack(self: any, default: bool = True) -> list: + args = [] + argidx = len(self.input_name) + len(self.output_name) + for i in range(len(self.attr_list)): + att = self.attr_list[i] + arg = att + if default and self.argsdefv[i + argidx] is not None: + if self.attr_val.get(att).get('type') == 'str': + arg += '="' + self.argsdefv[i + argidx] + '"' + elif self.attr_val.get(att).get('type') == 'bool': + arg += '=' + self.argsdefv[i + argidx].capitalize() + else: + arg += '=' + self.argsdefv[i + argidx] + args.append(arg) + return args + + def _build_paralist(self: any, default: bool = True) -> str: + args = [] + args.extend(self._ip_argpack(default)) + args.extend(self._op_argpack(default)) + args.extend(self._attr_argpack(default)) + return ', '.join(args) + + def _io_parachk(self: any, types: list, type_name: str) -> list: + chk = [] + for iot in types: + if iot == 'optional': + ptype = 'OPTION' + else: + ptype = iot.upper() + chk.append('para_check.{}_{}'.format(ptype, type_name)) + return chk + + def _attr_parachk(self: any) -> list: + chk = [] + for att in self.attr_list: + if self.attr_val.get(att).get('paramType') == 'optional': + pt = 'OPTION' + else: + pt = self.attr_val.get(att).get('paramType').upper() + att_type = self.attr_val.get(att).get('type').upper() + chk.append('para_check.{}_ATTR_{}'.format(pt, att_type)) + return chk + + def _build_parachk(self: any) -> str: + chk = [] + chk.extend(self._io_parachk(self.input_type, 'INPUT')) + chk.extend(self._io_parachk(self.output_type, 'OUTPUT')) + chk.extend(self._attr_parachk()) + chk.append('para_check.KERNEL_NAME') + return ', '.join(chk) + + def _build_paradefault(self: any): + optional = False + argtypes = [] + argtypes.extend(self.input_type) + argtypes.extend(self.output_type) + for atype in argtypes: + if atype == 'optional': + optional = True + if optional: + self.argsdefv.append('None') + else: + self.argsdefv.append(None) + for attr in self.attr_list: + atype = self.attr_val.get(attr).get('paramType') + if atype == 'optional': + optional = True + attrval = self.attr_val.get(attr).get('defaultValue') + if attrval is not None: + optional = True + if type == "bool": + attrval = attrval.capitalize() + elif type == "str": + attrval = "\"" + attrval + "\"" + self.argsdefv.append(attrval) + continue + if optional: + self.argsdefv.append(ATTR_DEFAULT.get(self.attr_val.get(attr).get('type'))) + else: + self.argsdefv.append(None) + + def _write_head(self: any, fd: object): + fd.write(IMPL_HEAD.format(self.input_ori_name, self.output_ori_name)) + + def _write_argparse(self: any, fd: object): + args = self._build_paralist(False) + fd.write('def _build_args({}):\n'.format(args)) + fd.write(' __inputs__ = []\n') + fd.write(' for arg in [{}]:\n'.format(', '.join(self.input_name))) + fd.write(' if arg != None:\n') + fd.write(' if isinstance(arg, (list, tuple)):\n') + fd.write(' if len(arg) == 0:\n') + fd.write(' continue\n') + fd.write(' __inputs__.append(arg[0])\n') + fd.write(' else:\n') + fd.write(' __inputs__.append(arg)\n') + fd.write(' else:\n') + fd.write(' __inputs__.append(arg)\n') + fd.write(' __outputs__ = []\n') + fd.write(' for arg in [{}]:\n'.format(', '.join(self.output_name))) + fd.write(' if arg != None:\n') + fd.write(' if isinstance(arg, (list, tuple)):\n') + fd.write(' if len(arg) == 0:\n') + fd.write(' continue\n') + fd.write(' __outputs__.append(arg[0])\n') + fd.write(' else:\n') + fd.write(' __outputs__.append(arg)\n') + fd.write(' else:\n') + fd.write(' __outputs__.append(arg)\n') + fd.write(' __attrs__ = []\n') + for attr in self.attr_list: + fd.write(' if {} != None:\n'.format(attr)) + fd.write(' attr = {}\n') + fd.write(' attr["name"] = "{}"\n'.format(attr)) + fd.write(' attr["dtype"] = "{}"\n'.format(self.attr_val.get(attr).get('type'))) + fd.write(' attr["value"] = {}\n'.format(attr)) + fd.write(' __attrs__.append(attr)\n') + fd.write(' return __inputs__, __outputs__, __attrs__\n') + + def _write_impl(self: any, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + pchk = self._build_parachk() + if len(self.kern_name) > 0: + kern_name = self.kern_name + else: + kern_name = self.op_intf + src = self.op_file + '.cpp' + fd.write(IMPL_API.format(self.op_type, pchk, self.op_intf, argsdef, kern_name, argsval,\ + self.custom_compile_options, self.custom_all_compile_options, self.op_intf,\ + optype_snake_ex(self.op_type), optype_snake(self.op_type), src)) + if self.op_replay_flag: + fd.write(REPLAY_OP_API.format(self.op_type, kern_name, self.op_file, self.op_type, self.op_file,\ + self.op_compile_option)) + else: + fd.write(COMPILE_OP_API.format(self.op_type, self.op_type, ', '.join(self.input_name),\ + ', '.join(self.output_name), False, self.input_type + self.output_type, self.op_compile_option)) + + def _write_cap(self: any, cap_name: str, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + if cap_name == 'check_supported': + fd.write(SUP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) + else: + fd.write(CAP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) + + def _write_glz(self: any, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + fd.write(GLZ_API.format(self.op_type, self.op_intf, argsdef, argsval, self.op_type)) + + +def write_scripts(cfgfile: str, cfgs: dict, dirs: dict, ops: list = None, op_compile_option:list = None): + batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') + iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') + file_map = {} + op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, AdpBuilder,\ + ops, dirs.get(const_var.AUTO_GEN_DIR)) + for op_desc in op_descs: + op_desc.write_adapt(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), op_compile_option) + file_map[op_desc.op_type] = op_desc.op_file + return file_map + +if __name__ == '__main__': + if len(sys.argv) <= 6: + raise RuntimeError('arguments must greater equal than 6') + rep_cfg = {} + rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] + rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] + cfg_dir = {} + cfg_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] + cfg_dir[const_var.CFG_OUT_DIR] = sys.argv[5] + cfg_dir[const_var.AUTO_GEN_DIR] = sys.argv[6] + write_scripts(cfgfile=sys.argv[1], cfgs=rep_cfg, dirs=cfg_dir) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_ops_config.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_ops_config.py new file mode 100755 index 000000000..b138047af --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_ops_config.py @@ -0,0 +1,120 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import glob +import json +import argparse +import const_var + + +def load_json(json_file: str): + with open(json_file, encoding='utf-8') as file: + json_content = json.load(file) + return json_content + + +def get_specified_suffix_file(root_dir, suffix): + specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) + all_suffix_files = glob.glob(specified_suffix, recursive=True) + return all_suffix_files + + +def add_simplified_config(op_type, key, core_type, task_ration, objfile, config): + simple_cfg = config.get('binary_info_config.json') + op_cfg = simple_cfg.get(op_type) + if not op_cfg: + op_cfg = {} + op_cfg['dynamicRankSupport'] = True + op_cfg['simplifiedKeyMode'] = 0 + op_cfg['binaryList'] = [] + simple_cfg[op_type] = op_cfg + bin_list = op_cfg.get('binaryList') + if core_type == 0 and task_ration == "tilingKey": + bin_list.append({'coreType': core_type, 'simplifiedKey': key, \ + 'multiKernelType':1, 'binPath': objfile}) + else: + bin_list.append({'coreType': core_type, 'simplifiedKey': key, 'binPath': objfile}) + + +def add_op_config(op_file, bin_info, config): + op_cfg = config.get(op_file) + if not op_cfg: + op_cfg = {} + op_cfg['binList'] = [] + config[op_file] = op_cfg + op_cfg.get('binList').append(bin_info) + + +def gen_ops_config(json_file, soc, config): + core_type_map = {'MIX': 0, 'AiCore': 1, 'VectorCore': 2, 'MIX_AICORE':3, 'MIX_VECTOR_CORE':4} + contents = load_json(json_file) + if ('binFileName' not in contents) or ('supportInfo' not in contents): + return + json_base_name = os.path.basename(json_file) + op_dir = os.path.basename(os.path.dirname(json_file)) + support_info = contents.get('supportInfo') + bin_name = contents.get('binFileName') + bin_suffix = contents.get('binFileSuffix') + task_ration = contents.get('taskRation') + core_type = core_type_map.get(contents.get("coreType")) + bin_file_name = bin_name + bin_suffix + op_type = bin_name.split('_')[0] + op_file = op_dir + '.json' + bin_info = {} + keys = support_info.get('simplifiedKey') + if keys: + bin_info['simplifiedKey'] = keys + for key in keys: + add_simplified_config(op_type, key, core_type, task_ration, os.path.join(soc, op_dir, bin_file_name), \ + config) + bin_info['staticKey'] = support_info.get('staticKey') + bin_info['int64Mode'] = support_info.get('int64Mode') + bin_info['inputs'] = support_info.get('inputs') + bin_info['outputs'] = support_info.get('outputs') + if support_info.get('attrs'): + bin_info['attrs'] = support_info.get('attrs') + bin_info['binInfo'] = {'jsonFilePath': os.path.join(soc, op_dir, json_base_name)} + add_op_config(op_file, bin_info, config) + + +def gen_all_config(root_dir, soc): + suffix = 'json' + config = {} + config['binary_info_config.json'] = {} + all_json_files = get_specified_suffix_file(root_dir, suffix) + for _json in all_json_files: + gen_ops_config(_json, soc, config) + for cfg_key in config.keys(): + cfg_file = os.path.join(root_dir, cfg_key) + with os.fdopen(os.open(cfg_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + json.dump(config.get(cfg_key), fd, indent=' ') + + +def args_prase(): + parser = argparse.ArgumentParser() + parser.add_argument('-p', + '--path', + nargs='?', + required=True, + help='Parse the path of the json file.') + parser.add_argument('-s', + '--soc', + nargs='?', + required=True, + help='Parse the soc_version of ops.') + return parser.parse_args() + + +def main(): + args = args_prase() + gen_all_config(args.path, args.soc) + + +if __name__ == '__main__': + main() diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_replay_build.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_replay_build.py new file mode 100755 index 000000000..1cac7d911 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_replay_build.py @@ -0,0 +1,65 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import opdesc_parser +import replay_codegen +import const_var +from replay_codegen import ReplayCodeGenParams + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + + +class ReplayBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + super().__init__(op_type) + + def gen_replay_source(self: any, impl_path: str, out_path: str, ops_product: str): + if not self.op_replay_flag: + print('{} replay not enabled'.format(self.op_type)) + return + argn = len(self.input_name) + len(self.output_name) + 1 + if self.op_replay_batch: + print('{} replay in batch mode'.format(self.op_type)) + else: + print('{} replay in normal mode'.format(self.op_type)) + if impl_path.endswith('op_kernel'): + implf = os.path.join(impl_path, self.op_file + '.cpp') + tiling_file = os.path.join(impl_path, "../op_host", self.op_file + '_tiling.h') + else: + if self.dynamic_shape: + dyn_path = 'dynamic' + else: + dyn_path = '' + implf = os.path.join(impl_path, dyn_path, self.op_file + '.cpp') + tiling_file = os.path.join(impl_path, "../../op_tiling", self.op_file + '_tiling.h') + rep_conf = replay_codegen.ReplayCodeGen(ReplayCodeGenParams(self.op_type, implf, tiling_file, self.op_file, \ + self.op_intf, argn, self.op_replay_batch, self.max_block_dim, self.max_shape_size)) + rep_conf.set_batch(self.op_replay_batch) + rep_conf.set_outdir(out_path) + rep_conf.gen_replay(ops_product) + + +def gen_replay(cfgfile: str, cfgs: dict, dirs: dict, ops_product: str, ops: list = None): + batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') + iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') + op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, ReplayBuilder, ops) + for op_desc in op_descs: + op_desc.gen_replay_source(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), ops_product) + + +if __name__ == '__main__': + if len(sys.argv) <= 6: + raise RuntimeError('arguments must greater than 6') + rep_cfg = {} + rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] + rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] + rep_dir = {} + rep_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] + rep_dir[const_var.CFG_OUT_DIR] = sys.argv[5] + gen_replay(sys.argv[1], rep_cfg, rep_dir, sys.argv[6]) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/batch_replay_impl.temp b/atb_operator/AtbEyeSample/EyeOp/cmake/util/batch_replay_impl.temp new file mode 100755 index 000000000..0e8834664 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/batch_replay_impl.temp @@ -0,0 +1,117 @@ +#include +#include +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#include "register/op_check.h" +#define __ASCENDC_REPLAY_CODE__ +#include + +using namespace std; +using namespace optiling; +using namespace AscendCReplay; + +extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); +extern "C" int elf_batch_append(char *elf, uint32_t elfSize, char *jit, int kernum, char *atext[], int alen[], + int atlen, const char* kernelname[]); + +#define KERNEL_N 1 +#define ARG_N (__ARG_NUM__) +#define MAX_L (1024 * 1024 * 100) +#define MAX_E (1024 * 1024) + +int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) +{ + // gen type 1 : direct call codes 0: load .o file + if (param.gentype < 0 || param.gentype > 1) { + printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); + return 0; + } else if (param.gentype == 1 && param.objptr == nullptr) { + printf("Error: call replay with direct call mode, but code obj addr is null\n"); + return 0; + } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { + printf("Error: call replay with object file mode, but object file path is null\n"); + return 0; + } + // core_type 0:MIX 1:CUBE 2:VEC + if (core_type < 0 || core_type > 2) { + printf("Error: call replay core type is %d !\n", core_type); + return 0; + } + g_coreType = __CORE_TYPE__; + g_taskRation = param.task_ration; + g_tilingKey = param.tiling_key; + + unsigned char *buf, *jit; + char *kernel[KERNEL_N]; + int len[KERNEL_N]; + block_idx = 0; + block_num = param.block_dim; + g_ubBase = block_num; + uint8_t *code = (uint8_t *)malloc(MAX_L); + uint8_t *pos = code; + struct timespec tp1, tp2; + + clock_gettime(CLOCK_MONOTONIC, &tp1); + if (block_num > 32) { + printf("Error: block_num > 32\n"); + return 0; + } + //__OP_FOPEN__ + for (int i = 0; i < KERNEL_N; i++) { + //__OP_SET_KERNEL__ + for (int j = 0; j < ARG_N; j++) + AddArg(j, ARG_STEP * (j + 1)); +#ifdef FP_CEILING + SetCtrlFloatEnable(); +#else + SetCtrlFloatDisable(); +#endif + CodeInit(pos, true); + __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); + CodeEnd(); + kernel[i] = (char *)pos; + len[i] = CodeLen(); + pos += len[i]; + } + //__OP_FCLOSE__ + clock_gettime(CLOCK_MONOTONIC, &tp2); + buf = (unsigned char *)malloc(MAX_E); + int fd = open(param.entry_file, O_RDONLY); + if (fd < 0) { + printf("[error]: cannot find entry.o : %s\n", param.entry_file); + return 0; + } + uint32_t bufSize = read(fd, buf, MAX_E); + if (bufSize <= 0) { + printf("[error]: entry.o : %s is too small ! \n", param.entry_file); + } + close(fd); + jit = (unsigned char *)malloc(MAX_L); + printf("total code generated %ld\n", pos - code); + int sz = elf_batch_append((char *)buf, bufSize, (char *)jit, KERNEL_N, kernel, len, pos - code, ¶m.kernel_name); + if (tp1.tv_sec != tp2.tv_sec) { + printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); + } else { + printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); + } + printf("new elf size %d\n", sz); + if (param.gentype == 0) { + fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); + (void)write(fd, jit, sz); + close(fd); + free(jit); + } else if (param.gentype == 1) { + *param.objptr = (char*)jit; + } + free(buf); + free(code); + return sz; +} + +REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/code_channel_infer.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/code_channel_infer.py new file mode 100755 index 000000000..d4e6a540c --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/code_channel_infer.py @@ -0,0 +1,58 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" +import os +import stat +import ctypes +import collections +import shutil +import subprocess +import copy + +"""CODE_* is used to cube/vector api is called in operator code +CODE_MIX means both cube and vector api is called +CODE_CUBE means only cube api is called +CODE_VEC means only vector api is called +""" +CODE_MIX = 0 +CODE_CUBE = 1 +CODE_VEC = 2 + + +def _is_v220(op_product: str): + """return if current soc version is V220 + + Returns: + res: True means V220 + """ + if op_product in ["ascend910b", "ascend910c"]: + return True + return False + + +InfoCodeChanelParams = collections.namedtuple('InfoCodeChanelParams',\ +['src_file', 'tiling_header', 'kernel_name', 'outdir', 'op_product', 'compile_options']) + + +def infer_code_channel(params: InfoCodeChanelParams): + """get code channel for v220, return CODE_MIX if soc version is not V220 + + Args: + src_file (str): AscendC operator code file + src_file (str): AscendC operator tiling header file + kernel_name (str): kernel function name + optype (str): operator type + compile_options (list): compile options for bisheng cmd + + Raises: + Exception: if not exist L1/L0/UB if code, it's not a aicore code + + Returns: + res (int): CODE_MIX/CODE_CUBE/CODE_VEC + """ + if not _is_v220(params.op_product): + return CODE_MIX + return CODE_VEC diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/const_var.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/const_var.py new file mode 100755 index 000000000..3e71f5ab8 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/const_var.py @@ -0,0 +1,33 @@ + +#!/usr/bin/env python +# coding=utf-8 +""" +Function: +The replay funtion entry +Copyright Information: +Huawei Technologies Co., Ltd. All Rights Reserved © 2020 +""" + +import os +import stat + + +REPLAY_BATCH = 'batch' +REPLAY_ITERATE = 'iterate' +CFG_IMPL_DIR = 'impl_dir' +CFG_OUT_DIR = 'out_dir' +AUTO_GEN_DIR = 'auto_gen_dir' +WFLAGS = os.O_WRONLY | os.O_CREAT | os.O_TRUNC +WMODES = stat.S_IWUSR | stat.S_IRUSR +SOC_MAP_EXT = {'ascend310p': 'Ascend310P3', 'ascend310b': 'Ascend310B1', + 'ascend910': 'Ascend910A', 'ascend910b': 'Ascend910B1', 'ascend910c': 'Ascend910C1'} +BIN_CMD = 'opc $1 --main_func={fun} --input_param={param} --soc_version={soc} \ +--output=$2 --impl_mode={impl} --simplified_key_mode=0 --op_mode=dynamic\n' +CHK_CMD = ''' +if ! test -f $2/{res_file} ; then + echo "$2/{res_file} not generated!" + exit 1 +fi +''' +ATTR_DEF_VAL = {'str' : '', 'int': 0, 'float': 0.0, 'bool': False, 'list_bool': [], + 'list_int': [], 'list_float': [], 'list_list_int': [[]]} diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_impl_and_mrege_json.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_impl_and_mrege_json.sh new file mode 100755 index 000000000..492034e49 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_impl_and_mrege_json.sh @@ -0,0 +1,20 @@ +#!/usr/bin/bash + +project_path=$1 +build_path=$2 +vendor_name=customize +if [[ ! -d "$project_path" ]]; then + echo "[ERROR] No projcet path is provided" + exit 1 +fi + +if [[ ! -d "$build_path" ]]; then + echo "[ERROR] No build path is provided" + exit 1 +fi + +# copy aicpu kernel so operators +if [[ -d "${project_path}/cpukernel/aicpu_kernel_lib" ]]; then + cp -f ${project_path}/cpukernel/aicpu_kernel_lib/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/cpu/aicpu_kernel/impl + rm -rf ${project_path}/cpukernel/aicpu_kernel_lib +fi diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_ops_filter.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_ops_filter.sh new file mode 100755 index 000000000..d4c27d17f --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_ops_filter.sh @@ -0,0 +1,62 @@ +#!/bin/bash +# Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +# Description: Generate npu_supported_ops.json +# ============================================================================== + +if [[ -z "$1" ]]; then + echo -e "[ERROR] No source dir provided" + exit 1 +fi + +if [[ -z "$2" ]]; then + echo -e "[ERROR] No destination dir provided" + exit 1 +fi + +src=$1 +dest_file=$2/npu_supported_ops.json + +if [ -f "$dest_file" ];then + chmod u+w $dest_file +fi + +echo $* + +add_ops() { + name=$1 + isHeavy=$2 + file=$3 + grep -w "\"$name\"" ${file} >/dev/null + if [ $? == 0 ];then + return + fi + echo " \"${name}\": {" >> ${file} + echo " \"isGray\": false," >> ${file} + echo " \"isHeavy\": ${isHeavy}" >> ${file} + echo " }," >> ${file} +} + +echo "{" > ${dest_file} +ini_files=$(find ${src} -name "*.ini") +for file in ${ini_files} ; do + name=$(grep '^\[' ${file} | sed 's/\[//g' | sed 's/]//g' | sed 's/\r//g') + grep 'heavyOp.flag' ${file} >/dev/null + if [ $? == 0 ];then + isHeavy=$(grep 'heavyOp.flag' ${file} | awk -F= '{print $2}') + else + isHeavy="false" + fi + for op in ${name} ; do + add_ops ${op} "false" ${dest_file} + done +done +echo "}" >> ${dest_file} +file_count=$(cat ${dest_file} | wc -l) +line=$(($file_count-1)) +sed -i "${line}{s/,//g}" ${dest_file} + +chmod 640 "${dest_file}" +echo -e "[INFO] Succed generated ${dest_file}" + +exit 0 + diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_version_info.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_version_info.sh new file mode 100755 index 000000000..a06cfc78d --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_version_info.sh @@ -0,0 +1,6 @@ +ascend_install_dir=$1 +gen_file_dir=$2 + +# create version.info +compiler_version=$(grep "Version" -w ${ascend_install_dir}/compiler/version.info | awk -F = '{print $2}') +echo "custom_opp_compiler_version=${compiler_version}" > ${gen_file_dir}/version.info \ No newline at end of file diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_op_info.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_op_info.py new file mode 100755 index 000000000..28ba08757 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_op_info.py @@ -0,0 +1,36 @@ +# -*- coding: utf-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" +import json +import os +import sys +import stat +import const_var + + +if __name__ == '__main__': + if len(sys.argv) != 3: + print(sys.argv) + print('argv error, inert_op_info.py your_op_file lib_op_file') + sys.exit(2) + + with open(sys.argv[1], 'r') as load_f: + insert_operator = json.load(load_f) + + all_operators = {} + if os.path.exists(sys.argv[2]): + if os.path.getsize(sys.argv[2]) != 0: + with open(sys.argv[2], 'r') as load_f: + all_operators = json.load(load_f) + + for k in insert_operator.keys(): + if k in all_operators.keys(): + print('replace op:[', k, '] success') + else: + print('insert op:[', k, '] success') + all_operators[k] = insert_operator[k] + + with os.fdopen(os.open(sys.argv[2], const_var.WFLAGS, const_var.WMODES), 'w') as json_file: + json_file.write(json.dumps(all_operators, indent=4)) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_simplified_keys.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_simplified_keys.py new file mode 100755 index 000000000..7e8dc2c55 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_simplified_keys.py @@ -0,0 +1,242 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import os +import re +import glob +import json +import argparse + + +DATA_TPYE_DICT = { + 'float32': 0, + 'float16': 1, + 'int8': 2, + 'int16': 6, + 'uint16': 7, + 'uint8': 4, + 'int32': 3, + 'int64': 9, + 'uint32': 8, + 'uint64': 10, + 'bool': 12, + 'double': 11, + 'complex64': 16, + 'complex128': 17, + 'qint8': 18, + 'qint16': 19, + 'qint32': 20, + 'quint8': 21, + 'quint16': 22, + 'resource': 23, + 'string': 24, + 'dual': 25, + 'variant': 26, + 'bf16': 27, + 'bfloat16': 27, + 'undefined': 28, + 'int4': 29, + 'uint1': 30, + 'int2': 31 +} + +FORMAT_DICT = { + 'NCHW': 0, + 'NHWC': 1, + 'ND': 2, + 'NC1HWC0': 3, + 'FRACTAL_Z': 4, + 'NC1C0HWPAD': 5, + 'NHWC1C0': 6, + 'FSR_NCHW': 7, + 'FRACTAL_DECONV': 8, + 'C1HWNC0': 9, + 'FRACTAL_DECONV_TRANSPOSE': 10, + 'FRACTAL_DECONV_SP_STRIDE_TRANS': 11, + 'NC1HWC0_C04': 12, + 'FRACTAL_Z_C04': 13, + 'CHWN': 14, + 'FRACTAL_DECONV_SP_STRIDE8_TRANS': 15, + 'HWCN': 16, + 'NC1KHKWHWC0': 17, + 'BN_WEIGHT': 18, + 'FILTER_HWCK': 19, + 'HASHTABLE_LOOKUP_LOOKUPS': 20, + 'HASHTABLE_LOOKUP_KEYS': 21, + 'HASHTABLE_LOOKUP_VALUE': 22, + 'HASHTABLE_LOOKUP_OUTPUT': 23, + 'HASHTABLE_LOOKUP_HITS': 24, + 'C1HWNCoC0': 25, + 'MD': 26, + 'NDHWC': 27, + 'FRACTAL_ZZ': 28, + 'FRACTAL_NZ': 29, + 'NCDHW': 30, + 'DHWCN': 31, + 'NDC1HWC0': 32, + 'FRACTAL_Z_3D': 33, + 'CN': 34, + 'NC': 35, + 'DHWNC': 36, + 'FRACTAL_Z_3D_TRANSPOSE': 37, + 'FRACTAL_ZN_LSTM': 38, + 'FRACTAL_Z_G': 39, + 'RESERVED': 40, + 'ALL': 41, + 'NULL': 42, + 'ND_RNN_BIAS': 43, + 'FRACTAL_ZN_RNN': 44, + 'NYUV': 45, + 'NYUV_A': 46 +} + + +def load_json(json_file: str): + with open(json_file, encoding='utf-8') as file: + json_content = json.load(file) + return json_content + + +def get_specified_suffix_file(root_dir, suffix): + specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) + all_suffix_files = glob.glob(specified_suffix, recursive=True) + return all_suffix_files + + +def get_deterministic_value(support_info): + deterministic_key = 'deterministic' + if deterministic_key not in support_info: + return 0 + deterministic_value = support_info.get(deterministic_key) + if deterministic_value == 'true': + return 1 + else: + return 0 + + +def get_precision_value(support_info): + precision_key = 'implMode' + precision_value = support_info.get(precision_key) + if precision_value == 'high_performance': + _value = 1 + elif precision_value == 'high_precision': + _value = 2 + else: + _value = 0 + return _value + + +def get_overflow_value(support_info): + return 0 + + +def get_parameters(info): + if info: + if 'dtype' in info: + data_type = info['dtype'] + data_type_value = DATA_TPYE_DICT.get(data_type) + else: + data_type_value = 0 + if 'format' in info: + _format = info['format'] + _format_value = FORMAT_DICT.get(_format) + else: + _format_value = 0 + else: + data_type_value = 0 + _format_value = 0 + return str(data_type_value), str(_format_value) + + +def get_dynamic_parameters(info): + # 动态输入时只需获取第一个参数 + return get_parameters(info[0]) + + +def get_all_parameters(support_info, _type): + result_list = list() + info_lists = support_info.get(_type) + if info_lists: + for _info in info_lists: + # 输入为列表时是动态输入 + if isinstance(_info, (list, tuple)): + data_type_value, _format_value = get_dynamic_parameters(_info) + else: + data_type_value, _format_value = get_parameters(_info) + result_list.append("{},{}".format(data_type_value, _format_value)) + return result_list + + +def get_all_input_parameters(support_info): + result = get_all_parameters(support_info, 'inputs') + return '/'.join(result) + + +def insert_content_into_file(input_file, content): + with open(input_file, 'r+') as file: + lines = file.readlines() + for index, line in enumerate(lines): + match_result = re.search(r'"staticKey":', line) + if match_result: + count = len(line) - len(line.lstrip()) + new_content = "{}{}".format(' ' * count, content) + # 插入到前一行,防止插入最后时还需要考虑是否添加逗号 + lines.insert(index, new_content) + break + file.seek(0) + file.write(''.join(lines)) + + +def insert_simplified_keys(json_file): + contents = load_json(json_file) + # 不存在'binFileName'或者'supportInfo'字段时,非需要替换的解析json文件 + if ('binFileName' not in contents) or ('supportInfo' not in contents): + return + support_info = contents.get('supportInfo') + bin_file_name = contents.get('binFileName') + # 'simplifiedKey'字段已经存在时,直接返回,不重复生成 + if 'simplifiedKey' in support_info: + return + op_type = bin_file_name.split('_')[0] + deterministic = str(get_deterministic_value(support_info)) + precision = str(get_precision_value(support_info)) + overflow = str(get_overflow_value(support_info)) + input_parameters = get_all_input_parameters(support_info) + key = '{}/d={},p={},o={}/{}/'.format( + op_type, + deterministic, + precision, + overflow, + input_parameters) + result = '"simplifiedKey": "' + key + '",\n' + insert_content_into_file(json_file, result) + + +def insert_all_simplified_keys(root_dir): + suffix = 'json' + all_json_files = get_specified_suffix_file(root_dir, suffix) + for _json in all_json_files: + insert_simplified_keys(_json) + + +def args_prase(): + parser = argparse.ArgumentParser() + parser.add_argument('-p', + '--path', + nargs='?', + required=True, + help='Parse the path of the json file.') + return parser.parse_args() + + +def main(): + args = args_prase() + insert_all_simplified_keys(args.path) + + +if __name__ == '__main__': + main() diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_entry.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_entry.py new file mode 100755 index 000000000..2b77c970d --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_entry.py @@ -0,0 +1,115 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + + +def gen_fun_def(title, kernel, argn, arg_type, arg_name): + entry = [] + entry.append(title) + entry.append(kernel) + entry.append('(') + args = [] + for i in range(0, argn): + args.append(arg_type + ' ' + arg_name + str(i)) + entry.append(', '.join(args)) + entry.append(')') + return ' '.join(entry) + + +def gen_batch_kernel_body(fname, argn, arg_name): + body = [] + body.append('{') + fun = [] + fun.append(fname) + fun.append('(') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + fun.append(', '.join(args)) + fun.append(');') + body.append(' '.join(fun)) + body.append('}') + return '\n'.join(body) + + +def gen_mc_kernel_body(kn, argn, arg_name, blknum): + body = [] + body.append('{') + body.append(' switch(block_idx) {') + for blk in range(0, blknum): + fun = [] + fun.append('{}_blk{:02d}'.format(kn, blk)) + fun.append('(') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + fun.append(', '.join(args)) + fun.append(')') + body.append(' case {}: {}; break;'.format(blk, ' '.join(fun))) + body.append(' default: break;') + body.append(' }') + body.append('}') + return '\n'.join(body) + + +def gen_proc_body(argn, arg_name): + body = [] + body.append('{') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + body.append('uint64_t __x = (uint64_t)' + ' + (uint64_t)'.join(args) + ';') + body.append('__asm__ ("NOP");') + body.append('__asm__ ("NOP");') + body.append('__asm__ ("NOP");') + body.append('}') + return '\n'.join(body) + + +def batch_code_gen(kn, argn, argt): + codes = [] + kernel_name = kn + proc_name = kernel_name + '_percore' + arg_num = int(argn) + data_type = argt + arg_type = '__gm__ ' + data_type + '* __restrict__' + arg_name = 'arg' + kernel_title = 'extern \"C\" __global__ __aicore__ void' + proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' + codes.append('#ifndef __aicore__') + codes.append('#define __aicore__ [aicore]') + codes.append('#endif') + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') + codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) + codes.append(gen_batch_kernel_body(proc_name, arg_num, arg_name)) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) + codes.append(gen_proc_body(arg_num, arg_name)) + return '\n'.join(codes) + '\n' + + +def mc_code_gen(kn, argn, argt, blknum): + codes = [] + kernel_name = kn + core_num = int(blknum) + arg_num = int(argn) + data_type = argt + arg_type = '__gm__ ' + data_type + '* __restrict__' + arg_name = 'arg' + kernel_title = 'extern \"C\" __global__ __aicore__ void' + proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' + codes.append('#ifndef __aicore__') + codes.append('#define __aicore__ [aicore]') + codes.append('#endif') + for i in range(0, core_num): + proc_name = '{}_blk{:02d}'.format(kernel_name, i) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') + codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) + codes.append(gen_mc_kernel_body(kernel_name, arg_num, arg_name, core_num)) + for i in range(0, core_num): + proc_name = '{}_blk{:02d}'.format(kernel_name, i) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) + codes.append(gen_proc_body(arg_num, arg_name)) + return '\n'.join(codes) + '\n' diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_impl.temp b/atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_impl.temp new file mode 100755 index 000000000..5079a1043 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_impl.temp @@ -0,0 +1,10 @@ +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#define __ASCENDC_REPLAY_CODE__ +#include "__CCE_FILE__" diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/COPYING b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/COPYING new file mode 100755 index 000000000..d159169d1 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/COPYING @@ -0,0 +1,339 @@ + GNU GENERAL PUBLIC LICENSE + Version 2, June 1991 + + Copyright (C) 1989, 1991 Free Software Foundation, Inc., + 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + Everyone is permitted to copy and distribute verbatim copies + of this license document, but changing it is not allowed. + + Preamble + + The licenses for most software are designed to take away your +freedom to share and change it. By contrast, the GNU General Public +License is intended to guarantee your freedom to share and change free +software--to make sure the software is free for all its users. This +General Public License applies to most of the Free Software +Foundation's software and to any other program whose authors commit to +using it. (Some other Free Software Foundation software is covered by +the GNU Lesser General Public License instead.) You can apply it to +your programs, too. + + When we speak of free software, we are referring to freedom, not +price. Our General Public Licenses are designed to make sure that you +have the freedom to distribute copies of free software (and charge for +this service if you wish), that you receive source code or can get it +if you want it, that you can change the software or use pieces of it +in new free programs; and that you know you can do these things. + + To protect your rights, we need to make restrictions that forbid +anyone to deny you these rights or to ask you to surrender the rights. +These restrictions translate to certain responsibilities for you if you +distribute copies of the software, or if you modify it. + + For example, if you distribute copies of such a program, whether +gratis or for a fee, you must give the recipients all the rights that +you have. You must make sure that they, too, receive or can get the +source code. And you must show them these terms so they know their +rights. + + We protect your rights with two steps: (1) copyright the software, and +(2) offer you this license which gives you legal permission to copy, +distribute and/or modify the software. + + Also, for each author's protection and ours, we want to make certain +that everyone understands that there is no warranty for this free +software. If the software is modified by someone else and passed on, we +want its recipients to know that what they have is not the original, so +that any problems introduced by others will not reflect on the original +authors' reputations. + + Finally, any free program is threatened constantly by software +patents. We wish to avoid the danger that redistributors of a free +program will individually obtain patent licenses, in effect making the +program proprietary. To prevent this, we have made it clear that any +patent must be licensed for everyone's free use or not licensed at all. + + The precise terms and conditions for copying, distribution and +modification follow. + + GNU GENERAL PUBLIC LICENSE + TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION + + 0. This License applies to any program or other work which contains +a notice placed by the copyright holder saying it may be distributed +under the terms of this General Public License. The "Program", below, +refers to any such program or work, and a "work based on the Program" +means either the Program or any derivative work under copyright law: +that is to say, a work containing the Program or a portion of it, +either verbatim or with modifications and/or translated into another +language. (Hereinafter, translation is included without limitation in +the term "modification".) Each licensee is addressed as "you". + +Activities other than copying, distribution and modification are not +covered by this License; they are outside its scope. The act of +running the Program is not restricted, and the output from the Program +is covered only if its contents constitute a work based on the +Program (independent of having been made by running the Program). +Whether that is true depends on what the Program does. + + 1. You may copy and distribute verbatim copies of the Program's +source code as you receive it, in any medium, provided that you +conspicuously and appropriately publish on each copy an appropriate +copyright notice and disclaimer of warranty; keep intact all the +notices that refer to this License and to the absence of any warranty; +and give any other recipients of the Program a copy of this License +along with the Program. + +You may charge a fee for the physical act of transferring a copy, and +you may at your option offer warranty protection in exchange for a fee. + + 2. You may modify your copy or copies of the Program or any portion +of it, thus forming a work based on the Program, and copy and +distribute such modifications or work under the terms of Section 1 +above, provided that you also meet all of these conditions: + + a) You must cause the modified files to carry prominent notices + stating that you changed the files and the date of any change. + + b) You must cause any work that you distribute or publish, that in + whole or in part contains or is derived from the Program or any + part thereof, to be licensed as a whole at no charge to all third + parties under the terms of this License. + + c) If the modified program normally reads commands interactively + when run, you must cause it, when started running for such + interactive use in the most ordinary way, to print or display an + announcement including an appropriate copyright notice and a + notice that there is no warranty (or else, saying that you provide + a warranty) and that users may redistribute the program under + these conditions, and telling the user how to view a copy of this + License. (Exception: if the Program itself is interactive but + does not normally print such an announcement, your work based on + the Program is not required to print an announcement.) + +These requirements apply to the modified work as a whole. If +identifiable sections of that work are not derived from the Program, +and can be reasonably considered independent and separate works in +themselves, then this License, and its terms, do not apply to those +sections when you distribute them as separate works. But when you +distribute the same sections as part of a whole which is a work based +on the Program, the distribution of the whole must be on the terms of +this License, whose permissions for other licensees extend to the +entire whole, and thus to each and every part regardless of who wrote it. + +Thus, it is not the intent of this section to claim rights or contest +your rights to work written entirely by you; rather, the intent is to +exercise the right to control the distribution of derivative or +collective works based on the Program. + +In addition, mere aggregation of another work not based on the Program +with the Program (or with a work based on the Program) on a volume of +a storage or distribution medium does not bring the other work under +the scope of this License. + + 3. You may copy and distribute the Program (or a work based on it, +under Section 2) in object code or executable form under the terms of +Sections 1 and 2 above provided that you also do one of the following: + + a) Accompany it with the complete corresponding machine-readable + source code, which must be distributed under the terms of Sections + 1 and 2 above on a medium customarily used for software interchange; or, + + b) Accompany it with a written offer, valid for at least three + years, to give any third party, for a charge no more than your + cost of physically performing source distribution, a complete + machine-readable copy of the corresponding source code, to be + distributed under the terms of Sections 1 and 2 above on a medium + customarily used for software interchange; or, + + c) Accompany it with the information you received as to the offer + to distribute corresponding source code. (This alternative is + allowed only for noncommercial distribution and only if you + received the program in object code or executable form with such + an offer, in accord with Subsection b above.) + +The source code for a work means the preferred form of the work for +making modifications to it. For an executable work, complete source +code means all the source code for all modules it contains, plus any +associated interface definition files, plus the scripts used to +control compilation and installation of the executable. However, as a +special exception, the source code distributed need not include +anything that is normally distributed (in either source or binary +form) with the major components (compiler, kernel, and so on) of the +operating system on which the executable runs, unless that component +itself accompanies the executable. + +If distribution of executable or object code is made by offering +access to copy from a designated place, then offering equivalent +access to copy the source code from the same place counts as +distribution of the source code, even though third parties are not +compelled to copy the source along with the object code. + + 4. You may not copy, modify, sublicense, or distribute the Program +except as expressly provided under this License. Any attempt +otherwise to copy, modify, sublicense or distribute the Program is +void, and will automatically terminate your rights under this License. +However, parties who have received copies, or rights, from you under +this License will not have their licenses terminated so long as such +parties remain in full compliance. + + 5. You are not required to accept this License, since you have not +signed it. However, nothing else grants you permission to modify or +distribute the Program or its derivative works. These actions are +prohibited by law if you do not accept this License. Therefore, by +modifying or distributing the Program (or any work based on the +Program), you indicate your acceptance of this License to do so, and +all its terms and conditions for copying, distributing or modifying +the Program or works based on it. + + 6. Each time you redistribute the Program (or any work based on the +Program), the recipient automatically receives a license from the +original licensor to copy, distribute or modify the Program subject to +these terms and conditions. You may not impose any further +restrictions on the recipients' exercise of the rights granted herein. +You are not responsible for enforcing compliance by third parties to +this License. + + 7. If, as a consequence of a court judgment or allegation of patent +infringement or for any other reason (not limited to patent issues), +conditions are imposed on you (whether by court order, agreement or +otherwise) that contradict the conditions of this License, they do not +excuse you from the conditions of this License. If you cannot +distribute so as to satisfy simultaneously your obligations under this +License and any other pertinent obligations, then as a consequence you +may not distribute the Program at all. For example, if a patent +license would not permit royalty-free redistribution of the Program by +all those who receive copies directly or indirectly through you, then +the only way you could satisfy both it and this License would be to +refrain entirely from distribution of the Program. + +If any portion of this section is held invalid or unenforceable under +any particular circumstance, the balance of the section is intended to +apply and the section as a whole is intended to apply in other +circumstances. + +It is not the purpose of this section to induce you to infringe any +patents or other property right claims or to contest validity of any +such claims; this section has the sole purpose of protecting the +integrity of the free software distribution system, which is +implemented by public license practices. Many people have made +generous contributions to the wide range of software distributed +through that system in reliance on consistent application of that +system; it is up to the author/donor to decide if he or she is willing +to distribute software through any other system and a licensee cannot +impose that choice. + +This section is intended to make thoroughly clear what is believed to +be a consequence of the rest of this License. + + 8. If the distribution and/or use of the Program is restricted in +certain countries either by patents or by copyrighted interfaces, the +original copyright holder who places the Program under this License +may add an explicit geographical distribution limitation excluding +those countries, so that distribution is permitted only in or among +countries not thus excluded. In such case, this License incorporates +the limitation as if written in the body of this License. + + 9. The Free Software Foundation may publish revised and/or new versions +of the General Public License from time to time. Such new versions will +be similar in spirit to the present version, but may differ in detail to +address new problems or concerns. + +Each version is given a distinguishing version number. If the Program +specifies a version number of this License which applies to it and "any +later version", you have the option of following the terms and conditions +either of that version or of any later version published by the Free +Software Foundation. If the Program does not specify a version number of +this License, you may choose any version ever published by the Free Software +Foundation. + + 10. If you wish to incorporate parts of the Program into other free +programs whose distribution conditions are different, write to the author +to ask for permission. For software which is copyrighted by the Free +Software Foundation, write to the Free Software Foundation; we sometimes +make exceptions for this. Our decision will be guided by the two goals +of preserving the free status of all derivatives of our free software and +of promoting the sharing and reuse of software generally. + + NO WARRANTY + + 11. BECAUSE THE PROGRAM IS LICENSED FREE OF CHARGE, THERE IS NO WARRANTY +FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN +OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES +PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED +OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF +MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS +TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE +PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, +REPAIR OR CORRECTION. + + 12. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING +WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MAY MODIFY AND/OR +REDISTRIBUTE THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, +INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING +OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED +TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY +YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER +PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE +POSSIBILITY OF SUCH DAMAGES. + + END OF TERMS AND CONDITIONS + + How to Apply These Terms to Your New Programs + + If you develop a new program, and you want it to be of the greatest +possible use to the public, the best way to achieve this is to make it +free software which everyone can redistribute and change under these terms. + + To do so, attach the following notices to the program. It is safest +to attach them to the start of each source file to most effectively +convey the exclusion of warranty; and each file should have at least +the "copyright" line and a pointer to where the full notice is found. + + + Copyright (C) + + This program is free software; you can redistribute it and/or modify + it under the terms of the GNU General Public License as published by + the Free Software Foundation; either version 2 of the License, or + (at your option) any later version. + + This program is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + GNU General Public License for more details. + + You should have received a copy of the GNU General Public License along + with this program; if not, write to the Free Software Foundation, Inc., + 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. + +Also add information on how to contact you by electronic and paper mail. + +If the program is interactive, make it output a short notice like this +when it starts in an interactive mode: + + Gnomovision version 69, Copyright (C) year name of author + Gnomovision comes with ABSOLUTELY NO WARRANTY; for details type `show w'. + This is free software, and you are welcome to redistribute it + under certain conditions; type `show c' for details. + +The hypothetical commands `show w' and `show c' should show the appropriate +parts of the General Public License. Of course, the commands you use may +be called something other than `show w' and `show c'; they could even be +mouse-clicks or menu items--whatever suits your program. + +You should also get your employer (if you work as a programmer) or your +school, if any, to sign a "copyright disclaimer" for the program, if +necessary. Here is a sample; alter the names: + + Yoyodyne, Inc., hereby disclaims all copyright interest in the program + `Gnomovision' (which makes passes at compilers) written by James Hacker. + + , 1 April 1989 + Ty Coon, President of Vice + +This General Public License does not permit incorporating your program into +proprietary programs. If your program is a subroutine library, you may +consider it more useful to permit linking proprietary applications with the +library. If this is what you want to do, use the GNU Lesser General +Public License instead of this License. diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/README.md b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/README.md new file mode 100755 index 000000000..b41f01682 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/README.md @@ -0,0 +1,246 @@ +[![License: GPL v2](https://img.shields.io/badge/License-GPL%20v2-blue.svg)](https://www.gnu.org/licenses/old-licenses/gpl-2.0.en.html) +![Build Status](https://github.com/megastep/makeself/workflows/CI/badge.svg) + +# makeself - Make self-extractable archives on Unix + +[makeself.sh][1] is a small shell script that generates a self-extractable +compressed tar archive from a directory. The resulting file appears as a shell script +(many of those have a **.run** suffix), and can be launched as is. The archive +will then uncompress itself to a temporary directory and an optional arbitrary +command will be executed (for example an installation script). This is pretty +similar to archives generated with WinZip Self-Extractor in the Windows world. +Makeself archives also include checksums for integrity self-validation (CRC +and/or MD5/SHA256 checksums). + +The makeself.sh script itself is used only to create the archives from a +directory of files. The resultant archive is actually a compressed (using +gzip, bzip2, or compress) TAR archive, with a small shell script stub at the +beginning. This small stub performs all the steps of extracting the files, +running the embedded command, and removing the temporary files when done. +All the user has to do to install the software contained in such an +archive is to "run" the archive, i.e **sh nice-software.run**. I recommend +using the ".run" (which was introduced by some Makeself archives released by +Loki Software) or ".sh" suffix for such archives not to confuse the users, +so that they will know they are actually shell scripts (with quite a lot of binary data +attached to them though!). + +I am trying to keep the code of this script as portable as possible, i.e it is +not relying on any bash-specific features and only calls commands that are +installed on any functioning UNIX-compatible system. This script as well as +the archives it generates should run on any Unix flavor, with any compatible +Bourne shell, provided of course that the compression programs are available. + +As of version 2.1, Makeself has been rewritten and tested on the following +platforms : + + * Linux (all distributions) + * Sun Solaris (8 and above) + * HP-UX (tested on 11.0 and 11i on HPPA RISC) + * SCO OpenUnix and OpenServer + * IBM AIX 5.1L + * macOS (Darwin) + * SGI IRIX 6.5 + * FreeBSD + * UnicOS / Cray + * Cygwin (Windows) + +If you successfully run Makeself and/or archives created with it on another +system, then please [let me know][2]! + +Examples of publicly available archives made using makeself are : + + * Game patches and installers for [Id Software][3] games like Quake 3 for Linux or Return To Castle Wolfenstein ; + * All game patches released by [Loki Software][4] for the Linux version of popular games ; + * The [nVidia drivers][5] for Linux + * The installer for the Linux version of [Google Earth][6] + * The [VirtualBox][7] installers for Linux + * The [Makeself][1] distribution itself ;-) + * and countless others... + +**Important note for Apache users:** By default, most Web servers will think that Makeself archives are regular text files and thus they may show up as text in a Web browser. The correct way to prevent this is to add a MIME type for this file format, like so (in httpd.conf) : + +`AddType application/x-makeself .run` + +**Important note for certain GNU/Linux distributions:** Archives created with Makeself prior to v2.1.2 were using an old syntax for the _head_ and _tail_ Unix commands that is being progressively obsoleted in their GNU forms. Therefore you may have problems uncompressing some of these archives. A workaround for this is to set the environment variable $_POSIX2_VERSION to enable the old syntax, i.e. : + +`export _POSIX2_VERSION=199209` + +## Usage + +The syntax of makeself is the following: + +``` +makeself.sh [args] archive_dir file_name label startup_script [script_args] +``` + + * _args_ are optional options for Makeself. The available ones are : + + * **`--version`** : Prints the version number on stdout, then exits immediately + * **`--gzip`** : Use gzip for compression (the default on platforms on which gzip is commonly available, like Linux) + * **`--bzip2`** : Use bzip2 instead of gzip for better compression. The bzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. + * **`--pbzip2`** : Use pbzip2 instead of gzip for better and faster compression on machines having multiple CPUs. The pbzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. + * **`--xz`** : Use xz instead of gzip for better compression. The xz command must be available in the command path. It is recommended that the archive prefix be set to something like '.xz.run' for the archive, so that potential users know that they'll need xz to extract it. + * **`--lzo`** : Use lzop instead of gzip for better compression. The lzop command must be available in the command path. It is recommended that the archive prefix be set to something like `.lzo.run` for the archive, so that potential users know that they'll need lzop to extract it. + * **`--lz4`** : Use lz4 instead of gzip for better compression. The lz4 command must be available in the command path. It is recommended that the archive prefix be set to something like '.lz4.run' for the archive, so that potential users know that they'll need lz4 to extract it. + * **`--zstd`** : Use zstd instead of gzip for better compression. The zstd command must be available in the command path. It is recommended that the archive prefix be set to something like '.zstd.run' for the archive, so that potential users know that they'll need zstd to extract it. + * **`--pigz`** : Use pigz for compression. + * **`--base64`** : Encode the archive to ASCII in Base64 format instead of compressing (base64 command required). + * **`--gpg-encrypt`** : Encrypt the archive using `gpg -ac -z $COMPRESS_LEVEL`. This will prompt for a password to encrypt with. Assumes that potential users have `gpg` installed. + * **`--ssl-encrypt`** : Encrypt the archive using `openssl aes-256-cbc -a -salt`. This will prompt for a password to encrypt with. Assumes that the potential users have the OpenSSL tools installed. + * **`--compress`** : Use the UNIX `compress` command to compress the data. This should be the default on all platforms that don't have gzip available. + * **`--nocomp`** : Do not use any compression for the archive, which will then be an uncompressed TAR. + * **`--complevel`** : Specify the compression level for gzip, bzip2, pbzip2, zstd, xz, lzo or lz4. (defaults to 9) + * **`--threads`** : Specify the number of threads to be used by compressors that support parallelization. Omit to use compressor's default. Most useful (and required) for opting into xz's threading, usually with `--threads=0` for all available cores. pbzip2 and pigz are parallel by default, and setting this value allows limiting the number of threads they use. + * **`--notemp`** : The generated archive will not extract the files to a temporary directory, but in a new directory created in the current directory. This is better to distribute software packages that may extract and compile by themselves (i.e. launch the compilation through the embedded script). + * **`--current`** : Files will be extracted to the current directory, instead of in a subdirectory. This option implies `--notemp` above. + * **`--follow`** : Follow the symbolic links inside of the archive directory, i.e. store the files that are being pointed to instead of the links themselves. + * **`--append`** _(new in 2.1.x)_: Append data to an existing archive, instead of creating a new one. In this mode, the settings from the original archive are reused (compression type, label, embedded script), and thus don't need to be specified again on the command line. + * **`--header`** : Makeself uses a separate file to store the header stub, called `makeself-header.sh`. By default, it is assumed that it is stored in the same location as makeself.sh. This option can be used to specify its actual location if it is stored someplace else. + * **`--cleanup`** : Specify a script that is run when execution is interrupted or finishes successfully. The script is executed with the same environment and initial `script_args` as `startup_script`. + * **`--copy`** : Upon extraction, the archive will first extract itself to a temporary directory. The main application of this is to allow self-contained installers stored in a Makeself archive on a CD, when the installer program will later need to unmount the CD and allow a new one to be inserted. This prevents "Filesystem busy" errors for installers that span multiple CDs. + * **`--nox11`** : Disable the automatic spawning of a new terminal in X11. + * **`--nowait`** : When executed from a new X11 terminal, disable the user prompt at the end of the script execution. + * **`--nomd5`** and **`--nocrc`** : Disable the creation of a MD5 / CRC checksum for the archive. This speeds up the extraction process if integrity checking is not necessary. + * **`--sha256`** : Adds a SHA256 checksum for the archive. This is in addition to the MD5 / CRC checksums unless `--nomd5` is also used. + * **`--lsm` _file_** : Provide and LSM file to makeself, that will be embedded in the generated archive. LSM files are describing a software package in a way that is easily parseable. The LSM entry can then be later retrieved using the `--lsm` argument to the archive. An example of a LSM file is provided with Makeself. + * **`--tar-format opt`** : Specify the tar archive format (default is ustar); you may use any value accepted by your tar command (such as posix, v7, etc). + * **`--tar-extra opt`** : Append more options to the tar command line. + + For instance, in order to exclude the `.git` directory from the packaged archive directory using the GNU `tar`, one can use `makeself.sh --tar-extra "--exclude=.git" ...` + + * **`--keep-umask`** : Keep the umask set to shell default, rather than overriding when executing self-extracting archive. + * **`--packaging-date date`** : Use provided string as the packaging date instead of the current date. + * **`--license`** : Append a license file. + * **`--nooverwrite`** : Do not extract the archive if the specified target directory already exists. + * **`--help-header file`** : Add a header to the archive's `--help` output. + * `archive_dir` is the name of the directory that contains the files to be archived + * `file_name` is the name of the archive to be created + * `label` is an arbitrary text string describing the package. It will be displayed while extracting the files. + * `startup_script` is the command to be executed _from within_ the directory of extracted files. Thus, if you wish to execute a program contained in this directory, you must prefix your command with `./`. For example, `./program` will be fine. The `script_args` are additional arguments for this command. + +Here is an example, assuming the user has a package image stored in a **/home/joe/mysoft**, and he wants to generate a self-extracting package named +**mysoft.sh**, which will launch the "setup" script initially stored in /home/joe/mysoft : + +`makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup +` + +Here is also how I created the [makeself.run][9] archive which contains the Makeself distribution : + +`makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" ` + +Archives generated with Makeself can be passed the following arguments: + + * **`--keep`** : Prevent the files to be extracted in a temporary directory that will be removed after the embedded script's execution. The files will then be extracted in the current working directory and will stay here until you remove them. + * **`--verbose`** : Will prompt the user before executing the embedded command + * **`--target dir`** : Allows to extract the archive in an arbitrary place. + * **`--nox11`** : Do not spawn a X11 terminal. + * **`--confirm`** : Prompt the user for confirmation before running the embedded command. + * **`--info`** : Print out general information about the archive (does not extract). + * **`--lsm`** : Print out the LSM entry, if it is present. + * **`--list`** : List the files in the archive. + * **`--check`** : Check the archive for integrity using the embedded checksums. Does not extract the archive. + * **`--nochown`** : By default, a `chown -R` command is run on the target directory after extraction, so that all files belong to the current user. This is mostly needed if you are running as root, as tar will then try to recreate the initial user ownerships. You may disable this behavior with this flag. + * **`--tar`** : Run the tar command on the contents of the archive, using the following arguments as parameter for the command. + * **`--noexec`** : Do not run the embedded script after extraction. + * **`--noexec-cleanup`** : Do not run the embedded cleanup script. + * **`--nodiskspace`** : Do not check for available disk space before attempting to extract. + * **`--cleanup-args`** : Specify arguments to be passed to the cleanup script. Wrap value in quotes to specify multiple arguments. + +Any subsequent arguments to the archive will be passed as additional arguments to the embedded command. You must explicitly use the `--` special command-line construct before any such options to make sure that Makeself will not try to interpret them. + +## Startup Script + +The startup script must be a regular Shell script. + +Within the startup script, you can use the `$USER_PWD` variable to get the path of the folder from which the self-extracting script is executed. This is especially useful to access files that are located in the same folder as the script, as shown in the example below. + +`my-self-extracting-script.sh --fooBarFileParameter foo.bar` + +## Building and Testing + +Clone the git repo and execute `git submodule update --init --recursive` to obtain all submodules. + +* To make a release: `make` +* To run all tests: `make test` + +## Maven Usage + +Makeself is now supported by the following maven plugin [makeself-maven-plugin](https://github.com/hazendaz/makeself-maven-plugin). Please refer to project for usage and report any bugs in regards to maven plugin on that project. + +## License + +Makeself itself is covered by the [GNU General Public License][8] (GPL) version 2 and above. Archives generated by Makeself don't have to be placed under this license (although I encourage it ;-)), since the archive itself is merely data for Makeself. + +## Contributing + +I will gladly consider merging your pull requests on the [GitHub][10] repository. However, please keep the following in mind: + + * One of the main purposes of Makeself is portability. Do not submit patches that will break supported platforms. The more platform-agnostic, the better. + * Please explain clearly what the purpose of the patch is, and how you achieved it. + +## Download + +Get the latest official distribution [here][9] (version 2.4.2). + +The latest development version can be grabbed from [GitHub][10]. Feel free to submit any patches there through the fork and pull request process. + +## Version history + + * **v1.0:** Initial public release + * **v1.1:** The archive can be passed parameters that will be passed on to the embedded script, thanks to John C. Quillan + * **v1.2:** Cosmetic updates, support for bzip2 compression and non-temporary archives. Many ideas thanks to Francois Petitjean. + * **v1.3:** More patches from Bjarni R. Einarsson and Francois Petitjean: Support for no compression (`--nocomp`), script is no longer mandatory, automatic launch in an xterm, optional verbose output, and -target archive option to indicate where to extract the files. + * **v1.4:** Many patches from Francois Petitjean: improved UNIX compatibility, automatic integrity checking, support of LSM files to get info on the package at run time.. + * **v1.5.x:** A lot of bugfixes, and many other patches, including automatic verification through the usage of checksums. Version 1.5.5 was the stable release for a long time, even though the Web page didn't get updated ;-). Makeself was also officially made a part of the [Loki Setup installer][11], and its source is being maintained as part of this package. + * **v2.0:** Complete internal rewrite of Makeself. The command-line parsing was vastly improved, the overall maintenance of the package was greatly improved by separating the stub from makeself.sh. Also Makeself was ported and tested to a variety of Unix platforms. + * **v2.0.1:** First public release of the new 2.0 branch. Prior versions are officially obsoleted. This release introduced the `--copy` argument that was introduced in response to a need for the [UT2K3][12] Linux installer. + * **v2.1.0:** Big change : Makeself can now support multiple embedded tarballs, each stored separately with their own checksums. An existing archive can be updated with the `--append` flag. Checksums are also better managed, and the `--nochown` option for archives appeared. + * **v2.1.1:** Fixes related to the Unix compression (compress command). Some Linux distributions made the insane choice to make it unavailable, even though gzip is capable of uncompressing these files, plus some more bugfixes in the extraction and checksum code. + * **v2.1.2:** Some bug fixes. Use head -n to avoid problems with POSIX conformance. + * **v2.1.3:** Bug fixes with the command line when spawning terminals. Added `--tar`, `--noexec` for archives. Added `--nomd5` and `--nocrc` to avoid creating checksums in archives. The embedded script is now run through "eval". The `--info` output now includes the command used to create the archive. A man page was contributed by Bartosz Fenski. + * **v2.1.4:** Fixed `--info` output. Generate random directory name when extracting files to . to avoid problems. Better handling of errors with wrong permissions for the directory containing the files. Avoid some race conditions, Unset the $CDPATH variable to avoid problems if it is set. Better handling of dot files in the archive directory. + * **v2.1.5:** Made the md5sum detection consistent with the header code. Check for the presence of the archive directory. Added `--encrypt` for symmetric encryption through gpg (Eric Windisch). Added support for the digest command on Solaris 10 for MD5 checksums. Check for available disk space before extracting to the target directory (Andreas Schweitzer). Allow extraction to run asynchronously (patch by Peter Hatch). Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo). + * **v2.1.6:** Replaced one dot per file progress with a realtime progress percentage and a spinning cursor. Added `--noprogress` to prevent showing the progress during the decompression. Added `--target` dir to allow extracting directly to a target directory. (Guy Baconniere) + * **v2.2.0:** First major new release in years! Includes many bugfixes and user contributions. Please look at the [project page on Github][10] for all the details. + * **v2.3.0:** Support for archive encryption via GPG or OpenSSL. Added LZO and LZ4 compression support. Options to set the packaging date and stop the umask from being overriden. Optionally ignore check for available disk space when extracting. New option to check for root permissions before extracting. + * **v2.3.1:** Various compatibility updates. Added unit tests for Travis CI in the GitHub repo. New `--tar-extra`, `--untar-extra`, `--gpg-extra`, `--gpg-asymmetric-encrypt-sign` options. + * **v2.4.0:** Added optional support for SHA256 archive integrity checksums. + * **v2.4.2:** New --cleanup and --cleanup-args arguments for cleanup scripts. Added threading support for supported compressors. Now supports zstd compression. + * **v2.4.3:** Make explicit POSIX tar archives for increased compatibility. + * **v2.4.4:** Fixed various compatibility issues (no longer use POSIX tar archives), Github Actions to check on Solaris and FreeBSD. + * **v2.4.5:** Added `--tar-format` option to set the tar archive format (default is ustar) + +## Links + + * Check out the ["Loki Setup"][11] installer, used to install many Linux games and other applications, and of which I am the co-author. Since the demise of Loki, I am now the official maintainer of the project, and it is now being hosted here on GitHub. + * Bjarni R. Einarsson also wrote the **setup.sh** installer script, inspired by Makeself. [Check it out !][14] + +## Contact + +This script was written by [Stéphane Peter][15] (megastep at megastep.org). Any enhancements and suggestions are welcome. + +Contributions were included from John C. Quillan, Bjarni R. Einarsson, +Francois Petitjean, Ryan C. Gordon, and many contributors on GitHub. If you think I forgot +your name, don't hesitate to contact me. + +This project is now hosted on GitHub. Feel free to submit patches and bug reports on the [project page][10]. + +* * * + +[Stephane Peter][2] + + [1]: http://makeself.io/ + [2]: mailto:megastep@megastep.org + [3]: http://www.idsoftware.com/ + [4]: http://www.lokigames.com/products/myth2/updates.php3 + [5]: http://www.nvidia.com/ + [6]: http://earth.google.com/ + [7]: http://www.virtualbox.org/ + [8]: http://www.gnu.org/copyleft/gpl.html + [9]: https://github.com/megastep/makeself/releases/download/release-2.4.5/makeself-2.4.5.run + [10]: https://github.com/megastep/makeself + [11]: https://github.com/megastep/loki_setup/ + [12]: http://www.unrealtournament2003.com/ + [13]: http://www.icculus.org/ + [14]: http://bre.klaki.net/programs/setup.sh/ + [15]: https://stephanepeter.com/ diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/VERSION b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/VERSION new file mode 100755 index 000000000..59aa62c1f --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/VERSION @@ -0,0 +1 @@ +2.4.5 diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/make-release.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/make-release.sh new file mode 100755 index 000000000..b5692d490 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/make-release.sh @@ -0,0 +1,9 @@ +#!/bin/sh +# +# Create a distributable archive of the current version of Makeself + +VER=`cat VERSION` +mkdir -p /tmp/makeself-$VER release +cp -pPR makeself* test README.md COPYING VERSION .gitmodules /tmp/makeself-$VER/ +./makeself.sh --notemp /tmp/makeself-$VER release/makeself-$VER.run "Makeself v$VER" echo "Makeself has extracted itself" + diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself-header.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself-header.sh new file mode 100755 index 000000000..940903148 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself-header.sh @@ -0,0 +1,660 @@ +cat << EOF > "$archname" +#!/bin/bash +# This script was generated using Makeself $MS_VERSION +# The license covering this archive and its contents, if any, is wholly independent of the Makeself license (GPL) +# 2022.3.19-Modified the MS_Help function and some options +# Huawei Technologies Co., Ltd. + +ORIG_UMASK=\`umask\` + +CRCsum="$CRCsum" +MD5="$MD5sum" +SHA="$SHAsum" +SIGNATURE="$Signature" +TMPROOT=\${TMPDIR:="\$HOME"} +if ! test -d "\$TMPROOT"; then + TMPROOT="\$PWD" +fi +export TMPDIR="\$TMPROOT" +USER_PWD="\$PWD" +if ! test -d "\$USER_PWD"; then + exit 1 +fi +export USER_PWD +ARCHIVE_DIR=\`dirname "\$0"\` +export ARCHIVE_DIR + +name_of_file="\$0 " +pwd_of_file="\$PWD" +label="$LABEL" +script="$SCRIPT" +scriptargs="$SCRIPTARGS" +cleanup_script="${CLEANUP_SCRIPT}" +licensetxt="$LICENSE" +helpheader='$HELPHEADER' +targetdir="$archdirname" +filesizes="$filesizes" +totalsize="$totalsize" +keep="$KEEP" +nooverwrite="$NOOVERWRITE" +quiet="n" +accept="n" +nodiskspace="n" +export_conf="$EXPORT_CONF" +decrypt_cmd="$DECRYPT_CMD" +skip="$SKIP" + +print_cmd_arg="" +if type printf > /dev/null; then + print_cmd="printf" +elif test -x /usr/ucb/echo; then + print_cmd="/usr/ucb/echo" +else + print_cmd="echo" +fi + +if test -d /usr/xpg4/bin; then + PATH=/usr/xpg4/bin:\$PATH + export PATH +fi + +if test -d /usr/sfw/bin; then + PATH=\$PATH:/usr/sfw/bin + export PATH +fi + +unset CDPATH + +MS_Printf() +{ + \$print_cmd \$print_cmd_arg "\$1" +} + +MS_PrintLicense() +{ + PAGER=\${PAGER:=more} + if test x"\$licensetxt" != x; then + PAGER_PATH=\`exec <&- 2>&-; which \$PAGER || command -v \$PAGER || type \$PAGER\` + if test -x "\$PAGER_PATH"; then + echo "\$licensetxt" | \$PAGER + else + echo "\$licensetxt" + fi + if test x"\$accept" != xy; then + while true + do + MS_Printf "Please type y to accept, n otherwise: " + read yn + if test x"\$yn" = xn; then + keep=n + eval \$finish; exit 1 + break; + elif test x"\$yn" = xy; then + break; + fi + done + fi + fi +} + +MS_diskspace() +{ + ( + df -kP "\$1" | tail -1 | awk '{ if (\$4 ~ /%/) {print \$3} else {print \$4} }' + ) +} + +MS_dd() +{ + blocks=\`expr \$3 / 1024\` + bytes=\`expr \$3 % 1024\` + # Test for ibs, obs and conv feature + if dd if=/dev/zero of=/dev/null count=1 ibs=512 obs=512 conv=sync 2> /dev/null; then + dd if="\$1" ibs=\$2 skip=1 obs=1024 conv=sync 2> /dev/null | \\ + { test \$blocks -gt 0 && dd ibs=1024 obs=1024 count=\$blocks ; \\ + test \$bytes -gt 0 && dd ibs=1 obs=1024 count=\$bytes ; } 2> /dev/null + else + dd if="\$1" bs=\$2 skip=1 2> /dev/null + fi +} + +MS_dd_Progress() +{ + if test x"\$noprogress" = xy; then + MS_dd "\$@" + return \$? + fi + file="\$1" + offset=\$2 + length=\$3 + pos=0 + bsize=4194304 + while test \$bsize -gt \$length; do + bsize=\`expr \$bsize / 4\` + done + blocks=\`expr \$length / \$bsize\` + bytes=\`expr \$length % \$bsize\` + ( + dd ibs=\$offset skip=1 2>/dev/null + pos=\`expr \$pos \+ \$bsize\` + MS_Printf " 0%% " 1>&2 + if test \$blocks -gt 0; then + while test \$pos -le \$length; do + dd bs=\$bsize count=1 2>/dev/null + pcent=\`expr \$length / 100\` + pcent=\`expr \$pos / \$pcent\` + if test \$pcent -lt 100; then + MS_Printf "\b\b\b\b\b\b\b" 1>&2 + if test \$pcent -lt 10; then + MS_Printf " \$pcent%% " 1>&2 + else + MS_Printf " \$pcent%% " 1>&2 + fi + fi + pos=\`expr \$pos \+ \$bsize\` + done + fi + if test \$bytes -gt 0; then + dd bs=\$bytes count=1 2>/dev/null + fi + MS_Printf "\b\b\b\b\b\b\b" 1>&2 + MS_Printf " 100%% " 1>&2 + ) < "\$file" +} + +MS_Help() +{ + cat << EOH >&2 +Usage: \$0 [options] +Options: + --help | -h Print this message + --info Print embedded info : title, default target directory, embedded script ... + --list Print the list of files in the archive + --check Checks integrity and version dependency of the archive + --quiet Quiet install mode, skip human-computer interactions + --nox11 Do not spawn an xterm + --noexec Do not run embedded script + --extract= Extract directly to a target directory (absolute or relative) + Usually used with --noexec to just extract files without running + --tar arg1 [arg2 ...] Access the contents of the archive through the tar command +\${helpheader} +EOH +} + +MS_Verify_Sig() +{ + GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` + MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` + test -x "\$GPG_PATH" || GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` + test -x "\$MKTEMP_PATH" || MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` + offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` + temp_sig=\`mktemp -t XXXXX\` + echo \$SIGNATURE | base64 --decode > "\$temp_sig" + gpg_output=\`MS_dd "\$1" \$offset \$totalsize | LC_ALL=C "\$GPG_PATH" --verify "\$temp_sig" - 2>&1\` + gpg_res=\$? + rm -f "\$temp_sig" + if test \$gpg_res -eq 0 && test \`echo \$gpg_output | grep -c Good\` -eq 1; then + if test \`echo \$gpg_output | grep -c \$sig_key\` -eq 1; then + test x"\$quiet" = xn && echo "GPG signature is good" >&2 + else + echo "GPG Signature key does not match" >&2 + exit 2 + fi + else + test x"\$quiet" = xn && echo "GPG signature failed to verify" >&2 + exit 2 + fi +} + +MS_Check() +{ + OLD_PATH="\$PATH" + PATH=\${GUESS_MD5_PATH:-"\$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} + MD5_ARG="" + MD5_PATH=\`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum\` + test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which md5 || command -v md5 || type md5\` + test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which digest || command -v digest || type digest\` + PATH="\$OLD_PATH" + + SHA_PATH=\`exec <&- 2>&-; which shasum || command -v shasum || type shasum\` + test -x "\$SHA_PATH" || SHA_PATH=\`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum\` + + if test x"\$quiet" = xn; then + MS_Printf "Verifying archive integrity..." + fi + offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` + fsize=\`cat "\$1" | wc -c | tr -d " "\` + if test \$totalsize -ne \`expr \$fsize - \$offset\`; then + echo " Unexpected archive size." >&2 + exit 2 + fi + verb=\$2 + i=1 + for s in \$filesizes + do + crc=\`echo \$CRCsum | cut -d" " -f\$i\` + if test -x "\$SHA_PATH"; then + if test x"\`basename \$SHA_PATH\`" = xshasum; then + SHA_ARG="-a 256" + fi + sha=\`echo \$SHA | cut -d" " -f\$i\` + if test x"\$sha" = x0000000000000000000000000000000000000000000000000000000000000000; then + test x"\$verb" = xy && echo " \$1 does not contain an embedded SHA256 checksum." >&2 + else + shasum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$SHA_PATH \$SHA_ARG" | cut -b-64\`; + if test x"\$shasum" != x"\$sha"; then + echo "Error in SHA256 checksums: \$shasum is different from \$sha" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " SHA256 checksums are OK." >&2 + fi + crc="0000000000"; + fi + fi + if test -x "\$MD5_PATH"; then + if test x"\`basename \$MD5_PATH\`" = xdigest; then + MD5_ARG="-a md5" + fi + md5=\`echo \$MD5 | cut -d" " -f\$i\` + if test x"\$md5" = x00000000000000000000000000000000; then + test x"\$verb" = xy && echo " \$1 does not contain an embedded MD5 checksum." >&2 + else + md5sum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$MD5_PATH \$MD5_ARG" | cut -b-32\`; + if test x"\$md5sum" != x"\$md5"; then + echo "Error in MD5 checksums: \$md5sum is different from \$md5" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " MD5 checksums are OK." >&2 + fi + crc="0000000000"; verb=n + fi + fi + if test x"\$crc" = x0000000000; then + test x"\$verb" = xy && echo " \$1 does not contain a CRC checksum." >&2 + else + sum1=\`MS_dd_Progress "\$1" \$offset \$s | CMD_ENV=xpg4 cksum | awk '{print \$1}'\` + if test x"\$sum1" != x"\$crc"; then + echo "Error in checksums: \$sum1 is different from \$crc" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " CRC checksums are OK." >&2 + fi + fi + i=\`expr \$i + 1\` + offset=\`expr \$offset + \$s\` + done + if test x"\$quiet" = xn; then + echo " All good." + fi +} + +MS_Decompress() +{ + if test x"\$decrypt_cmd" != x""; then + { eval "\$decrypt_cmd" || echo " ... Decryption failed." >&2; } | eval "$GUNZIP_CMD" + else + eval "$GUNZIP_CMD" + fi + + if test \$? -ne 0; then + echo " ... Decompression failed." >&2 + fi +} + +UnTAR() +{ + if test x"\$quiet" = xn; then + tar \$1vf - $UNTAR_EXTRA 2>&1 || { echo " ... Extraction failed." >&2; kill -15 \$$; } + else + tar \$1f - $UNTAR_EXTRA 2>&1 || { echo Extraction failed. >&2; kill -15 \$$; } + fi +} + +MS_exec_cleanup() { + if test x"\$cleanup" = xy && test x"\$cleanup_script" != x""; then + cleanup=n + cd "\$tmpdir" + eval "\"\$cleanup_script\" \$scriptargs \$cleanupargs" + fi +} + +MS_cleanup() +{ + echo 'Signal caught, cleaning up' >&2 + MS_exec_cleanup + cd "\$TMPROOT" + rm -rf "\$tmpdir" + eval \$finish; exit 15 +} + +Script_Args_Check() +{ + script_supported_args=\$(echo \${helpheader} | grep -o -E "\-\-[^ ]+" | awk -F"=" {'print \$1'}) + arg_to_test=\$(echo \$1|awk -F"=" {'print \$1'}) + + for arg in \${script_supported_args}; + do + if test x"\$arg_to_test" = x"\$arg" ;then + return + fi + done + + MS_Help + exit 1 +} + +finish=true +xterm_loop= +noprogress=$NOPROGRESS +nox11=$NOX11 +copy=$COPY +ownership=$OWNERSHIP +verbose=n +cleanup=y +cleanupargs= +sig_key= + +initargs="\$@" + +while [ -n "\$*" ] +do + case "\$1" in + -h | --help) + MS_Help + exit 0 + ;; + -q | --quiet) + quiet=y + noprogress=y + shift + ;; + --info) + echo Identification: "\$label" + echo Target directory: "\$targetdir" + echo Uncompressed size: $USIZE KB + echo Compression: $COMPRESS + if test x"$ENCRYPT" != x""; then + echo Encryption: $ENCRYPT + fi + echo Date of packaging: $DATE + echo Built with Makeself version $MS_VERSION + echo Build command was: "$MS_COMMAND" + if test x"\$script" != x; then + echo Script run after extraction: + echo " " \$script \$scriptargs + fi + if test x"$copy" = xcopy; then + echo "Archive will copy itself to a temporary location" + fi + if test x"$NEED_ROOT" = xy; then + echo "Root permissions required for extraction" + fi + if test x"$KEEP" = xy; then + echo "directory \$targetdir is permanent" + else + echo "\$targetdir will be removed after extraction" + fi + exit 0 + ;; + --list) + echo Target directory: \$targetdir + offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + for s in \$filesizes + do + MS_dd "\$0" \$offset \$s | MS_Decompress | UnTAR t + offset=\`expr \$offset + \$s\` + done + exit 0 + ;; + --tar) + offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + arg1="\$2" + shift 2 || { MS_Help; exit 1; } + for s in \$filesizes + do + MS_dd "\$0" \$offset \$s | MS_Decompress | tar "\$arg1" - "\$@" + offset=\`expr \$offset + \$s\` + done + exit 0 + ;; + --check) + MS_Check "\$0" y + scriptargs="\$scriptargs \$1" + shift + ;; + --noexec) + script="" + cleanup_script="" + shift + ;; + --extract=*) + keep=y + targetdir=\`echo \$1 | cut -d"=" -f2 \` + if ! shift; then MS_Help; exit 1; fi + ;; + --nox11) + nox11=y + shift + ;; + --xwin) + if test "$NOWAIT" = n; then + finish="echo Press Return to close this window...; read junk" + fi + xterm_loop=1 + shift + ;; + --phase2) + copy=phase2 + shift + ;; + --repack | --repack-path=*) + Script_Args_Check \$1 + scriptargs="\$scriptargs '\$1'" + shift + if [[ ! "\$1" =~ ^-.* ]]; then + scriptargs="\$scriptargs '\$1'" + shift + fi + ;; + *) + Script_Args_Check \$1 + scriptargs="\$scriptargs '\$1'" + shift + ;; + esac +done + +quiet_para="" +if test x"\$quiet" = xy; then + quiet_para="--quiet " +fi +scriptargs="--\$name_of_file""--\"\$pwd_of_file\""" \$quiet_para""\$scriptargs" + +if test x"\$quiet" = xy -a x"\$verbose" = xy; then + echo Cannot be verbose and quiet at the same time. >&2 + exit 1 +fi + +if test x"$NEED_ROOT" = xy -a \`id -u\` -ne 0; then + echo "Administrative privileges required for this archive (use su or sudo)" >&2 + exit 1 +fi + +if test x"\$copy" \!= xphase2; then + MS_PrintLicense +fi + +case "\$copy" in +copy) + tmpdir="\$TMPROOT"/makeself.\$RANDOM.\`date +"%y%m%d%H%M%S"\`.\$\$ + mkdir "\$tmpdir" || { + echo "Could not create temporary directory \$tmpdir" >&2 + exit 1 + } + SCRIPT_COPY="\$tmpdir/makeself" + echo "Copying to a temporary location..." >&2 + cp "\$0" "\$SCRIPT_COPY" + chmod +x "\$SCRIPT_COPY" + cd "\$TMPROOT" + exec "\$SCRIPT_COPY" --phase2 -- \$initargs + ;; +phase2) + finish="\$finish ; rm -rf \`dirname \$0\`" + ;; +esac + +if test x"\$nox11" = xn; then + if tty -s; then # Do we have a terminal? + : + else + if test x"\$DISPLAY" != x -a x"\$xterm_loop" = x; then # No, but do we have X? + if xset q > /dev/null 2>&1; then # Check for valid DISPLAY variable + GUESS_XTERMS="xterm gnome-terminal rxvt dtterm eterm Eterm xfce4-terminal lxterminal kvt konsole aterm terminology" + for a in \$GUESS_XTERMS; do + if type \$a >/dev/null 2>&1; then + XTERM=\$a + break + fi + done + chmod a+x \$0 || echo Please add execution rights on \$0 + if test \`echo "\$0" | cut -c1\` = "/"; then # Spawn a terminal! + exec \$XTERM -e "\$0 --xwin \$initargs" + else + exec \$XTERM -e "./\$0 --xwin \$initargs" + fi + fi + fi + fi +fi + +if test x"\$targetdir" = x.; then + tmpdir="." +else + if test x"\$keep" = xy; then + if test x"\$nooverwrite" = xy && test -d "\$targetdir"; then + echo "Target directory \$targetdir already exists, aborting." >&2 + exit 1 + fi + if test x"\$quiet" = xn; then + echo "Creating directory \$targetdir" >&2 + fi + tmpdir="\$targetdir" + dashp="-p" + else + tmpdir="\$TMPROOT/selfgz\$\$\$RANDOM" + dashp="" + fi + mkdir \$dashp "\$tmpdir" || { + echo 'Cannot create target directory' \$tmpdir >&2 + echo 'You should try option --extract=' >&2 + eval \$finish + exit 1 + } +fi + +location="\`pwd\`" +if test x"\$SETUP_NOCHECK" != x1; then + MS_Check "\$0" +fi +offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + +if test x"\$verbose" = xy; then + MS_Printf "About to extract $USIZE KB in \$tmpdir ... Proceed ? [Y/n] " + read yn + if test x"\$yn" = xn; then + eval \$finish; exit 1 + fi +fi + +if test x"\$quiet" = xn; then + # Decrypting with openssl will ask for password, + # the prompt needs to start on new line + if test x"$ENCRYPT" = x"openssl"; then + echo "Decrypting and uncompressing \$label..." + else + MS_Printf "Uncompressing \$label" + fi +fi +res=3 +if test x"\$keep" = xn; then + trap MS_cleanup 1 2 3 15 +fi + +if test x"\$nodiskspace" = xn; then + leftspace=\`MS_diskspace "\$tmpdir"\` + if test -n "\$leftspace"; then + if test "\$leftspace" -lt $USIZE; then + echo + echo "Not enough space left in "\`dirname \$tmpdir\`" (\$leftspace KB) to decompress \$0 ($USIZE KB)" >&2 + if test x"\$keep" = xn; then + echo "Consider setting TMPDIR to a directory with more free space." + fi + eval \$finish; exit 1 + fi + fi +fi + +for s in \$filesizes +do + if MS_dd_Progress "\$0" \$offset \$s | MS_Decompress | ( cd "\$tmpdir"; umask \$ORIG_UMASK ; UnTAR xp ) 1>/dev/null; then + if test x"\$ownership" = xy; then + (cd "\$tmpdir"; chown -R \`id -u\` .; chgrp -R \`id -g\` .) + fi + else + echo >&2 + echo "Unable to decompress \$0" >&2 + eval \$finish; exit 1 + fi + offset=\`expr \$offset + \$s\` +done +if test x"\$quiet" = xn; then + echo +fi + +cd "\$tmpdir" +res=0 +if test x"\$script" != x; then + if test x"\$export_conf" = x"y"; then + MS_BUNDLE="\$0" + MS_LABEL="\$label" + MS_SCRIPT="\$script" + MS_SCRIPTARGS="\$scriptargs" + MS_ARCHDIRNAME="\$archdirname" + MS_KEEP="\$KEEP" + MS_NOOVERWRITE="\$NOOVERWRITE" + MS_COMPRESS="\$COMPRESS" + MS_CLEANUP="\$cleanup" + export MS_BUNDLE MS_LABEL MS_SCRIPT MS_SCRIPTARGS + export MS_ARCHDIRNAME MS_KEEP MS_NOOVERWRITE MS_COMPRESS + fi + + if test x"\$verbose" = x"y"; then + yn="x" + while test x"\$yn" != x -a x"\$yn" != xy -a x"\$yn" != xY -a x"\$yn" != xn -a x"\$yn" != xN + do + MS_Printf "OK to execute: \$script \$scriptargs \$* ? [Y/n] " + read yn + if test x"\$yn" = x -o x"\$yn" = xy -o x"\$yn" = xY; then + eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$?; + elif test x"\$yn" = xn -o x"\$yn" = xN; then + echo "Unable to decompress \$script ,because of aborting! ";res=\$? + else + echo "Input value is unacceptable,please try again." + fi + done + else + eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$? + fi + if test "\$res" -ne 0; then + test x"\$verbose" = xy && echo "The program '\$script' returned an error code (\$res)" >&2 + fi +fi + +MS_exec_cleanup + +if test x"\$keep" = xn; then + cd "\$TMPROOT" + rm -rf "\$tmpdir" +fi +eval \$finish; exit \$res +EOF diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.1 b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.1 new file mode 100755 index 000000000..81bf6e4ff --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.1 @@ -0,0 +1,110 @@ +.TH "MAKESELF" "1" "2.4.5" +.SH "NAME" +makeself \- An utility to generate self-extractable archives. +.SH "SYNTAX" +.B makeself [\fIoptions\fP] archive_dir file_name label +.B [\fIstartup_script\fP] [\fIargs\fP] +.SH "DESCRIPTION" +This program is a free (GPL) utility designed to create self-extractable +archives from a directory. +.SH "OPTIONS" +The following options are supported. +.TP 15 +.B -v, --version +Prints out the makeself version number and exits. +.TP +.B -h, --help +Print out help information. +.TP +.B --tar-quietly +Suppress verbose output from the tar command +.TP +.B --quiet +Do not print any messages other than errors +.TP +.B --gzip +Compress using gzip (default if detected). +.TP +.B --bzip2 +Compress using bzip2. +.TP +.B --pbzip2 +Compress using pbzip2. +.TP +.B --xz +Compress using xz. +.TP +.B --lzo +Compress using lzop. +.TP +.B --lz4 +Compress using lz4. +.TP +.B --compress +Compress using the UNIX 'compress' command. +.TP +.B --nocomp +Do not compress the data. +.TP +.B --complevel lvl +Specify the compression level for gzip,bzip2,pbzui2,xz,lzo or lz4 +.TP +.B --notemp +The archive will create archive_dir in the current directory and +uncompress in ./archive_dir. +.TP +.B --copy +Upon extraction, the archive will first copy itself to a temporary directory. +.TP +.B --append +Append more files to an existing makeself archive. The label and startup scripts will then be ignored. +.TP +.B --current +Files will be extracted to the current directory. Both --current and --target dir imply --notemp. +.TP +.B --target dir +Extract directly to a target directory. Directory path can be either absolute or relative. +.TP +.B --header file +Specify location of the header script. +.TP +.B --cleanup file +Specify a cleanup script that executes on interrupt and when finished successfully. +.TP +.B --follow +Follow the symlinks in the archive. +.TP +.B --noprogress +Do not show the progress during the decompression. +.TP +.B --nox11 +Disable automatic spawn of an xterm if running in X11. +.TP +.B --nowait +Do not wait for user input after executing embedded program from an xterm. +.TP +.B --nomd5 +Do not create a MD5 checksum for the archive. +.TP +.B --nocrc +Do not create a CRC32 checksum for the archive. +.TP +.B --lsm file +LSM file describing the package. +.B --packaging-date date +Use provided string as the packaging date instead of the current date. +.SH "EXAMPLES" +Here is an example, assuming the user has a package image stored in a /home/joe/mysoft, +and he wants to generate a self-extracting package named mysoft.sh, which will launch +the "setup" script initially stored in /home/joe/mysoft: +.TP +makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup +.TP +Here is also how I created the makeself.run archive which contains the Makeself distribution: +.TP +makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" +.SH "AUTHORS" +Makeself has been written by Stéphane Peter . +.BR +This man page was originally written by Bartosz Fenski for the +Debian GNU/Linux distribution (but it may be used by others). diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.lsm b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.lsm new file mode 100755 index 000000000..3c4cea8c1 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.lsm @@ -0,0 +1,16 @@ +Begin3 +Title: makeself.sh +Version: 2.4.5 +Description: makeself.sh is a shell script that generates a self-extractable + tar.gz archive from a directory. The resulting file appears as a shell + script, and can be launched as is. The archive will then uncompress + itself to a temporary directory and an arbitrary command will be + executed (for example an installation script). This is pretty similar + to archives generated with WinZip Self-Extractor in the Windows world. +Keywords: Installation archive tar winzip +Author: Stephane Peter (megastep@megastep.org) +Maintained-by: Stephane Peter (megastep@megastep.org) +Original-site: https://makeself.io/ +Platform: Unix +Copying-policy: GPL +End diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.sh new file mode 100755 index 000000000..c8ea56597 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.sh @@ -0,0 +1,822 @@ +#!/bin/sh +# +# Makeself version 2.4.x +# by Stephane Peter +# +# Utility to create self-extracting tar.gz archives. +# The resulting archive is a file holding the tar.gz archive with +# a small Shell script stub that uncompresses the archive to a temporary +# directory and then executes a given script from withing that directory. +# +# Makeself home page: https://makeself.io/ +# +# Version 2.0 is a rewrite of version 1.0 to make the code easier to read and maintain. +# +# Version history : +# - 1.0 : Initial public release +# - 1.1 : The archive can be passed parameters that will be passed on to +# the embedded script, thanks to John C. Quillan +# - 1.2 : Package distribution, bzip2 compression, more command line options, +# support for non-temporary archives. Ideas thanks to Francois Petitjean +# - 1.3 : More patches from Bjarni R. Einarsson and Francois Petitjean: +# Support for no compression (--nocomp), script is no longer mandatory, +# automatic launch in an xterm, optional verbose output, and -target +# archive option to indicate where to extract the files. +# - 1.4 : Improved UNIX compatibility (Francois Petitjean) +# Automatic integrity checking, support of LSM files (Francois Petitjean) +# - 1.5 : Many bugfixes. Optionally disable xterm spawning. +# - 1.5.1 : More bugfixes, added archive options -list and -check. +# - 1.5.2 : Cosmetic changes to inform the user of what's going on with big +# archives (Quake III demo) +# - 1.5.3 : Check for validity of the DISPLAY variable before launching an xterm. +# More verbosity in xterms and check for embedded command's return value. +# Bugfix for Debian 2.0 systems that have a different "print" command. +# - 1.5.4 : Many bugfixes. Print out a message if the extraction failed. +# - 1.5.5 : More bugfixes. Added support for SETUP_NOCHECK environment variable to +# bypass checksum verification of archives. +# - 1.6.0 : Compute MD5 checksums with the md5sum command (patch from Ryan Gordon) +# - 2.0 : Brand new rewrite, cleaner architecture, separated header and UNIX ports. +# - 2.0.1 : Added --copy +# - 2.1.0 : Allow multiple tarballs to be stored in one archive, and incremental updates. +# Added --nochown for archives +# Stopped doing redundant checksums when not necesary +# - 2.1.1 : Work around insane behavior from certain Linux distros with no 'uncompress' command +# Cleaned up the code to handle error codes from compress. Simplified the extraction code. +# - 2.1.2 : Some bug fixes. Use head -n to avoid problems. +# - 2.1.3 : Bug fixes with command line when spawning terminals. +# Added --tar for archives, allowing to give arbitrary arguments to tar on the contents of the archive. +# Added --noexec to prevent execution of embedded scripts. +# Added --nomd5 and --nocrc to avoid creating checksums in archives. +# Added command used to create the archive in --info output. +# Run the embedded script through eval. +# - 2.1.4 : Fixed --info output. +# Generate random directory name when extracting files to . to avoid problems. (Jason Trent) +# Better handling of errors with wrong permissions for the directory containing the files. (Jason Trent) +# Avoid some race conditions (Ludwig Nussel) +# Unset the $CDPATH variable to avoid problems if it is set. (Debian) +# Better handling of dot files in the archive directory. +# - 2.1.5 : Made the md5sum detection consistent with the header code. +# Check for the presence of the archive directory +# Added --encrypt for symmetric encryption through gpg (Eric Windisch) +# Added support for the digest command on Solaris 10 for MD5 checksums +# Check for available disk space before extracting to the target directory (Andreas Schweitzer) +# Allow extraction to run asynchronously (patch by Peter Hatch) +# Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo) +# - 2.1.6 : Replaced one dot per file progress with a realtime progress percentage and a spining cursor (Guy Baconniere) +# Added --noprogress to prevent showing the progress during the decompression (Guy Baconniere) +# Added --target dir to allow extracting directly to a target directory (Guy Baconniere) +# - 2.2.0 : Many bugfixes, updates and contributions from users. Check out the project page on Github for the details. +# - 2.3.0 : Option to specify packaging date to enable byte-for-byte reproducibility. (Marc Pawlowsky) +# - 2.4.0 : Optional support for SHA256 checksums in archives. +# - 2.4.2 : Add support for threads for several compressors. (M. Limber) +# Added zstd support. +# - 2.4.3 : Make explicit POSIX tar archives for increased compatibility. +# - 2.4.5 : Added --tar-format to override ustar tar archive format +# +# (C) 1998-2021 by Stephane Peter +# +# This software is released under the terms of the GNU GPL version 2 and above +# Please read the license at http://www.gnu.org/copyleft/gpl.html +# Self-extracting archives created with this script are explictly NOT released under the term of the GPL +# + +MS_VERSION=2.4.5 +MS_COMMAND="$0" +unset CDPATH + +for f in ${1+"$@"}; do + MS_COMMAND="$MS_COMMAND \\\\ + \\\"$f\\\"" +done + +# For Solaris systems +if test -d /usr/xpg4/bin; then + PATH=/usr/xpg4/bin:$PATH + export PATH +fi + +# Procedures + +MS_Usage() +{ + echo "Usage: $0 [args] archive_dir file_name label startup_script [script_args]" + echo "args can be one or more of the following :" + echo " --version | -v : Print out Makeself version number and exit" + echo " --help | -h : Print out this help message" + echo " --tar-quietly : Suppress verbose output from the tar command" + echo " --quiet | -q : Do not print any messages other than errors." + echo " --gzip : Compress using gzip (default if detected)" + echo " --pigz : Compress with pigz" + echo " --zstd : Compress with zstd" + echo " --bzip2 : Compress using bzip2 instead of gzip" + echo " --pbzip2 : Compress using pbzip2 instead of gzip" + echo " --xz : Compress using xz instead of gzip" + echo " --lzo : Compress using lzop instead of gzip" + echo " --lz4 : Compress using lz4 instead of gzip" + echo " --compress : Compress using the UNIX 'compress' command" + echo " --complevel lvl : Compression level for gzip pigz zstd xz lzo lz4 bzip2 and pbzip2 (default 9)" + echo " --threads thds : Number of threads to be used by compressors that support parallelization." + echo " Omit to use compressor's default. Most useful (and required) for opting" + echo " into xz's threading, usually with '--threads=0' for all available cores." + echo " pbzip2 and pigz are parallel by default, and setting this value allows" + echo " limiting the number of threads they use." + echo " --base64 : Instead of compressing, encode the data using base64" + echo " --gpg-encrypt : Instead of compressing, encrypt the data using GPG" + echo " --gpg-asymmetric-encrypt-sign" + echo " : Instead of compressing, asymmetrically encrypt and sign the data using GPG" + echo " --gpg-extra opt : Append more options to the gpg command line" + echo " --ssl-encrypt : Instead of compressing, encrypt the data using OpenSSL" + echo " --ssl-passwd pass : Use the given password to encrypt the data using OpenSSL" + echo " --ssl-pass-src src : Use the given src as the source of password to encrypt the data" + echo " using OpenSSL. See \"PASS PHRASE ARGUMENTS\" in man openssl." + echo " If this option is not supplied, the user will be asked to enter" + echo " encryption password on the current terminal." + echo " --ssl-no-md : Do not use \"-md\" option not supported by older OpenSSL." + echo " --nochown : Do not give the target folder to the current user (default)" + echo " --chown : Give the target folder to the current user recursively" + echo " --nocomp : Do not compress the data" + echo " --notemp : The archive will create archive_dir in the" + echo " current directory and uncompress in ./archive_dir" + echo " --needroot : Check that the root user is extracting the archive before proceeding" + echo " --copy : Upon extraction, the archive will first copy itself to" + echo " a temporary directory" + echo " --append : Append more files to an existing Makeself archive" + echo " The label and startup scripts will then be ignored" + echo " --target dir : Extract directly to a target directory" + echo " directory path can be either absolute or relative" + echo " --nooverwrite : Do not extract the archive if the specified target directory exists" + echo " --current : Files will be extracted to the current directory" + echo " Both --current and --target imply --notemp" + echo " --tar-format opt : Specify a tar archive format (default is ustar)" + echo " --tar-extra opt : Append more options to the tar command line" + echo " --untar-extra opt : Append more options to the during the extraction of the tar archive" + echo " --nomd5 : Don't calculate an MD5 for archive" + echo " --nocrc : Don't calculate a CRC for archive" + echo " --sha256 : Compute a SHA256 checksum for the archive" + echo " --header file : Specify location of the header script" + echo " --cleanup file : Specify a cleanup script that executes on interrupt and when finished successfully." + echo " --follow : Follow the symlinks in the archive" + echo " --noprogress : Do not show the progress during the decompression" + echo " --nox11 : Disable automatic spawn of a xterm" + echo " --nowait : Do not wait for user input after executing embedded" + echo " program from an xterm" + echo " --sign passphrase : Signature private key to sign the package with" + echo " --lsm file : LSM file describing the package" + echo " --license file : Append a license file" + echo " --help-header file : Add a header to the archive's --help output" + echo " --packaging-date date" + echo " : Use provided string as the packaging date" + echo " instead of the current date." + echo + echo " --keep-umask : Keep the umask set to shell default, rather than overriding when executing self-extracting archive." + echo " --export-conf : Export configuration variables to startup_script" + echo + echo "Do not forget to give a fully qualified startup script name" + echo "(i.e. with a ./ prefix if inside the archive)." + exit 1 +} + +# Default settings +if type gzip >/dev/null 2>&1; then + COMPRESS=gzip +elif type compress >/dev/null 2>&1; then + COMPRESS=compress +else + echo "ERROR: missing commands: gzip, compress" >&2 + MS_Usage +fi +ENCRYPT=n +PASSWD="" +PASSWD_SRC="" +OPENSSL_NO_MD=n +COMPRESS_LEVEL=9 +DEFAULT_THREADS=123456 # Sentinel value +THREADS=$DEFAULT_THREADS +KEEP=n +CURRENT=n +NOX11=n +NOWAIT=n +APPEND=n +TAR_QUIETLY=n +KEEP_UMASK=n +QUIET=n +NOPROGRESS=n +COPY=none +NEED_ROOT=n +TAR_ARGS=rvf +TAR_FORMAT=ustar +TAR_EXTRA="" +GPG_EXTRA="" +DU_ARGS=-ks +HEADER=`dirname "$0"`/makeself-header.sh +SIGNATURE="" +TARGETDIR="" +NOOVERWRITE=n +DATE=`LC_ALL=C date` +EXPORT_CONF=n +SHA256=n +OWNERSHIP=n +SIGN=n +GPG_PASSPHRASE="" + +# LSM file stuff +LSM_CMD="echo No LSM. >> \"\$archname\"" + +while true +do + case "$1" in + --version | -v) + echo Makeself version $MS_VERSION + exit 0 + ;; + --pbzip2) + COMPRESS=pbzip2 + shift + ;; + --bzip2) + COMPRESS=bzip2 + shift + ;; + --gzip) + COMPRESS=gzip + shift + ;; + --pigz) + COMPRESS=pigz + shift + ;; + --zstd) + COMPRESS=zstd + shift + ;; + --xz) + COMPRESS=xz + shift + ;; + --lzo) + COMPRESS=lzo + shift + ;; + --lz4) + COMPRESS=lz4 + shift + ;; + --compress) + COMPRESS=compress + shift + ;; + --base64) + COMPRESS=base64 + shift + ;; + --gpg-encrypt) + COMPRESS=gpg + shift + ;; + --gpg-asymmetric-encrypt-sign) + COMPRESS=gpg-asymmetric + shift + ;; + --gpg-extra) + GPG_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-encrypt) + ENCRYPT=openssl + shift + ;; + --ssl-passwd) + PASSWD=$2 + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-pass-src) + PASSWD_SRC=$2 + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-no-md) + OPENSSL_NO_MD=y + shift + ;; + --nocomp) + COMPRESS=none + shift + ;; + --complevel) + COMPRESS_LEVEL="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --threads) + THREADS="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --nochown) + OWNERSHIP=n + shift + ;; + --chown) + OWNERSHIP=y + shift + ;; + --notemp) + KEEP=y + shift + ;; + --copy) + COPY=copy + shift + ;; + --current) + CURRENT=y + KEEP=y + shift + ;; + --tar-format) + TAR_FORMAT="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --tar-extra) + TAR_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --untar-extra) + UNTAR_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --target) + TARGETDIR="$2" + KEEP=y + shift 2 || { MS_Usage; exit 1; } + ;; + --sign) + SIGN=y + GPG_PASSPHRASE="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --nooverwrite) + NOOVERWRITE=y + shift + ;; + --needroot) + NEED_ROOT=y + shift + ;; + --header) + HEADER="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --cleanup) + CLEANUP_SCRIPT="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --license) + # We need to escape all characters having a special meaning in double quotes + LICENSE=$(sed 's/\\/\\\\/g; s/"/\\\"/g; s/`/\\\`/g; s/\$/\\\$/g' "$2") + shift 2 || { MS_Usage; exit 1; } + ;; + --follow) + TAR_ARGS=rvhf + DU_ARGS=-ksL + shift + ;; + --noprogress) + NOPROGRESS=y + shift + ;; + --nox11) + NOX11=y + shift + ;; + --nowait) + NOWAIT=y + shift + ;; + --nomd5) + NOMD5=y + shift + ;; + --sha256) + SHA256=y + shift + ;; + --nocrc) + NOCRC=y + shift + ;; + --append) + APPEND=y + shift + ;; + --lsm) + LSM_CMD="cat \"$2\" >> \"\$archname\"" + shift 2 || { MS_Usage; exit 1; } + ;; + --packaging-date) + DATE="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --help-header) + HELPHEADER=`sed -e "s/'/'\\\\\''/g" $2` + shift 2 || { MS_Usage; exit 1; } + [ -n "$HELPHEADER" ] && HELPHEADER="$HELPHEADER +" + ;; + --tar-quietly) + TAR_QUIETLY=y + shift + ;; + --keep-umask) + KEEP_UMASK=y + shift + ;; + --export-conf) + EXPORT_CONF=y + shift + ;; + -q | --quiet) + QUIET=y + shift + ;; + -h | --help) + MS_Usage + ;; + -*) + echo Unrecognized flag : "$1" + MS_Usage + ;; + *) + break + ;; + esac +done + +if test $# -lt 1; then + MS_Usage +else + if test -d "$1"; then + archdir="$1" + else + echo "Directory $1 does not exist." >&2 + exit 1 + fi +fi +archname="$2" + +if test "$QUIET" = "y" || test "$TAR_QUIETLY" = "y"; then + if test "$TAR_ARGS" = "rvf"; then + TAR_ARGS="rf" + elif test "$TAR_ARGS" = "rvhf"; then + TAR_ARGS="rhf" + fi +fi + +if test "$APPEND" = y; then + if test $# -lt 2; then + MS_Usage + fi + + # Gather the info from the original archive + OLDENV=`sh "$archname" --dumpconf` + if test $? -ne 0; then + echo "Unable to update archive: $archname" >&2 + exit 1 + else + eval "$OLDENV" + OLDSKIP=`expr $SKIP + 1` + fi +else + if test "$KEEP" = n -a $# = 3; then + echo "ERROR: Making a temporary archive with no embedded command does not make sense!" >&2 + echo >&2 + MS_Usage + fi + # We don't want to create an absolute directory unless a target directory is defined + if test "$CURRENT" = y; then + archdirname="." + elif test x"$TARGETDIR" != x; then + archdirname="$TARGETDIR" + else + archdirname=`basename "$1"` + fi + + if test $# -lt 3; then + MS_Usage + fi + + LABEL="$3" + SCRIPT="$4" + test "x$SCRIPT" = x || shift 1 + shift 3 + SCRIPTARGS="$*" +fi + +if test "$KEEP" = n -a "$CURRENT" = y; then + echo "ERROR: It is A VERY DANGEROUS IDEA to try to combine --notemp and --current." >&2 + exit 1 +fi + +case $COMPRESS in +gzip) + GZIP_CMD="gzip -c$COMPRESS_LEVEL" + GUNZIP_CMD="gzip -cd" + ;; +pigz) + GZIP_CMD="pigz -$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD --processes $THREADS" + fi + GUNZIP_CMD="gzip -cd" + ;; +zstd) + GZIP_CMD="zstd -$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD --threads=$THREADS" + fi + GUNZIP_CMD="zstd -cd" + ;; +pbzip2) + GZIP_CMD="pbzip2 -c$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD -p$THREADS" + fi + GUNZIP_CMD="bzip2 -d" + ;; +bzip2) + GZIP_CMD="bzip2 -$COMPRESS_LEVEL" + GUNZIP_CMD="bzip2 -d" + ;; +xz) + GZIP_CMD="xz -c$COMPRESS_LEVEL" + # Must opt-in by specifying a value since not all versions of xz support threads + if test $THREADS -ne $DEFAULT_THREADS; then + GZIP_CMD="$GZIP_CMD --threads=$THREADS" + fi + GUNZIP_CMD="xz -d" + ;; +lzo) + GZIP_CMD="lzop -c$COMPRESS_LEVEL" + GUNZIP_CMD="lzop -d" + ;; +lz4) + GZIP_CMD="lz4 -c$COMPRESS_LEVEL" + GUNZIP_CMD="lz4 -d" + ;; +base64) + GZIP_CMD="base64" + GUNZIP_CMD="base64 --decode -i -" + ;; +gpg) + GZIP_CMD="gpg $GPG_EXTRA -ac -z$COMPRESS_LEVEL" + GUNZIP_CMD="gpg -d" + ENCRYPT="gpg" + ;; +gpg-asymmetric) + GZIP_CMD="gpg $GPG_EXTRA -z$COMPRESS_LEVEL -es" + GUNZIP_CMD="gpg --yes -d" + ENCRYPT="gpg" + ;; +compress) + GZIP_CMD="compress -fc" + GUNZIP_CMD="(type compress >/dev/null 2>&1 && compress -fcd || gzip -cd)" + ;; +none) + GZIP_CMD="cat" + GUNZIP_CMD="cat" + ;; +esac + +if test x"$ENCRYPT" = x"openssl"; then + if test x"$APPEND" = x"y"; then + echo "Appending to existing archive is not compatible with OpenSSL encryption." >&2 + fi + + ENCRYPT_CMD="openssl enc -aes-256-cbc -salt" + DECRYPT_CMD="openssl enc -aes-256-cbc -d" + + if test x"$OPENSSL_NO_MD" != x"y"; then + ENCRYPT_CMD="$ENCRYPT_CMD -md sha256" + DECRYPT_CMD="$DECRYPT_CMD -md sha256" + fi + + if test -n "$PASSWD_SRC"; then + ENCRYPT_CMD="$ENCRYPT_CMD -pass $PASSWD_SRC" + elif test -n "$PASSWD"; then + ENCRYPT_CMD="$ENCRYPT_CMD -pass pass:$PASSWD" + fi +fi + +tmpfile="${TMPDIR:-/tmp}/mkself$$" + +if test -f "$HEADER"; then + oldarchname="$archname" + archname="$tmpfile" + # Generate a fake header to count its lines + SKIP=0 + . "$HEADER" + SKIP=`cat "$tmpfile" |wc -l` + # Get rid of any spaces + SKIP=`expr $SKIP` + rm -f "$tmpfile" + if test "$QUIET" = "n"; then + echo "Header is $SKIP lines long" >&2 + fi + archname="$oldarchname" +else + echo "Unable to open header file: $HEADER" >&2 + exit 1 +fi + +if test "$QUIET" = "n"; then + echo +fi + +if test "$APPEND" = n; then + if test -f "$archname"; then + echo "WARNING: Overwriting existing file: $archname" >&2 + fi +fi + +USIZE=`du $DU_ARGS "$archdir" | awk '{print $1}'` + +if test "." = "$archdirname"; then + if test "$KEEP" = n; then + archdirname="makeself-$$-`date +%Y%m%d%H%M%S`" + fi +fi + +test -d "$archdir" || { echo "Error: $archdir does not exist."; rm -f "$tmpfile"; exit 1; } +if test "$QUIET" = "n"; then + echo "About to compress $USIZE KB of data..." + echo "Adding files to archive named \"$archname\"..." +fi + +# See if we have GNU tar +TAR=`exec <&- 2>&-; which gtar || command -v gtar || type gtar` +test -x "$TAR" || TAR=tar + +tmparch="${TMPDIR:-/tmp}/mkself$$.tar" +( + if test "$APPEND" = "y"; then + tail -n "+$OLDSKIP" "$archname" | eval "$GUNZIP_CMD" > "$tmparch" + fi + cd "$archdir" + # "Determining if a directory is empty" + # https://www.etalabs.net/sh_tricks.html + find . \ + \( \ + ! -type d \ + -o \ + \( -links 2 -exec sh -c ' + is_empty () ( + cd "$1" + set -- .[!.]* ; test -f "$1" && return 1 + set -- ..?* ; test -f "$1" && return 1 + set -- * ; test -f "$1" && return 1 + return 0 + ) + is_empty "$0"' {} \; \ + \) \ + \) -print \ + | LC_ALL=C sort \ + | sed 's/./\\&/g' \ + | xargs $TAR $TAR_EXTRA --format $TAR_FORMAT -$TAR_ARGS "$tmparch" +) || { + echo "ERROR: failed to create temporary archive: $tmparch" + rm -f "$tmparch" "$tmpfile" + exit 1 +} + +USIZE=`du $DU_ARGS "$tmparch" | awk '{print $1}'` + +eval "$GZIP_CMD" <"$tmparch" >"$tmpfile" || { + echo "ERROR: failed to create temporary file: $tmpfile" + rm -f "$tmparch" "$tmpfile" + exit 1 +} +rm -f "$tmparch" + +if test x"$ENCRYPT" = x"openssl"; then + echo "About to encrypt archive \"$archname\"..." + { eval "$ENCRYPT_CMD -in $tmpfile -out ${tmpfile}.enc" && mv -f ${tmpfile}.enc $tmpfile; } || \ + { echo Aborting: could not encrypt temporary file: "$tmpfile".; rm -f "$tmpfile"; exit 1; } +fi + +fsize=`cat "$tmpfile" | wc -c | tr -d " "` + +# Compute the checksums + +shasum=0000000000000000000000000000000000000000000000000000000000000000 +md5sum=00000000000000000000000000000000 +crcsum=0000000000 + +if test "$NOCRC" = y; then + if test "$QUIET" = "n"; then + echo "skipping crc at user request" + fi +else + crcsum=`CMD_ENV=xpg4 cksum < "$tmpfile" | sed -e 's/ /Z/' -e 's/ /Z/' | cut -dZ -f1` + if test "$QUIET" = "n"; then + echo "CRC: $crcsum" + fi +fi + +if test "$SHA256" = y; then + SHA_PATH=`exec <&- 2>&-; which shasum || command -v shasum || type shasum` + if test -x "$SHA_PATH"; then + shasum=`eval "$SHA_PATH -a 256" < "$tmpfile" | cut -b-64` + else + SHA_PATH=`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum` + shasum=`eval "$SHA_PATH" < "$tmpfile" | cut -b-64` + fi + if test "$QUIET" = "n"; then + if test -x "$SHA_PATH"; then + echo "SHA256: $shasum" + else + echo "SHA256: none, SHA command not found" + fi + fi +fi +if test "$NOMD5" = y; then + if test "$QUIET" = "n"; then + echo "Skipping md5sum at user request" + fi +else + # Try to locate a MD5 binary + OLD_PATH=$PATH + PATH=${GUESS_MD5_PATH:-"$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} + MD5_ARG="" + MD5_PATH=`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum` + test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which md5 || command -v md5 || type md5` + test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which digest || command -v digest || type digest` + PATH=$OLD_PATH + if test -x "$MD5_PATH"; then + if test `basename ${MD5_PATH}`x = digestx; then + MD5_ARG="-a md5" + fi + md5sum=`eval "$MD5_PATH $MD5_ARG" < "$tmpfile" | cut -b-32` + if test "$QUIET" = "n"; then + echo "MD5: $md5sum" + fi + else + if test "$QUIET" = "n"; then + echo "MD5: none, MD5 command not found" + fi + fi +fi +if test "$SIGN" = y; then + GPG_PATH=`exec <&- 2>&-; which gpg || command -v gpg || type gpg` + if test -x "$GPG_PATH"; then + SIGNATURE=`$GPG_PATH --pinentry-mode=loopback --batch --yes --passphrase "$GPG_PASSPHRASE" --output - --detach-sig $tmpfile | base64 | tr -d \\\\n` + if test "$QUIET" = "n"; then + echo "Signature: $SIGNATURE" + fi + else + echo "Missing gpg command" >&2 + fi +fi + +totalsize=0 +for size in $fsize; +do + totalsize=`expr $totalsize + $size` +done + +if test "$APPEND" = y; then + mv "$archname" "$archname".bak || exit + + # Prepare entry for new archive + filesizes="$fsize" + CRCsum="$crcsum" + MD5sum="$md5sum" + SHAsum="$shasum" + Signature="$SIGNATURE" + # Generate the header + . "$HEADER" + # Append the new data + cat "$tmpfile" >> "$archname" + + chmod +x "$archname" + rm -f "$archname".bak + if test "$QUIET" = "n"; then + echo "Self-extractable archive \"$archname\" successfully updated." + fi +else + filesizes="$fsize" + CRCsum="$crcsum" + MD5sum="$md5sum" + SHAsum="$shasum" + Signature="$SIGNATURE" + + # Generate the header + . "$HEADER" + + # Append the compressed tar data after the stub + if test "$QUIET" = "n"; then + echo + fi + cat "$tmpfile" >> "$archname" + chmod +x "$archname" + if test "$QUIET" = "n"; then + echo Self-extractable archive \"$archname\" successfully created. + fi +fi +rm -f "$tmpfile" diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/run-tests.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/run-tests.sh new file mode 100755 index 000000000..31ee16511 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/run-tests.sh @@ -0,0 +1,8 @@ +#!/bin/sh +# Run every available test - Bash needed +cd test +for test in *test; +do + echo "Running test $test ..." + bash $test || { echo "*** ERROR: Test '$test' failed!"; exit 1; } +done diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/merge_aicpu_info_json.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/merge_aicpu_info_json.sh new file mode 100755 index 000000000..a977bd51d --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/merge_aicpu_info_json.sh @@ -0,0 +1,31 @@ +#!/bin/bash + +project_path=$1 +build_path=$2 +vendor_name=customize +echo $@ +if [[ ! -d "$project_path" ]]; then + echo "[ERROR] No projcet path is provided" + exit 1 +fi + +if [[ ! -d "$build_path" ]]; then + echo "[ERROR] No build path is provided" + exit 1 +fi + +if [[ ! -d "$ASCEND_OPP_PATH" ]]; then + echo "[ERROR] No opp install path is provided" + exit 1 +fi +custom_exist_info_json=$ASCEND_OPP_PATH/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json +custom_new_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json +temp_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/temp_cust_aicpu_kernel.json + +if [[ -f "$custom_exist_info_json" ]] && [[ -f "$custom_new_info_json" ]]; then + cp -f $custom_exist_info_json $temp_info_json + chmod +w $temp_info_json + python3 ${project_path}/cmake/util/insert_op_info.py ${custom_new_info_json} ${temp_info_json} + cp -f $temp_info_json $custom_new_info_json + rm -f $temp_info_json +fi diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/opdesc_parser.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/opdesc_parser.py new file mode 100755 index 000000000..0d3221602 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/opdesc_parser.py @@ -0,0 +1,298 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os + + +OP_ALL = '__ALLOP__' +SOC_ALL = '__ALLSOC__' +SOC_TO_SHORT_SOC_MAP = { + "ascend910a": "ascend910", + "ascend910proa": "ascend910", + "ascend910b": "ascend910", + "ascend910prob": "ascend910", + "ascend910premiuma": "ascend910", + "ascend910b1": "ascend910b", + "ascend910b2": "ascend910b", + "ascend910b2c": "ascend910b", + "ascend910b3": "ascend910b", + "ascend910b4": "ascend910b", + "ascend910c1": "ascend910c", + "ascend910c2": "ascend910c", + "ascend910c3": "ascend910c", + "ascend910c4": "ascend910c", + "ascend310p1": "ascend310p", + "ascend310p3": "ascend310p", + "ascend310p3vir01": "ascend310p", + "ascend310p3vir02": "ascend310p", + "ascend310p3vir04": "ascend310p", + "ascend310p3vir08": "ascend310p", + "ascend310b1": "ascend310b", + "bs9sx1aa": "bs9sx1a" +} +CONFLICT_KEYWORDS = { + "and", "as", "assert", "break", "class", "continue", "def", "del", "elif", "else", + "except", "finally", "for", "from", "global", "if", "import", "in", "is", "lambda", + "not", "or", "pass", "raise", "return", "try", "while", "with", "yield", "False", + "None", "True", "nonlocal", "arg", "__inputs__", "__outputs__", "options", "bisheng", + "bisheng_path", "tikcpp_path", "impl_mode", "custom_compile_options", + "custom_all_compile_options", "soc_version", "soc_short", "custom_compile_options_soc", + "custom_all_compile_options_soc", "origin_func_name", "ascendc_src_dir_ex", + "ascendc_src_dir", "ascendc_src_file", "src", "op_type", "code_channel", "op_info", + "compile_op", "get_code_channel", "result", "__attrs__", "isinstance", "attr", + "get_current_build_config", "_build_args", "get_dtype_fmt_options", "shutil", "os", + "get_kernel_source" +} + + +class OpDesc: + def __init__(self: any, op_type: str): + self.op_type = op_type + self.attr_list = [] + self.attr_val = {} + self.input_name = [] + self.input_ori_name = [] + self.input_type = [] + self.input_dtype = [] + self.input_fmt = [] + self.output_name = [] + self.output_ori_name = [] + self.output_type = [] + self.output_dtype = [] + self.output_fmt = [] + self.op_fmt_sel = False + self.op_chk_support = False + self.op_intf = '' + self.kern_name = '' + self.op_file = '' + self.op_replay_flag = False + self.op_replay_batch = False + self.input_idx = -1 + self.output_idx = -1 + self.max_block_dim = 32 + self.max_shape_size = 268435456 + self.dynamic_shape = False + self.op_range_limit = '' + self.custom_compile_options = {} + self.custom_all_compile_options = {} + + @staticmethod + def _parse_digit(conf: str) -> int: + return int(conf.split('=')[1]) + + @staticmethod + def _parse_flag(conf: str) -> bool: + if 'true' == conf.split('=')[1]: + return True + return False + + @staticmethod + def _parse_str(conf: str) -> str: + return conf.split('=')[1] + + @staticmethod + def _parse_list(conf: str) -> list: + return conf.split('=')[1].split(',') + + def parse_input(self: any, conf: str): + if conf.startswith('input{}.name'.format(int(self.input_idx) + 1)): + self.input_idx += 1 + self.input_ori_name.append(self._parse_str(conf)) + self.input_name.append(self.input_ori_name[-1] + '_in__') + elif conf.startswith('input{}.paramType'.format(int(self.input_idx))): + self.input_type.append(self._parse_str(conf)) + elif conf.startswith('input{}.dtype'.format(int(self.input_idx))): + self.input_dtype.append(self._parse_str(conf)) + elif conf.startswith('input{}.format'.format(int(self.input_idx))): + self.input_fmt.append(self._parse_str(conf)) + else: + return + + def parse_output(self: any, conf: str): + if conf.startswith('output{}.name'.format(int(self.output_idx) + 1)): + self.output_idx += 1 + self.output_ori_name.append(self._parse_str(conf)) + self.output_name.append(self.output_ori_name[-1] + '_out_') + elif conf.startswith('output{}.paramType'.format(int(self.output_idx))): + self.output_type.append(self._parse_str(conf)) + elif conf.startswith('output{}.dtype'.format(int(self.output_idx))): + self.output_dtype.append(self._parse_str(conf)) + elif conf.startswith('output{}.format'.format(int(self.output_idx))): + self.output_fmt.append(self._parse_str(conf)) + else: + return + + def parse_op_format(self: any, conf: str): + self.op_fmt_sel = self._parse_flag(conf) + + def parse_check_support(self: any, conf: str): + self.op_chk_support = self._parse_flag(conf) + + def parse_range_limit(self: any, conf: str): + self.op_range_limit = self._parse_str(conf) + + def parse_kern_name(self: any, conf: str): + self.kern_name = self._parse_str(conf) + + def parse_op_intf(self: any, conf: str): + self.op_intf = self._parse_str(conf) + + def parse_op_file(self: any, conf: str): + self.op_file = self._parse_str(conf) + + def parse_dynamic_shape(self: any, conf: str): + self.dynamic_shape = self._parse_flag(conf) + + def parse_attr_list(self: any, conf: str): + self.attr_list = self._parse_list(conf) + intersection_element = set(self.attr_list) & CONFLICT_KEYWORDS + if intersection_element: + raise Exception(f'The attribute name: {intersection_element} in op: {{\'{self.op_type}\'}} conflicts with\ + the built-in variable name. Use a complex name or prefix the operator name.') + + @staticmethod + def _camel_to_snake(camel_case_str: str): + snake_case_str = '' + for i, c in enumerate(camel_case_str): + if i == 0: + snake_case_str += c.lower() + elif c.isupper(): + snake_case_str += '_' + c.lower() + else: + snake_case_str += c + return snake_case_str + + def parse_attr_val(self: any, conf: str): + for attr in self.attr_list: + if self.attr_val.get(attr) is None: + self.attr_val[attr] = {} + if conf.startswith('attr_{}.type'.format(attr)): + self.attr_val.get(attr)['type'] = self._camel_to_snake(self._parse_str(conf)) + elif conf.startswith('attr_{}.paramType'.format(attr)): + self.attr_val.get(attr)['paramType'] = self._parse_str(conf) + elif conf.startswith('attr_{}.defaultValue'.format(attr)): + self.attr_val.get(attr)['defaultValue'] = self._parse_str(conf) + + def parse_replay_val(self: any, batch_list: list, iterator_list: list): + if self.op_type in batch_list: + self.op_replay_flag = True + self.op_replay_batch = True + elif self.op_type in iterator_list: + self.op_replay_flag = True + self.op_replay_batch = False + + +def _is_op_type_in_opdesc(op_descs: list, op_type: str): + for op in op_descs: + if op_type == op.op_type: + return True + return False + + +def _set_all_options_to_opdescs(op_descs, soc_ver_compile_options): + for op in op_descs: + op.custom_all_compile_options = soc_ver_compile_options + + +def _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options): + for op in op_descs: + if op.op_type != op_type: + continue + op.custom_compile_options = soc_ver_compile_options + + +def _trans_soc_ver_to_short(soc_ver: str): + low_soc_ver = soc_ver.lower() + if low_soc_ver not in SOC_TO_SHORT_SOC_MAP: + print(f'WARNING: caution: {soc_ver} will trans into ascend910, if not your intention,' + f'use ascend910b1~4 instead') + return SOC_TO_SHORT_SOC_MAP[low_soc_ver] + + +def _get_op_custom_options(op_descs: list, auto_gen_dir: str): + if auto_gen_dir is None: + return + file = os.path.join(auto_gen_dir, "custom_compile_options.ini") + if not os.path.exists(file): + print(f'WARNING: cannot find {auto_gen_dir}/custom_compile_options.ini') + return + with open(file, 'r') as fd: + lines = fd.readlines() + for line in lines: + param_list = str.split(line.rstrip('\n'), ',') + if len(param_list) != 3: + raise Exception(f'ERROR: custom compile option {param_list} len is not 3') + op_type = param_list[0] + if op_type.upper() == 'ALL': + op_type = OP_ALL + if op_type != OP_ALL and _is_op_type_in_opdesc(op_descs, op_type) == False: + print(f'WARNING: op: {op_type} are not exists in this project') + continue + soc_ver_compile_options = {} + soc_ver = param_list[1] + options_str = param_list[2] + options = str.split(options_str, ';') + if soc_ver == '': + soc_ver_compile_options[SOC_ALL] = options + else: + soc_ver_list = str.split(soc_ver, ';') + for ver in soc_ver_list: + short_ver = _trans_soc_ver_to_short(ver) + soc_ver_compile_options[short_ver] = options + if op_type == OP_ALL: + _set_all_options_to_opdescs(op_descs, soc_ver_compile_options) + else: + _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options) + + +def get_op_desc(file: str, batch_list: list, iterator_list: list, builder: any, + op_type: list, auto_gen_dir: str = None) -> list: + op_descs = [] + op_match = False + with open (file, 'r') as fd: + lines = fd.readlines() + for line in lines: + line = line.strip() + if line.startswith('['): + name = line[1:-1] + if op_type is None or name in op_type: + op_match = True + op_desc = builder(name) + op_desc.parse_replay_val(batch_list, iterator_list) + op_descs.append(op_desc) + else: + op_match = False + if op_type is not None and len(op_descs) == len(op_type): + return op_descs + continue + if not op_match: + continue + if line.startswith('input'): + op_desc.parse_input(line) + elif line.startswith('output'): + op_desc.parse_output(line) + elif line.startswith('dynamicFormat.flag'): + op_desc.parse_op_format(line) + elif line.startswith('needCheckSupport.flag'): + op_desc.parse_check_support(line) + elif line.startswith('rangeLimit.value'): + op_desc.parse_range_limit(line) + elif line.startswith('opInterface.value'): + op_desc.parse_op_intf(line) + elif line.startswith('kernel.name'): + op_desc.parse_kern_name(line) + elif line.startswith('opFile.value'): + op_desc.parse_op_file(line) + elif line.startswith('dynamicShapeSupport.flag'): + op_desc.parse_dynamic_shape(line) + elif line.startswith('attr.list'): + op_desc.parse_attr_list(line) + elif line.startswith('attr_'): + op_desc.parse_attr_val(line) + _get_op_custom_options(op_descs, auto_gen_dir) + return op_descs diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/parse_ini_to_json.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/parse_ini_to_json.py new file mode 100755 index 000000000..ac931e5d3 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/parse_ini_to_json.py @@ -0,0 +1,338 @@ +# Copyright 2020-2021 Huawei Technologies Co., Ltd +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +""" +parser ini to json +""" + +import json +import os +import stat +import sys + + +ATTR_TYPE_LIST = ["int", "float", "bool", "str", "listInt", "listFloat", "listBool", "listStr", "listListInt", + "type", "listType", "tensor", "listTensor"] +ATTR_PARAMTYPE_LIST = ["optional", "required"] +BOOL_FLAG_KEY = ["dynamicFormat", "dynamicShapeSupport", "dynamicRankSupport", "precision_reduce", "heavyOp", + "needCheckSupport", "enableVectorCore"] +BOOL_LIST = ["true", "false"] +DTYPE_LIST = ["float16", "float", "float32", "int8", "int16", "int32", "uint8", "uint16", "uint32", "bool", + "int64", "uint64", "qint8", "qint16", "qint32", "quint8", "quint16", "double", "complex64", + "complex128", "string", "resource", "dual", "dual_sub_int8", "dual_sub_uint8", "string_ref", + "int4", "bfloat16", "uint1"] +FORMAT_LIST = ["NCHW", "NHWC", "ND", "NC1HWC0", "FRACTAL_Z", "NC1C0HWPAD", "NHWC1C0", "FSR_NCHW", "FRACTAL_DECONV", + "C1HWNC0", "FRACTAL_DECONV_TRANSPOSE", "FRACTAL_DECONV_SP_STRIDE_TRANS", "NC1HWC0_C04", + "FRACTAL_Z_C04", "CHWN", "FRACTAL_DECONV_SP_STRIDE8_TRANS", "HWCN", "NC1KHKWHWC0", "BN_WEIGHT", + "FILTER_HWCK", "HASHTABLE_LOOKUP_LOOKUPS", "HASHTABLE_LOOKUP_KEYS", "HASHTABLE_LOOKUP_VALUE", + "HASHTABLE_LOOKUP_OUTPUT", "HASHTABLE_LOOKUP_HITS", "C1HWNCoC0", "MD", "NDHWC", "FRACTAL_ZZ", + "FRACTAL_NZ", "NCDHW", "DHWCN", "NDC1HWC0", "FRACTAL_Z_3D", "CN", "NC", "DHWNC", + "FRACTAL_Z_3D_TRANSPOSE", "FRACTAL_ZN_LSTM", "FRACTAL_ZN_RNN", "FRACTAL_Z_G", "NULL"] + + +def parse_ini_files(ini_files): + """ + parse ini files to json + Parameters: + ---------------- + ini_files:input file list + return:ops_info + ---------------- + """ + tbe_ops_info = {} + for ini_file in ini_files: + check_file_size(ini_file) + parse_ini_to_obj(ini_file, tbe_ops_info) + return tbe_ops_info + + +def check_file_size(input_file): + try: + file_size = os.path.getsize(input_file) + except OSError as os_error: + print('[ERROR] Failed to open "%s". %s' % (input_file, str(os_error))) + raise OSError from os_error + if file_size > 10*1024*1024: + print('[WARN] The size of %s exceeds 10MB, it may take more time to run, please wait.' % input_file) + + +def parse_ini_to_obj(ini_file, tbe_ops_info): + """ + parse ini file to json obj + Parameters: + ---------------- + ini_file:ini file path + tbe_ops_info:ops_info + ---------------- + """ + with open(ini_file) as ini_file: + lines = ini_file.readlines() + op_dict = {} + op_name = "" + find_op_type = False + for line in lines: + line = line.rstrip() + if line == "": + continue + if line.startswith("["): + if line.endswith("]"): + op_name = line[1:-1] + op_dict = {} + tbe_ops_info[op_name] = op_dict + find_op_type = True + elif "=" in line: + key1 = line[:line.index("=")] + key2 = line[line.index("=")+1:] + key1_0, key1_1 = key1.split(".") + if key1_0 not in op_dict: + op_dict[key1_0] = {} + if key1_1 in op_dict.get(key1_0): + raise RuntimeError("Op:" + op_name + " " + key1_0 + " " + + key1_1 + " is repeated!") + dic_key = op_dict.get(key1_0) + dic_key[key1_1] = key2 + else: + continue + if not find_op_type: + raise RuntimeError("Not find OpType in .ini file.") + + +def check_output_exist(op_dict, is_valid): + """ + Function Description: + Check output is exist + Parameter: op_dict + Parameter: is_valid + """ + if "output0" in op_dict: + output0_dict = op_dict.get("output0") + if output0_dict.get("name", None) is None: + is_valid = False + print("output0.name is required in .ini file!") + else: + is_valid = False + print("output0 is required in .ini file!") + return is_valid + + +def check_attr_dict(attr_dict, is_valid, attr): + """ + Function Description: + Check attr_dict + Parameter: attr_dict + Parameter: is_valid + Parameter: attr + """ + attr_type = attr_dict.get("type") + value = attr_dict.get("value") + param_type = attr_dict.get("paramType") + if attr_type is None or value is None: + is_valid = False + print("If attr.list is exist, {0}.type and {0}.value is required".format(attr)) + if param_type and param_type not in ATTR_PARAMTYPE_LIST: + is_valid = False + print("{0}.paramType only support {1}.".format(attr, ATTR_PARAMTYPE_LIST)) + if attr_type and attr_type not in ATTR_TYPE_LIST: + is_valid = False + print("{0}.type only support {1}.".format(attr, ATTR_TYPE_LIST)) + return is_valid + + +def check_attr(op_dict, is_valid): + """ + Function Description: + Check attr + Parameter: op_dict + Parameter: is_valid + """ + if "attr" in op_dict: + attr_dict = op_dict.get("attr") + attr_list_str = attr_dict.get("list", None) + if attr_list_str is None: + is_valid = False + print("attr.list is required in .ini file!") + else: + attr_list = attr_list_str.split(",") + for attr_name in attr_list: + attr = "attr_" + attr_name.strip() + attr_dict = op_dict.get(attr) + if attr_dict: + is_valid = check_attr_dict(attr_dict, is_valid, attr) + else: + is_valid = False + print("%s is required in .ini file, when attr.list is %s!" % (attr, attr_list_str)) + return is_valid + + +def check_bool_flag(op_dict, is_valid): + """ + Function Description: + check_bool_flag + Parameter: op_dict + Parameter: is_valid + """ + for key in BOOL_FLAG_KEY: + if key in op_dict: + op_bool_key = op_dict.get(key) + if op_bool_key.get("flag").strip() not in BOOL_LIST: + is_valid = False + print("{0}.flag only support {1}.".format(key, BOOL_LIST)) + return is_valid + + +def check_type_format(op_info, is_valid, op_info_key): + """ + Function Description: + Check type and format + Parameter: op_info + Parameter: is_valid + Parameter: op_info_key + """ + op_info_dtype_str = op_info.get("dtype") + op_info_dtype_num = 0 + op_info_format_num = 0 + if op_info_dtype_str: + op_info_dtype = op_info_dtype_str.split(",") + op_info_dtype_num = len(op_info_dtype) + for dtype in op_info_dtype: + if dtype.strip() not in DTYPE_LIST: + is_valid = False + print("{0}.dtype not support {1}.".format(op_info_key, dtype)) + op_info_format_str = op_info.get("format") + if op_info_format_str: + op_info_format = op_info_format_str.split(",") + op_info_format_num = len(op_info_format) + for op_format in op_info_format: + if op_format.strip() not in FORMAT_LIST: + is_valid = False + print("{0}.format not support {1}.".format(op_info_key, op_format)) + if op_info_dtype_num > 0 and op_info_format_num > 0: + if op_info_dtype_num != op_info_format_num: + is_valid = False + print("The number of {0}.dtype not match the number of {0}.format.".format(op_info_key)) + return is_valid + + +def check_op_info(tbe_ops): + """ + Function Description: + Check info. + Parameter: tbe_ops + Return Value: is_valid + """ + print("\n\n==============check valid for ops info start==============") + required_op_input_info_keys = ["paramType", "name"] + required_op_output_info_keys = ["paramType", "name"] + param_type_valid_value = ["dynamic", "optional", "required"] + is_valid = True + for op_key in tbe_ops: + op_dict = tbe_ops[op_key] + for op_info_key in op_dict: + if op_info_key.startswith("input"): + op_input_info = op_dict[op_info_key] + missing_keys = [] + for required_op_input_info_key in required_op_input_info_keys: + if required_op_input_info_key not in op_input_info: + missing_keys.append(required_op_input_info_key) + if len(missing_keys) > 0: + print("op: " + op_key + " " + op_info_key + " missing: " + + ",".join(missing_keys)) + is_valid = False + else: + if not op_input_info["paramType"] in param_type_valid_value: + print("op: " + op_key + " " + op_info_key + \ + " paramType not valid, valid key:[dynamic, " + "optional, required]") + is_valid = False + is_valid = check_type_format(op_input_info, is_valid, op_info_key) + if op_info_key.startswith("output"): + op_input_info = op_dict[op_info_key] + missing_keys = [] + for required_op_input_info_key in required_op_output_info_keys: + if required_op_input_info_key not in op_input_info: + missing_keys.append(required_op_input_info_key) + if len(missing_keys) > 0: + print("op: " + op_key + " " + op_info_key + " missing: " + + ",".join(missing_keys)) + is_valid = False + else: + if not op_input_info["paramType"] in param_type_valid_value: + print("op: " + op_key + " " + op_info_key + + " paramType not valid, valid key:[dynamic, " + "optional, required]") + is_valid = False + is_valid = check_type_format(op_input_info, is_valid, op_info_key) + is_valid = check_attr(op_dict, is_valid) + is_valid = check_bool_flag(op_dict, is_valid) + print("==============check valid for ops info end================\n\n") + return is_valid + + +def write_json_file(tbe_ops_info, json_file_path): + """ + Save info to json file + Parameters: + ---------------- + tbe_ops_info: ops_info + json_file_path: json file path + ---------------- + """ + json_file_real_path = os.path.realpath(json_file_path) + wr_flag = os.O_WRONLY | os.O_CREAT + wr_mode = stat.S_IWUSR | stat.S_IRUSR + with os.fdopen(os.open(json_file_real_path, wr_flag, wr_mode), 'w') as file_path: + # The owner have all rights,group only have read rights + os.chmod(json_file_real_path, stat.S_IWUSR + stat.S_IRGRP + + stat.S_IRUSR) + json.dump(tbe_ops_info, file_path, sort_keys=True, indent=4, + separators=(',', ':')) + print("Compile op info cfg successfully.") + + +def parse_ini_to_json(ini_file_paths, outfile_path): + """ + parse ini files to json file + Parameters: + ---------------- + ini_file_paths: list of ini file path + outfile_path: output file path + ---------------- + """ + tbe_ops_info = parse_ini_files(ini_file_paths) + if not check_op_info(tbe_ops_info): + print("Compile op info cfg failed.") + return False + write_json_file(tbe_ops_info, outfile_path) + return True + + +if __name__ == '__main__': + args = sys.argv + + OUTPUT_FILE_PATH = "tbe_ops_info.json" + ini_file_path_list = [] + + for arg in args: + if arg.endswith("ini"): + ini_file_path_list.append(arg) + OUTPUT_FILE_PATH = arg.replace(".ini", ".json") + if arg.endswith("json"): + OUTPUT_FILE_PATH = arg + + if len(ini_file_path_list) == 0: + ini_file_path_list.append("tbe_ops_info.ini") + + if not parse_ini_to_json(ini_file_path_list, OUTPUT_FILE_PATH): + sys.exit(1) + sys.exit(0) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/preset_parse.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/preset_parse.py new file mode 100755 index 000000000..8f1124b1d --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/preset_parse.py @@ -0,0 +1,23 @@ +import json +import sys +import os + + +def get_config_opts(file): + src_dir = os.path.abspath(os.path.dirname(file)) + opts = '' + with open(file, 'r') as fd: + config = json.load(fd) + for conf in config: + if conf == 'configurePresets': + for node in config[conf]: + macros = node.get('cacheVariables') + if macros is not None: + for key in macros: + opts += '-D{}={} '.format(key, macros[key]['value']) + opts = opts.replace('${sourceDir}', src_dir) + print(opts) + + +if __name__ == "__main__": + get_config_opts(sys.argv[1]) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_codegen.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_codegen.py new file mode 100755 index 000000000..1baa364ef --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_codegen.py @@ -0,0 +1,105 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import os +import stat +import collections +import kernel_entry as keb +from tiling_data_def_build import gen_tiling +import code_channel_infer +import const_var + +PYF_PATH = os.path.dirname(__file__) + +ReplayCodeGenParams = collections.namedtuple('ReplayCodeGenParams',\ +['op_type', 'impl', 'tiling_file', 'kernel', 'entry', 'argn', 'op_replay_batch', 'max_block_dim', 'max_shape_size']) + + +class ReplayCodeGen: + def __init__(self, replayCodeGenParams): + self.op_type = replayCodeGenParams.op_type + self.impl = replayCodeGenParams.impl + self.tiling_file = replayCodeGenParams.tiling_file + self.tiling_data_file = '' + self.kernel = replayCodeGenParams.kernel + self.entry = replayCodeGenParams.entry + self.argn = replayCodeGenParams.argn + self.batch = False + self.outdir = '' + self.data_type = 'uint8_t' + self.blknum = 32 + self.op_replay_batch = replayCodeGenParams.op_replay_batch + self.max_block_dim = replayCodeGenParams.max_block_dim + self.max_shape_size = replayCodeGenParams.max_shape_size + + def set_batch(self, is_batch): + self.batch = is_batch + + def set_outdir(self, outdir): + self.outdir = outdir + + def gen_replay(self, ops_product: str): + kerentry = os.path.join(self.outdir, self.kernel + '_entry.cce') + kerimpl = os.path.join(self.outdir, self.kernel + '_impl.cpp') + replayimpl = os.path.join(self.outdir, self.kernel + '_replay.cpp') + if self.batch: + reptmp = os.path.join(PYF_PATH, 'batch_replay_impl.temp') + else: + reptmp = os.path.join(PYF_PATH, 'replay_impl.temp') + kertmp = os.path.join(PYF_PATH, 'kernel_impl.temp') + self._gen_kentry(kerentry) + self._gen_kimpl_code(kerimpl, kertmp) + self._gen_tiling_data_header() + self._gen_replay_code(replayimpl, reptmp, ops_product) + + def _gen_tiling_data_header(self): + self.tiling_data_file = os.path.join(self.outdir, self.kernel + '_tiling_data.h') + gen_tiling(self.tiling_file, self.tiling_data_file) + + def _gen_kimpl_code(self, src, tmpfile): + with open(tmpfile, 'r') as fd: + temp = fd.read() + temp = temp.replace('__CCE_FILE__', self.impl) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(temp) + + def _gen_replay_code(self, src, tmpfile, ops_product: str): + with open(tmpfile, 'r') as fd: + temp = fd.read() + temp = temp.replace('__ARG_NUM__', str(self.argn)) + argdef = [] + kargs = [] + for i in range(0, self.argn): + argdef.append('{} *'.format(self.data_type)) + kargs.append('({} *)GetArg({})'.format(self.data_type, i)) + temp = temp.replace('__ARGS_DEF__', ', '.join(argdef)) + temp = temp.replace('__KERNEL_ARGS__', ', '.join(kargs)) + temp = temp.replace('__KERNEL_FUN__', self.entry) + core_type_infer = 'core_type' + code_channel = code_channel_infer.infer_code_channel(code_channel_infer.InfoCodeChanelParams(self.impl,\ + self.tiling_data_file, self.kernel, self.outdir, ops_product, None)) + if code_channel == code_channel_infer.CODE_VEC: + core_type_infer = '0' + elif code_channel == code_channel_infer.CODE_CUBE: + core_type_infer = '1' + temp = temp.replace('__CORE_TYPE__', core_type_infer) + # regist function + temp = temp.replace('__OPS_PRODUCT__', ops_product) + temp = temp.replace('__OPTYPE__', self.op_type) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(temp) + + def _gen_kentry(self, src): + kf = '' + pre_alloc_str = 'A' * 256 + if self.batch: + kf += keb.batch_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str), self.argn, self.data_type) + else: + kf += keb.mc_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str),\ + self.argn, self.data_type, self.blknum) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(kf) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_impl.temp b/atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_impl.temp new file mode 100755 index 000000000..1d30dd865 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_impl.temp @@ -0,0 +1,120 @@ +#include +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#include "register/op_check.h" +#define __ASCENDC_REPLAY_CODE__ +using namespace std; +using namespace optiling; +using namespace AscendCReplay; + +extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); +extern "C" int elf_append(char *elf, uint32_t elfSize, char *jit, int kernum, int blknum[], char *atext[], + int alen[], int atlen, const char* kernelname[]); + +#define KERNEL_N 1 +#define ARG_N (__ARG_NUM__) +#define MAX_L (1024 * 1024 * 100) +#define MAX_E (1024 * 1024) + +int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) +{ + // gen type 1 : direct call codes 0: load .o file + if (param.gentype < 0 || param.gentype > 1) { + printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); + return 0; + } else if (param.gentype == 1 && param.objptr == nullptr) { + printf("Error: call replay with direct call mode, but code obj addr is null\n"); + return 0; + } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { + printf("Error: call replay with object file mode, but object file path is null\n"); + return 0; + } + // core_type 0:MIX 1:CUBE 2:VEC + if (core_type < 0 || core_type > 2) { + printf("Error: call replay core type is %d !\n", core_type); + return 0; + } + g_coreType = __CORE_TYPE__; + g_taskRation = param.task_ration; + g_tilingKey = param.tiling_key; + + unsigned char *buf, *jit; + char *kernel[KERNEL_N * 32]; + int len[KERNEL_N * 32]; + int blknum[KERNEL_N]; + int max; + block_num = param.block_dim; + g_ubBase = block_num; + uint8_t *code = (uint8_t *)malloc(MAX_L); + uint8_t *pos = code; + struct timespec tp1, tp2; + + clock_gettime(CLOCK_MONOTONIC, &tp1); + if (block_num > 32) { + printf("Error: block_num > 32\n"); + return 0; + } + //__OP_FOPEN__ + for (int i = 0; i < KERNEL_N; i++) { + for (int j = 0; j < ARG_N; j++) + AddArg(j, ARG_STEP * (j + 1)); + for (block_idx = 0; block_idx < block_num; block_idx++) { + //__OP_SET_KERNEL__ + int code_idx = i * block_num + block_idx; +#ifdef FP_CEILING + SetCtrlFloatEnable(); +#else + SetCtrlFloatDisable(); +#endif + CodeInit(pos, false); + __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); + CodeEnd(); + kernel[code_idx] = (char *)pos; + len[code_idx] = CodeLen(); + pos += len[code_idx]; + printf("kernel %d core %ld code generated len %d\n", i, block_idx, len[code_idx]); + } + blknum[i] = block_num; + } + //__OP_FCLOSE__ + clock_gettime(CLOCK_MONOTONIC, &tp2); + buf = (unsigned char *)malloc(MAX_E); + int fd = open(param.entry_file, O_RDONLY); + if (fd < 0) { + printf("[error]: cannot find entry.o : %s\n", param.entry_file); + return 0; + } + uint32_t bufSize = read(fd, buf, MAX_E); + if (bufSize <= 0) { + printf("[error]: entry.o : %s is too small ! \n", param.entry_file); + } + close(fd); + jit = (unsigned char *)malloc(MAX_L); + printf("total code generated %ld\n", pos - code); + int sz = elf_append((char *)buf, bufSize, (char *)jit, KERNEL_N, blknum, kernel, len, pos - code, ¶m.kernel_name); + if (tp1.tv_sec != tp2.tv_sec) { + printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); + } else { + printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); + } + printf("new elf size %d\n", sz); + if (param.gentype == 0) { + fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); + (void)write(fd, jit, sz); + close(fd); + free(jit); + } else if (param.gentype == 1) { + *param.objptr = (char*)jit; + } + free(buf); + free(code); + return sz; +} + +REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/tiling_data_def_build.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/tiling_data_def_build.py new file mode 100755 index 000000000..4c5b099b7 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/cmake/util/tiling_data_def_build.py @@ -0,0 +1,87 @@ +#!/usr/bin/env python +# coding=utf-8 +""" +Function: +The replay funtion entry +Copyright Information: +Huawei Technologies Co., Ltd. All Rights Reserved © 2020 +""" + +import sys +import os +import stat +import re +import const_var + + +def gen_tiling(tiling_header_file: str, tiling_file_out: str): + if not os.path.exists(tiling_header_file): + print("warning: no userdef tiling header file: ", tiling_header_file) + return + print("generate tiling def header file: ", tiling_file_out) + tmp_name = os.path.splitext(os.path.basename(tiling_header_file))[0].upper() + tiling_source = '#ifndef __{}_H__\n'.format(tmp_name) + tiling_source += '#define __{}_H__\n\n'.format(tmp_name) + tiling_source += '#include \n' + tiling_source += '#include \n\n' + tiling_source += '#include "kernel_tiling/kernel_tiling.h"\n\n' + end_source = "" + pattern = re.compile(r'[(](.*)[)]', re.S) + with open(tiling_header_file, 'r') as fd: + lines = fd.readlines() + for line in lines: + line = line.strip() + if (line.startswith('BEGIN_TILING_DATA_DEF')): + tiling_source += '#pragma pack(1)\n' + tiling_source += 'struct ' + struct_def = re.findall(pattern, line)[0] + tiling_source += struct_def + ' {\n' + elif (line.startswith('TILING_DATA_FIELD_DEF_ARR')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {}[{}] = {{}};\n'.format(fds[0].strip(), fds[2].strip(), fds[1].strip()) + elif (line.startswith('TILING_DATA_FIELD_DEF_STRUCT')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {};\n'.format(fds[0].strip(), fds[1].strip()) + elif (line.startswith('TILING_DATA_FIELD_DEF')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {} = 0;\n'.format(fds[0].strip(), fds[1].strip()) + elif (line.startswith('END_TILING_DATA_DEF')): + tiling_source += '};\n' + tiling_source += '#pragma pack()\n\n' + tiling_source += '#ifdef __NPU_TILING__\n' + tiling_source += \ + 'inline [aicore] void Init{stru}(const __gm__ uint8_t* tiling, {stru}* const_data)\n'\ + .format(stru=struct_def) + tiling_source += '{\n' + tiling_source += ' const __gm__ uint32_t *src = (const __gm__ uint32_t *)tiling;\n' + tiling_source += ' uint32_t *dst = (uint32_t *)const_data;\n' + tiling_source += ' for (auto i = 0; i < sizeof({}) / 4; i++) *(dst + i) = *(src + i);\n'\ + .format(struct_def) + tiling_source += '}\n' + tiling_source += '#else\n' + tiling_source += 'inline void Init{stru}(uint8_t* tiling, {stru}* const_data)\n'.format(stru=struct_def) + tiling_source += '{\n' + tiling_source += ' uint64_t *src = (uint64_t *)tiling;\n' + tiling_source += ' uint64_t *dst = (uint64_t *)const_data;\n' + tiling_source += ' for (auto i = 0; i < sizeof({}) / 8; i++) *(dst + i) = *(src + i);\n'\ + .format(struct_def) + tiling_source += '}\n' + tiling_source += '#endif\n\n' + end_source = ''' +#define GET_TILING_DATA(tiling_data, tiling_arg) \\ +{stru} tiling_data; \\ +Init{stru}(tiling_arg, &tiling_data)\n +'''.format(stru=struct_def) + tiling_source += end_source + tiling_source += '#endif' + with os.fdopen(os.open(tiling_file_out, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(tiling_source) + + +if __name__ == '__main__': + if len(sys.argv) <= 2: + raise RuntimeError('arguments must greater than 2') + gen_tiling(sys.argv[1], sys.argv[2]) diff --git a/atb_operator/AtbEyeSample/EyeOp/framework/CMakeLists.txt b/atb_operator/AtbEyeSample/EyeOp/framework/CMakeLists.txt new file mode 100755 index 000000000..b6be9b492 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/framework/CMakeLists.txt @@ -0,0 +1,11 @@ +if(NOT EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/mindspore") + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/caffe_plugin") + add_subdirectory(caffe_plugin) + endif() + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/tf_plugin") + add_subdirectory(tf_plugin) + endif() + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/onnx_plugin") + add_subdirectory(onnx_plugin) + endif() +endif() diff --git a/atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/CMakeLists.txt b/atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/CMakeLists.txt new file mode 100755 index 000000000..a6aba5c20 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/CMakeLists.txt @@ -0,0 +1,14 @@ + +aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} plugin_srcs) +add_library(cust_tf_parsers SHARED ${plugin_srcs}) +target_compile_definitions(cust_tf_parsers PRIVATE google=ascend_private) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_tf_parsers PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_tf_parsers PRIVATE intf_pub graph) +install(TARGETS cust_tf_parsers + LIBRARY DESTINATION packages/vendors/${vendor_name}/framework/tensorflow +) diff --git a/atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/tensorflow_eye_plugin.cc b/atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/tensorflow_eye_plugin.cc new file mode 100755 index 000000000..c48076386 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/tensorflow_eye_plugin.cc @@ -0,0 +1,12 @@ +/* 版权所有 (c) 华为技术有限公司 2020-2021 + * 注册 Eye 操作信息到 GE + */ +#include "register/register.h" + +namespace domi { +// register op info to GE +REGISTER_CUSTOM_OP("Eye") + .FrameworkType(TENSORFLOW) // type: CAFFE, TENSORFLOW + .OriginOpType("Eye") // name in tf module + .ParseParamsByOperatorFn(AutoMappingByOpFn); +} // namespace domi diff --git a/atb_operator/AtbEyeSample/EyeOp/op_host/CMakeLists.txt b/atb_operator/AtbEyeSample/EyeOp/op_host/CMakeLists.txt new file mode 100755 index 000000000..40dd51cfa --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/op_host/CMakeLists.txt @@ -0,0 +1,82 @@ + +aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} ops_srcs) + +opbuild(OPS_SRC ${ops_srcs} + OUT_DIR ${ASCEND_AUTOGEN_PATH} +) + +add_library(cust_op_proto SHARED ${ops_srcs} ${ASCEND_AUTOGEN_PATH}/op_proto.cc) +target_compile_definitions(cust_op_proto PRIVATE OP_PROTO_LIB) +target_compile_options(cust_op_proto PRIVATE + -fvisibility=hidden +) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_op_proto PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_op_proto PRIVATE + intf_pub + exe_graph + register + tiling_api + -Wl,--whole-archive + rt2_registry + -Wl,--no-whole-archive +) +set_target_properties(cust_op_proto PROPERTIES OUTPUT_NAME + cust_opsproto_rt2.0 +) +add_library(cust_optiling SHARED ${ops_srcs}) +target_compile_definitions(cust_optiling PRIVATE OP_TILING_LIB) +target_compile_options(cust_optiling PRIVATE + -fvisibility=hidden +) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_optiling PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_optiling PRIVATE + intf_pub + exe_graph + register + tiling_api + -Wl,--whole-archive + rt2_registry + -Wl,--no-whole-archive +) +set_target_properties(cust_optiling PROPERTIES OUTPUT_NAME + cust_opmaster_rt2.0 +) + +file(GLOB aclnn_src ${ASCEND_AUTOGEN_PATH}/aclnn_*.cpp) +file(GLOB aclnn_inc ${ASCEND_AUTOGEN_PATH}/aclnn_*.h) +add_library(cust_opapi SHARED ${aclnn_src}) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_opapi PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_opapi PRIVATE intf_pub ascendcl nnopbase) + +add_custom_target(optiling_compat ALL + COMMAND ln -sf lib/linux/${CMAKE_SYSTEM_PROCESSOR}/$ + ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so +) + +install(TARGETS cust_op_proto + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_proto/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) +install(FILES ${ASCEND_AUTOGEN_PATH}/op_proto.h + DESTINATION packages/vendors/${vendor_name}/op_proto/inc) +install(TARGETS cust_optiling + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) +install(FILES ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so + DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling) +install(TARGETS cust_opapi + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_api/lib) +install(FILES ${aclnn_inc} + DESTINATION packages/vendors/${vendor_name}/op_api/include) diff --git a/atb_operator/AtbEyeSample/EyeOp/op_host/eye.cpp b/atb_operator/AtbEyeSample/EyeOp/op_host/eye.cpp new file mode 100755 index 000000000..cf72d4d17 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/op_host/eye.cpp @@ -0,0 +1,173 @@ +/** +* @file eye.cpp +* +* Copyright (C) 2023. Huawei Technologies Co., Ltd. All rights reserved. +* +* This program is distributed in the hope that it will be useful, +* but WITHOUT ANY WARRANTY; without even the implied warranty of +* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. +*/ + +#include "eye_tiling.h" +#include "register/op_def_registry.h" +#include "tiling/platform/platform_ascendc.h" + +namespace optiling { + const uint32_t BLOCK_SIZE = 32; + constexpr int BATCH_SIZE_DIM_THRESHOLD = 2; + constexpr uint32_t DATATYPE_SIZE_2 = 2; + constexpr uint32_t DATATYPE_SIZE_4 = 4; + constexpr uint32_t TYPE_KEY_0 = 0; + constexpr uint32_t TYPE_KEY_1 = 1; + constexpr uint32_t DATA_NUM = 4; + static ge::graphStatus TilingFunc(gert::TilingContext* context) { + EyeTilingData tiling; + uint32_t sizeofdatatype; + uint32_t totalLengthAligned; + + // 1. 获取平台信息 + uint64_t ub_size; + auto ascendcPlatform = platform_ascendc::PlatformAscendC(context->GetPlatformInfo()); + ascendcPlatform.GetCoreMemSize(platform_ascendc::CoreMemType::UB, ub_size); + + // 2. 获取数据信息 + uint32_t dataNum; + uint32_t totalLength = context->GetInputTensor(0)->GetShapeSize(); + uint32_t tanhxLength = context->GetOptionalInputTensor(0)->GetShapeSize(); + const int64_t* pnum_rows = context->GetAttrs()->GetInt(0); + const int64_t* pnum_columns = context->GetAttrs()->GetInt(1); + const int64_t* pdtype = context->GetAttrs()->GetInt(2); + auto *pbatch_shapecv = context->GetAttrs()->GetListInt(0); + int32_t num_rows = *pnum_rows; + int32_t num_columns = *pnum_columns; + int32_t dtype = *pdtype; + const int64_t *pbatch_shape = pbatch_shapecv->GetData(); + int32_t batchShapeSize = pbatch_shapecv->GetSize(); + int32_t batch_shape_list[10]; + if(num_columns == 0){ + num_columns = num_rows; + } + + int32_t mark = 0; + int32_t batchNum = 1; + int32_t batchSize = 0; + auto shape = context->GetInputTensor(0)->GetOriginShape(); + const uint32_t dimNum = shape.GetDimNum(); + if(dimNum > BATCH_SIZE_DIM_THRESHOLD){ + mark = 1; + batchSize = num_rows * num_columns; + batchNum = totalLength / batchSize; + } + + tiling.set_num_columns(num_columns); + tiling.set_num_rows(num_rows); + tiling.set_batch_shape(batch_shape_list); + tiling.set_dtype(dtype); + tiling.set_mark(mark); + tiling.set_batchNum(batchNum); + tiling.set_batchSize(batchSize); + + auto dt = context->GetInputTensor(0)->GetDataType(); + uint32_t typeKey; + if (dt == 1) { + sizeofdatatype = DATATYPE_SIZE_2; + dataNum = DATA_NUM; + typeKey = TYPE_KEY_0; + }else{ + sizeofdatatype = DATATYPE_SIZE_4; + dataNum = DATA_NUM; + typeKey = TYPE_KEY_1; + } + + tiling.set_typeKey(typeKey); + + // 3. 填满UB大小 + uint32_t ub_block_num = ub_size / BLOCK_SIZE / dataNum - 256; + if (ub_block_num % 2 != 0) { + ub_block_num = ub_block_num - 1; + } + + // 4. 输入向量满足32字节对齐 + uint32_t ALIGN_NUM = BLOCK_SIZE / sizeofdatatype; + if (totalLength % ALIGN_NUM != 0) { //不对齐,先32位对齐 + totalLengthAligned = ((totalLength + ALIGN_NUM - 1) / ALIGN_NUM) * ALIGN_NUM; + } else { + totalLengthAligned = totalLength; + } + + // 5. Tiling参数计算 + uint32_t tile_num, block_dim = 1; + context->SetBlockDim(block_dim); + uint32_t blockLength = 0; + uint32_t tileLength = 0; + uint32_t lasttileLength = 0; + blockLength = totalLengthAligned / block_dim; + tile_num = blockLength / ALIGN_NUM / ub_block_num; + + if (tile_num == 0) { // 不足一个ub的情况 + tile_num = 1; + tileLength = ((blockLength / ALIGN_NUM) + 1) / 2 * 2 * ALIGN_NUM; + lasttileLength = tileLength; + } else if((blockLength / ALIGN_NUM) % ub_block_num == 0){ // 核内能均分 + tileLength = ub_block_num * ALIGN_NUM; + lasttileLength = tileLength; + }else{ // 核内不能均分 + tile_num = tile_num + 1; // 加一个小包的数量 + tileLength = ub_block_num * ALIGN_NUM; + lasttileLength = blockLength - (tile_num - 1) * tileLength; + lasttileLength = ((lasttileLength / ALIGN_NUM) + 1) / 2 * 2 * ALIGN_NUM; + } + + tiling.set_blockLength(blockLength); + tiling.set_tileNum(tile_num); + tiling.set_tileLength(tileLength); + tiling.set_lasttileLength(lasttileLength); + tiling.SaveToBuffer(context->GetRawTilingData()->GetData(), + context->GetRawTilingData()->GetCapacity()); + context->GetRawTilingData()->SetDataSize(tiling.GetDataSize()); + size_t* currentWorkspace = context->GetWorkspaceSizes(1); + currentWorkspace[0] = 0; + return ge::GRAPH_SUCCESS; + } +} + +namespace ge { + static ge::graphStatus InferShape(gert::InferShapeContext* context) { + const gert::Shape* x1_shape = context->GetInputShape(0); + gert::Shape* y_shape = context->GetOutputShape(0); + *y_shape = *x1_shape; + return GRAPH_SUCCESS; + } +} + + +namespace ops { +class Eye : public OpDef { +public: + explicit Eye(const char* name) : OpDef(name) { + this->Input("y") + .ParamType(REQUIRED) + .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32}) + .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) + .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); + this->Output("y") + .ParamType(REQUIRED) + .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32}) + .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) + .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); + this->Attr("num_rows").Int(); + this->Attr("num_columns").AttrType(OPTIONAL).Int(0); + this->Attr("batch_shape").AttrType(OPTIONAL).ListInt({}); + this->Attr("dtype").AttrType(OPTIONAL).Int(0); + + this->SetInferShape(ge::InferShape); + + this->AICore() + .SetTiling(optiling::TilingFunc); + this->AICore().AddConfig("ascend310b") + .AddConfig("ascend910b"); + } +}; + +OP_ADD(Eye); +} diff --git a/atb_operator/AtbEyeSample/EyeOp/op_host/eye_tiling.h b/atb_operator/AtbEyeSample/EyeOp/op_host/eye_tiling.h new file mode 100755 index 000000000..6aadcaa35 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/op_host/eye_tiling.h @@ -0,0 +1,31 @@ +/** +* @file eye_tiling.h +* +* Copyright (C) 2023. Huawei Technologies Co., Ltd. All rights reserved. +* +* This program is distributed in the hope that it will be useful, +* but WITHOUT ANY WARRANTY; without even the implied warranty of +* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. +*/ + +#include "register/tilingdata_base.h" + +namespace optiling { +BEGIN_TILING_DATA_DEF(EyeTilingData) + TILING_DATA_FIELD_DEF(uint32_t, typeKey); + TILING_DATA_FIELD_DEF(uint32_t, blockLength); + TILING_DATA_FIELD_DEF(uint32_t, tileNum); + TILING_DATA_FIELD_DEF(uint32_t, tileLength); + TILING_DATA_FIELD_DEF(uint32_t, lasttileLength); + TILING_DATA_FIELD_DEF(int32_t, num_rows); + TILING_DATA_FIELD_DEF(int32_t, num_columns); + TILING_DATA_FIELD_DEF(int32_t, dtype); + TILING_DATA_FIELD_DEF(int32_t, mark); + TILING_DATA_FIELD_DEF(int32_t, batchNum); + TILING_DATA_FIELD_DEF(int32_t, batchSize); + TILING_DATA_FIELD_DEF_ARR(int32_t, 10, batch_shape); + +END_TILING_DATA_DEF; + +REGISTER_TILING_DATA_CLASS(Eye, EyeTilingData) +} diff --git a/atb_operator/AtbEyeSample/EyeOp/op_kernel/CMakeLists.txt b/atb_operator/AtbEyeSample/EyeOp/op_kernel/CMakeLists.txt new file mode 100755 index 000000000..8c94a952d --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/op_kernel/CMakeLists.txt @@ -0,0 +1,68 @@ +# set custom compile options +if ("${CMAKE_BUILD_TYPE}x" STREQUAL "Debugx") + add_ops_compile_options(ALL OPTIONS -g -O0) +endif() + +foreach(compute_unit ${ASCEND_COMPUTE_UNIT}) + + # generate aic-${compute_unit}-ops-info.json + add_ops_info_target(TARGET ops_info_gen_${compute_unit} + OUTPUT ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core/${compute_unit}/aic-${compute_unit}-ops-info.json + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/config/${compute_unit} + ) + + # generate ascendc impl py once + if (NOT TARGET ascendc_impl_gen) + add_ops_impl_target(TARGET ascendc_impl_gen + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl + ) + endif() + + # dynamic shape binary compile + if (${ENABLE_BINARY_PACKAGE} AND NOT ${ENABLE_CROSS_COMPILE}) + add_bin_compile_target(TARGET ascendc_bin_${compute_unit} + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} + ADP_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/dynamic + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/binary/${compute_unit} + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/kernel + COMPUTE_UNIT ${compute_unit} + ) + add_dependencies(ascendc_bin_${compute_unit} ascendc_impl_gen) + endif() + + if (${ENABLE_CROSS_COMPILE} AND ${ENABLE_BINARY_PACKAGE}) + add_cross_compile_target( + TARGET bin_${compute_unit} + OUT_DIR ${CMAKE_CURRENT_SOURCE_DIR}/../kernel + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/ + ) + endif() +endforeach() + +# generate npu_supported_ops.json +add_npu_support_target(TARGET npu_supported_ops + OPS_INFO_DIR ${ASCEND_AUTOGEN_PATH} + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core + INSTALL_DIR packages/vendors/${vendor_name}/framework/${ASCEND_FRAMEWORK_TYPE} +) + +if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) + add_subdirectory(testcases) +endif() + +# install kernel file +if (${ENABLE_SOURCE_PACKAGE}) + file(GLOB KERNEL_FILES + ${CMAKE_CURRENT_SOURCE_DIR}/*.cpp + ${CMAKE_CURRENT_SOURCE_DIR}/*.h + ${CMAKE_CURRENT_SOURCE_DIR}/*.py + ) + install(FILES ${KERNEL_FILES} + DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl/dynamic + ) +endif() diff --git a/atb_operator/AtbEyeSample/EyeOp/op_kernel/eye.cpp b/atb_operator/AtbEyeSample/EyeOp/op_kernel/eye.cpp new file mode 100755 index 000000000..9aef97978 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/op_kernel/eye.cpp @@ -0,0 +1,91 @@ +/** +* @file eye.cpp +* +* Copyright (C) 2023. Huawei Technologies Co., Ltd. All rights reserved. +* +* This program is distributed in the hope that it will be useful, +* but WITHOUT ANY WARRANTY; without even the implied warranty of +* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. +*/ + +#include "kernel_operator.h" + +using namespace AscendC; +constexpr int32_t BUFFER_NUM = 2; + +class KernelEye { +public: + __aicore__ inline KernelEye() {} + __aicore__ inline void Init(GM_ADDR y, GM_ADDR y_ref, + uint32_t typeKey, uint32_t blockLength, uint32_t tileNum, + uint32_t tileLength, uint32_t lasttileLength, int32_t num_columns, + int32_t num_rows, int32_t *batch_shape, int32_t dtype, + int32_t mark, int32_t batchSize, int32_t batchNum) { + + this->batchSize = batchSize; + this->batchNum = batchNum; + this->mark = mark; + this->num_rows = num_rows; + this->num_columns = num_columns; + this->batch_shape = batch_shape; + this->dtype = dtype; + this->blockLength = blockLength; + this->tileNum = tileNum; + this->tileLength = tileLength / BUFFER_NUM; + this->lasttileLength = lasttileLength / BUFFER_NUM; + this->typeKey = typeKey; + + yGm.SetGlobalBuffer((__gm__ DTYPE_Y*)y_ref + this->blockLength * GetBlockIdx(), this->blockLength); + } + __aicore__ inline void Process() { + int32_t index, t; + if(mark == 0){ + for(int32_t i = 0; i < num_rows; i++){ + index = i * num_columns + i; + if(i < num_columns){ + yGm.SetValue(index, 1); + } + } + }else{ + for(int32_t i = 0; i < batchNum; i++){ + for(int32_t j = 0; j < num_rows; j++){ + if(j < num_columns){ + t = j * num_columns + j; + index = i * batchSize + t; + yGm.SetValue(index, 1); + } + } + } + } + } + +private: + TPipe pipe; + TBuf tmpBuf1, tmpBuf2, tmpBuf3; + TQue inQueueIN; + TQue outQueueOUT; + GlobalTensor yGm; + uint32_t blockLength; + uint32_t tileNum; + uint32_t tileLength; + uint32_t lasttileLength; + uint32_t typeKey; + int32_t *batch_shape; + int32_t dtype; + int32_t num_columns; + int32_t num_rows; + int32_t mark; + int32_t batchSize, batchNum; +}; + +extern "C" __global__ __aicore__ void eye(GM_ADDR y, GM_ADDR y_ref, GM_ADDR workspace, GM_ADDR tiling) { + GET_TILING_DATA(tiling_data, tiling); + + KernelEye op; + + op.Init(y, y_ref, tiling_data.typeKey, tiling_data.blockLength, + tiling_data.tileNum, tiling_data.tileLength, tiling_data.lasttileLength, + tiling_data.num_columns, tiling_data.num_rows, tiling_data.batch_shape, + tiling_data.dtype, tiling_data.mark, tiling_data.batchSize, tiling_data.batchNum); + op.Process(); +} \ No newline at end of file diff --git a/atb_operator/AtbEyeSample/EyeOp/scripts/install.sh b/atb_operator/AtbEyeSample/EyeOp/scripts/install.sh new file mode 100755 index 000000000..8468c5a25 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/scripts/install.sh @@ -0,0 +1,318 @@ +#!/bin/bash +vendor_name=customize +targetdir=/usr/local/Ascend/opp +target_custom=0 + +sourcedir=$PWD/packages +vendordir=vendors/$vendor_name + +QUIET="y" + +while true +do + case $1 in + --quiet) + QUIET="y" + shift + ;; + --install-path=*) + INSTALL_PATH=$(echo $1 | cut -d"=" -f2-) + INSTALL_PATH=${INSTALL_PATH%*/} + shift + ;; + --*) + shift + ;; + *) + break + ;; + esac +done + +log() { + cur_date=`date +"%Y-%m-%d %H:%M:%S"` + echo "[runtime] [$cur_date] "$1 +} + +if [ -n "${INSTALL_PATH}" ]; then + if [[ ! "${INSTALL_PATH}" = /* ]]; then + log "[ERROR] use absolute path for --install-path argument" + exit 1 + fi + if [ ! -d ${INSTALL_PATH} ]; then + mkdir ${INSTALL_PATH} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${INSTALL_PATH} failed" + exit 1 + fi + fi + targetdir=${INSTALL_PATH} +elif [ -n "${ASCEND_CUSTOM_OPP_PATH}" ]; then + if [ ! -d ${ASCEND_CUSTOM_OPP_PATH} ]; then + mkdir -p ${ASCEND_CUSTOM_OPP_PATH} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${ASCEND_CUSTOM_OPP_PATH} failed" + fi + fi + targetdir=${ASCEND_CUSTOM_OPP_PATH} +else + if [ "x${ASCEND_OPP_PATH}" == "x" ]; then + log "[ERROR] env ASCEND_OPP_PATH no exist" + exit 1 + fi + targetdir="${ASCEND_OPP_PATH}" +fi + +if [ ! -d $targetdir ];then + log "[ERROR] $targetdir no exist" + exit 1 +fi + +upgrade() +{ + if [ ! -d ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 files" + return 0 + fi + + if [ ! -d ${targetdir}/$vendordir/$1 ];then + log "[INFO] create ${targetdir}/$vendordir/$1." + mkdir -p ${targetdir}/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/$1 failed" + return 1 + fi + else + has_same_file=-1 + for file_a in ${sourcedir}/$vendordir/$1/*; do + file_b=${file_a##*/}; + if [ "ls ${targetdir}/$vendordir/$1" = "" ]; then + log "[INFO] ${targetdir}/$vendordir/$1 is empty !!" + return 1 + fi + grep -q $file_b <<<`ls ${targetdir}/$vendordir/$1`; + if [[ $? -eq 0 ]]; then + echo -n "${file_b} " + has_same_file=0 + fi + done + if [ 0 -eq $has_same_file ]; then + if test $QUIET = "n"; then + echo "[INFO]: has old version in ${targetdir}/$vendordir/$1, \ + you want to Overlay Installation , please enter:[o]; \ + or replace directory installation , please enter: [r]; \ + or not install , please enter:[n]." + + while true + do + read orn + if [ "$orn" = n ]; then + return 0 + elif [ "$orn" = m ]; then + break; + elif [ "$0rn" = r ]; then + [ -n "${targetdir}/$vendordir/$1/" ] && rm -rf "${targetdir}/$vendordir/$1"/* + break; + else + echo "[ERROR] input error, please input again!" + fi + done + fi + fi + log "[INFO] replace or merge old ops $1 files .g....." + fi + + log "copy new ops $1 files ......" + if [ -d ${targetdir}/$vendordir/$1/ ]; then + chmod -R +w "$targetdir/$vendordir/$1/" >/dev/null 2>&1 + fi + cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 files failed" + return 1 + fi + + return 0 +} +upgrade_proto() +{ + if [ ! -f ${sourcedir}/$vendordir/custom.proto ]; then + log "[INFO] no need to upgrade custom.proto files" + return 0 + fi + if [ ! -d ${targetdir}/$vendordir/framework/caffe ];then + log "[INFO] create ${targetdir}/$vendordir/framework/caffe." + mkdir -p ${targetdir}/$vendordir/framework/caffe + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/framework/caffe failed" + return 1 + fi + else + if [ -f ${targetdir}/$vendordir/framework/caffe/custom.proto ]; then + # 有老版本,判断是否要覆盖式安装 + if test $QUIET = "n"; then + echo "[INFO] ${targetdir}/$vendordir/framework/caffe has old version"\ + "custom.proto file. Do you want to replace? [y/n] " + + while true + do + read yn + if [ "$yn" = n ]; then + return 0 + elif [ "$yn" = y ]; then + break; + else + echo "[ERROR] input error, please input again!" + fi + done + fi + fi + log "[INFO] replace old caffe.proto files ......" + fi + chmod -R +w "$targetdir/$vendordir/framework/caffe/" >/dev/null 2>&1 + cp -rf ${sourcedir}/$vendordir/custom.proto ${targetdir}/$vendordir/framework/caffe/ + if [ $? -ne 0 ];then + log "[ERROR] copy new custom.proto failed" + return 1 + fi + log "[INFO] copy custom.proto success" + + return 0 +} + +upgrade_file() +{ + if [ ! -e ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 file" + return 0 + fi + + log "copy new $1 files ......" + cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 file failed" + return 1 + fi + + return 0 +} + +delete_optiling_file() +{ + if [ ! -d ${targetdir}/vendors ];then + log "[INFO] $1 not exist, no need to uninstall" + return 0 + fi + sys_info=$(uname -m) + if [ ! -d ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/lib/linux/${sys_info} ];then + rm -rf ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/liboptiling.so + fi + return 0 +} + +log "[INFO] copy uninstall sh success" + +if [ ! -d ${targetdir}/vendors ];then + log "[INFO] create ${targetdir}/vendors." + mkdir -p ${targetdir}/vendors + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/vendors failed" + return 1 + fi +fi +chmod u+w ${targetdir}/vendors + +echo "[ops_custom]upgrade framework" +upgrade framework +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op proto" +upgrade op_proto +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade version.info" +upgrade_file version.info +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op impl" +delete_optiling_file op_impl +upgrade op_impl +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op api" +upgrade op_api +if [ $? -ne 0 ];then + exit 1 +fi + +upgrade_proto +if [ $? -ne 0 ];then + exit 1 +fi + +# set the set_env.bash +if [ -n "${INSTALL_PATH}" ] && [ -d ${INSTALL_PATH} ]; then + _ASCEND_CUSTOM_OPP_PATH=${targetdir}/${vendordir} + bin_path="${_ASCEND_CUSTOM_OPP_PATH}/bin" + set_env_variable="#!/bin/bash\nexport ASCEND_CUSTOM_OPP_PATH=${_ASCEND_CUSTOM_OPP_PATH}:\${ASCEND_CUSTOM_OPP_PATH}" + if [ ! -d ${bin_path} ]; then + mkdir -p ${bin_path} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${bin_path} failed" + exit 1 + fi + fi + echo -e ${set_env_variable} > ${bin_path}/set_env.bash + if [ $? -ne 0 ]; then + log "[ERROR] write ASCEND_CUSTOM_OPP_PATH to set_env.bash failed" + exit 1 + else + log "[INFO] using requirements: when custom module install finished or before you run the custom module, \ + execute the command [ source ${bin_path}/set_env.bash ] to set the environment path" + fi +else + config_file=${targetdir}/vendors/config.ini + if [ ! -f ${config_file} ]; then + touch ${config_file} + chmod 640 ${config_file} + echo "load_priority=$vendor_name" > ${config_file} + if [ $? -ne 0 ];then + echo "echo load_priority failed" + exit 1 + fi + else + found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" + found_vendor=$(echo $found_vendors | sed "s/$vendor_name//g" | tr ',' ' ') + vendor=$(echo $found_vendor | tr -s ' ' ',') + if [ "$vendor" != "" ]; then + sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" + fi + fi +fi + +chmod u-w ${targetdir}/vendors + +if [ -d ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/ ]; then + chmod -R 440 ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/* >/dev/null 2>&1 +fi +if [ -f ${targetdir}/ascend_install.info ]; then + chmod -R 440 ${targetdir}/ascend_install.info +fi +if [ -f ${targetdir}/scene.info ]; then + chmod -R 440 ${targetdir}/scene.info +fi +if [ -f ${targetdir}/version.info ]; then + chmod -R 440 ${targetdir}/version.info +fi + +echo "SUCCESS" +exit 0 + diff --git a/atb_operator/AtbEyeSample/EyeOp/scripts/upgrade.sh b/atb_operator/AtbEyeSample/EyeOp/scripts/upgrade.sh new file mode 100755 index 000000000..e09173485 --- /dev/null +++ b/atb_operator/AtbEyeSample/EyeOp/scripts/upgrade.sh @@ -0,0 +1,151 @@ +#!/bin/bash +vendor_name=customize +targetdir=/usr/local/Ascend/opp +target_custom=0 + +sourcedir=$PWD/packages +vendordir=vendors/$vendor_name + +log() { + cur_date=`date +"%Y-%m-%d %H:%M:%S"` + echo "[runtime] [$cur_date] "$1 +} + +if [[ "x${ASCEND_OPP_PATH}" == "x" ]];then + log "[ERROR] env ASCEND_OPP_PATH no exist" + exit 1 +fi + +targetdir=${ASCEND_OPP_PATH} + +if [ ! -d $targetdir ];then + log "[ERROR] $targetdir no exist" + exit 1 +fi + +upgrade() +{ + if [ ! -d ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 files" + return 0 + fi + + if [ ! -d ${targetdir}/$vendordir/$1 ];then + log "[INFO] create ${targetdir}/$vendordir/$1." + mkdir -p ${targetdir}/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/$1 failed" + return 1 + fi + else + vendor_installed_dir=$(ls "$targetdir/vendors" 2> /dev/null) + for i in $vendor_installed_dir;do + vendor_installed_file=$(ls "$vendor_installed_dir/$vendor_name/$i" 2> /dev/null) + if [ "$i" = "$vendor_name" ] && [ "$vendor_installed_file" != "" ]; then + echo "[INFO]: $vendor_name custom opp package has been installed on the path $vendor_installed_dir, \ + you want to Overlay Installation , please enter:[o]; \ + or replace directory installation , please enter: [r]; \ + or not install , please enter:[n]." + fi + while true + do + read mrn + if [ "$mrn" = m ]; then + break + elif [ "$mrn" = r ]; then + [ -n "$vendor_installed_file"] && rm -rf "$vendor_installed_file" + break + elif [ "$mrn" = n ]; then + return 0 + else + echo "[WARNING]: Input error, please input m or r or n to choose!" + fi + done + done + log "[INFO] replace old ops $1 files ......" + fi + + log "copy new ops $1 files ......" + cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 files failed" + return 1 + fi + + return 0 +} + +upgrade_file() +{ + if [ ! -e ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 file" + return 0 + fi + + log "copy new $1 files ......" + cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 file failed" + return 1 + fi + + return 0 +} + +log "[INFO] copy uninstall sh success" + +echo "[ops_custom]upgrade framework" +upgrade framework +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op proto" +upgrade op_proto +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op impl" +upgrade op_impl +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op api" +upgrade op_api +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade version.info" +upgrade_file version.info +if [ $? -ne 0 ];then + exit 1 +fi + +config_file=${targetdir}/vendors/config.ini +found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" +found_vendor=$(echo $found_vendors | sed "s/$vendor_name//g" | tr ',' ' ') +vendor=$(echo $found_vendor | tr -s ' ' ',') +if [ "$vendor" != "" ]; then + sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" +fi + +changemode() +{ + if [ -d ${targetdir} ];then + chmod -R 550 ${targetdir}>/dev/null 2>&1 + fi + + return 0 +} +echo "[ops_custom]changemode..." +#changemode +if [ $? -ne 0 ];then + exit 1 +fi + +echo "SUCCESS" +exit 0 + diff --git a/atb_operator/AtbTrilSample/TrilOp/CMakeLists.txt b/atb_operator/AtbTrilSample/TrilOp/CMakeLists.txt new file mode 100755 index 000000000..584132d80 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/CMakeLists.txt @@ -0,0 +1,69 @@ +cmake_minimum_required(VERSION 3.16.0) +project(opp) +if(ENABLE_CROSS_COMPILE) + if(${CMAKE_SYSTEM_PROCESSOR} STREQUAL x86_64) + set(CROSS_COMPILE_PLATFORM aarch64) + else() + set(CROSS_COMPILE_PLATFORM x86_64) + endif() + set(PLATFORM ${CMAKE_SYSTEM_PROCESSOR}) + set(CMAKE_COMPILE_COMPILER_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/linux/${CROSS_COMPILE_PLATFORM}/) + set(CMAKE_COMPILE_RUNTIME_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/${CROSS_COMPILE_PLATFORM}/) + set(CMAKE_SYSTEM_PROCESSOR ${CROSS_COMPILE_PLATFORM}) + set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) + set(CMAKE_CXX_COMPILER ${CMAKE_CROSS_PLATFORM_COMPILER}) +else() + set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) +endif() + +include(cmake/config.cmake) +include(cmake/func.cmake) +include(cmake/intf.cmake) + +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/framework) + add_subdirectory(framework) +endif() +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_host) + add_subdirectory(op_host) +endif() +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_kernel) + add_subdirectory(op_kernel) +endif() +if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) + add_subdirectory(testcases) +endif() + +# modify vendor_name in install.sh and upgrade.sh +add_custom_command(OUTPUT ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh + COMMAND mkdir -p ${CMAKE_BINARY_DIR}/scripts + COMMAND cp -r ${CMAKE_SOURCE_DIR}/scripts/* ${CMAKE_BINARY_DIR}/scripts/ + COMMAND sed -i "s/vendor_name=customize/vendor_name=${vendor_name}/g" ${CMAKE_BINARY_DIR}/scripts/* +) +add_custom_target(modify_vendor ALL DEPENDS ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh) +install(DIRECTORY ${CMAKE_BINARY_DIR}/scripts/ DESTINATION . FILE_PERMISSIONS OWNER_EXECUTE OWNER_READ GROUP_READ) + +install(FILES ${CMAKE_SOURCE_DIR}/custom.proto DESTINATION packages OPTIONAL) + +get_system_info(SYSTEM_INFO) + +# gen version.info +add_custom_target(gen_version_info ALL + COMMAND bash ${CMAKE_CURRENT_SOURCE_DIR}/cmake/util/gen_version_info.sh ${ASCEND_CANN_PACKAGE_PATH} ${CMAKE_CURRENT_BINARY_DIR} +) + +install(FILES ${CMAKE_CURRENT_BINARY_DIR}/version.info + DESTINATION packages/vendors/${vendor_name}/) + +# CPack config +set(CPACK_PACKAGE_NAME ${CMAKE_PROJECT_NAME}) +set(CPACK_PACKAGE_VERSION ${CMAKE_PROJECT_VERSION}) +set(CPACK_PACKAGE_DESCRIPTION "CPack opp project") +set(CPACK_PACKAGE_DESCRIPTION_SUMMARY "CPack opp project") +set(CPACK_PACKAGE_DIRECTORY ${CMAKE_INSTALL_PREFIX}) +set(CPACK_PACKAGE_FILE_NAME "custom_opp_${SYSTEM_INFO}.run") +set(CPACK_GENERATOR External) +set(CPACK_CMAKE_GENERATOR "Unix Makefiles") +set(CPACK_EXTERNAL_ENABLE_STAGING TRUE) +set(CPACK_EXTERNAL_PACKAGE_SCRIPT ${CMAKE_SOURCE_DIR}/cmake/makeself.cmake) +set(CPACK_EXTERNAL_BUILT_PACKAGES ${CPACK_PACKAGE_DIRECTORY}/_CPack_Packages/Linux/External/${CPACK_PACKAGE_FILE_NAME}/${CPACK_PACKAGE_FILE_NAME}) +include(CPack) diff --git a/atb_operator/AtbTrilSample/TrilOp/CMakePresets.json b/atb_operator/AtbTrilSample/TrilOp/CMakePresets.json new file mode 100755 index 000000000..d32bf6d9a --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/CMakePresets.json @@ -0,0 +1,63 @@ +{ + "version": 1, + "cmakeMinimumRequired": { + "major": 3, + "minor": 19, + "patch": 0 + }, + "configurePresets": [ + { + "name": "default", + "displayName": "Default Config", + "description": "Default build using Unix Makefiles generator", + "generator": "Unix Makefiles", + "binaryDir": "${sourceDir}/build_out", + "cacheVariables": { + "CMAKE_BUILD_TYPE": { + "type": "STRING", + "value": "Release" + }, + "ENABLE_SOURCE_PACKAGE": { + "type": "BOOL", + "value": "True" + }, + "ENABLE_BINARY_PACKAGE": { + "type": "BOOL", + "value": "True" + }, + "ASCEND_COMPUTE_UNIT": { + "type": "STRING", + "value": "ascend310b;ascend910b" + }, + "ENABLE_TEST": { + "type": "BOOL", + "value": "True" + }, + "vendor_name": { + "type": "STRING", + "value": "customize" + }, + "ASCEND_CANN_PACKAGE_PATH": { + "type": "PATH", + "value": "/usr/local/Ascend/ascend-toolkit/latest" + }, + "ASCEND_PYTHON_EXECUTABLE": { + "type": "STRING", + "value": "python3" + }, + "CMAKE_INSTALL_PREFIX": { + "type": "PATH", + "value": "${sourceDir}/build_out" + }, + "ENABLE_CROSS_COMPILE": { + "type": "BOOL", + "value": "False" + }, + "CMAKE_CROSS_PLATFORM_COMPILER": { + "type": "PATH", + "value": "/usr/bin/aarch64-linux-gnu-g++" + } + } + } + ] +} \ No newline at end of file diff --git a/atb_operator/AtbTrilSample/TrilOp/README.md b/atb_operator/AtbTrilSample/TrilOp/README.md new file mode 100644 index 000000000..a8addc946 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/README.md @@ -0,0 +1,193 @@ +# 概述 +本样例基于TrilOp算子工程,介绍了单算子工程及aclnn接口说明。 + +## 算子工程介绍 +### 算子描述 +Tril算子是PyTorch中的一种常见矩阵构造函数。Tril函数默认情况下返回一个矩阵主对角线以下的下三角矩阵,其它元素全部为0。主对角线的偏移由可选参数diagonal决定,其缺省值为0。diagonal为正值时,主对角线向上偏移。当输入是一个多维张量时,其最后两个维度构成矩阵,Tril以迭代的方式处理多维张量中的每个矩阵,最终返回对应的下三角矩阵构成的多维张量。返回的多维张量与输入张量维度保持一致。 + +### 算子规格描述 + + + + + + + +
算子类型(OpType)Tril
算子输入nameshapedata typeformat默认值
x-float32, float16ND\
算子输出y-float32, float16ND\
attr属性diagonal\int\0
核函数名tril
+ + + +### 算子工程文件介绍 +其中,算子工程目录TrilOp包含算子实现的模板文件、编译脚本等,如下所示: +``` +├── TrilOp //Add自定义算子工程 +│ ├── cmake +│ ├── framework // 算子插件实现文件目录,单算子模型文件的生成不依赖算子适配插件,无需关注 +│ ├── op_host // host侧实现文件 +│ ├── op_kernel // kernel侧实现文件 +│ ├── scripts // 自定义算子工程打包相关脚本所在目录 +│ ├── build.sh // 编译入口脚本 +│ ├── CMakeLists.txt // 算子工程的CMakeLists.txt +│ └── CMakePresets.json // 编译配置项 +``` + +### 编译运行样例算子 +针对自定义算子工程,编译运行包含如下步骤: +- 编译自定义算子工程生成算子安装包; +- 安装自定义算子到算子库中; +- 调用执行自定义算子; + +详细操作如下所示。 +#### 1. 获取源码包 +编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 +#### 2. 编译算子工程 + 编译自定义算子工程,构建生成自定义算子包。 + + - 执行如下命令,切换到算子工程TrilOp目录。 + + ```bash + cd ${git_clone_path}/samples/atb_operator/AtbTrilSample/TrilOp + ``` + + - 修改CMakePresets.json中ASCEND_CANN_PACKAGE_PATH为CANN软件包安装后的实际路径。 + + + ```json + { + …… + "configurePresets": [ + { + …… + "ASCEND_CANN_PACKAGE_PATH": { + "type": "PATH", + "value": "/usr/local/Ascend/ascend-toolkit/latest" //请替换为CANN软件包安装后的实际路径。eg:/home/HwHiAiUser/Ascend/ascend-toolkit/latest + }, + …… + } + ] + } + ``` + - 在算子工程TrilOp目录下执行如下命令,进行算子工程编译。 + + ```bash + ./build.sh + ``` +编译成功后,会在当前目录下创建build_out目录,并在build_out目录下生成自定义算子安装包custom_opp_\_\.run,例如“custom_opp_ubuntu_x86_64.run”。 + +备注:如果要使用dump调试功能,需要移除op_host内和CMakeLists.txt内的Atlas 200/500 A2 推理产品的配置。 + +#### 3. 部署算子包 + +执行如下命令,在自定义算子安装包所在路径下,安装自定义算子包。 + ```bash + cd build_out + ./custom_opp__.run + ``` +命令执行成功后,自定义算子包中的相关文件将部署至当前环境的OPP算子库的vendors/customize目录中。 + +#### 4. 配置环境变量 + + 请根据当前环境上CANN开发套件包的[安装方式](https://hiascend.com/document/redirect/CannCommunityInstSoftware),选择对应配置环境变量的命令。 + - 默认路径,root用户安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=/usr/local/Ascend/ascend-toolkit/latest + ``` + - 默认路径,非root用户安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=$HOME/Ascend/ascend-toolkit/latest + ``` + - 指定路径install_path,安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=${install_path}/ascend-toolkit/latest + ``` + +## aclnn接口说明 + +自定义算子编译部署后,会自动生成单算子API,可以直接在应用程序中调用。算子API的形式一般定义为“两段式接口”,形如: + ```cpp + aclnnStatus aclnnTrilGetWorkspaceSize(const aclTensor *x, int64_t diagonal, const aclTensor *out, uint64_t *workspaceSize,aclOpExecutor **executor); + aclnnStatus aclnnTril(void *workspace, int64_t workspaceSize, aclOpExecutor **executor, aclrtStream stream); + ``` +其中aclnnTrilGetWorkspaceSize为第一段接口,主要用于计算本次API调用计算过程中需要多少的workspace内存。获取到本次API计算需要的workspace大小之后,按照workspaceSize大小申请Device侧内存,然后调用第二段接口aclnnTriu执行计算。具体参考[AscendCL单算子调用](https://hiascend.com/document/redirect/CannCommunityAscendCInVorkSingleOp)>单算子API执行 章节。 + +### 功能描述 +* 算子功能:完成加法计算。 +* 计算公式: +``` +z = x + y +``` + +#### 参数说明 +##### aclnnTrilGetWorkspaceSize: + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
接口参数描述
aclnnTrilGetWorkspaceSizex输入x的Tensor,支持flaot/half类型,ND排布格式
diagonal可选属性,对角线的位置,支持int64类型数值,默认值为0
out输出z的Tensor,支持flaot/half类型,ND排布格式
workspaceSize返回需要在Device侧申请的workspace大小
executor返回op执行器,包含了算子计算流程
+ + +| 返回值 | 描述 | +| ---- | ---- | +| ACL_SUCCESS | 成功 | +| ACL_ERROR | 失败 | + +##### aclnnTril + + + + + + + + + + + + + + + + + + + + + + + +
接口参数描述
aclnnTrilworkspace输入Device侧申请的workspace内存
workspaceSizeworkspace大小
executor输入op执行器,包含了算子计算流程
stream输入执行流
+ +| 返回值 | 描述 | +| ---- | ---- | +| ACL_SUCCESS | 成功 | +| ACL_ERROR | 失败 | + + +## 更新说明 +| 时间 | 更新事项 | +|----|------| +| 2025/03/27 | 新增README | \ No newline at end of file diff --git a/atb_operator/AtbTrilSample/TrilOp/build.sh b/atb_operator/AtbTrilSample/TrilOp/build.sh new file mode 100755 index 000000000..decae7c43 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/build.sh @@ -0,0 +1,74 @@ +#!/bin/bash +script_path=$(realpath $(dirname $0)) + + +mkdir -p build_out +rm -rf build_out/* +cd build_out + +opts=$(python3 $script_path/cmake/util/preset_parse.py $script_path/CMakePresets.json) +ENABLE_CROSS="-DENABLE_CROSS_COMPILE=True" +ENABLE_BINARY="-DENABLE_BINARY_PACKAGE=True" +cmake_version=$(cmake --version | grep "cmake version" | awk '{print $3}') + +cmake_run_package() +{ + target=$1 + cmake --build . --target $target -j16 + if [ $? -ne 0 ]; then exit 1; fi + + if [ $target = "package" ]; then + if test -d ./op_kernel/binary ; then + ./cust*.run + if [ $? -ne 0 ]; then exit 1; fi + cmake --build . --target binary -j16 + if [ $? -ne 0 ]; then exit 1; fi + cmake --build . --target $target -j16 + fi + fi +} + +if [[ $opts =~ $ENABLE_CROSS ]] && [[ $opts =~ $ENABLE_BINARY ]] +then + target=package + if [ "$1"x != ""x ]; then target=$1; fi + if [ "$cmake_version" \< "3.19.0" ] ; then + cmake .. $opts -DENABLE_CROSS_COMPILE=0 + else + cmake .. --preset=default -DENABLE_CROSS_COMPILE=0 + fi + cmake_run_package $target + cp -r kernel ../ + rm -rf * + if [ "$cmake_version" \< "3.19.0" ] ; then + cmake .. $opts + else + cmake .. --preset=default + fi + + cmake --build . --target $target -j16 + if [ $? -ne 0 ]; then exit 1; fi + if [ $target = "package" ]; then + if test -d ./op_kernel/binary ; then + ./cust*.run + fi + fi + rm -rf ../kernel + +else + target=package + if [ "$1"x != ""x ]; then target=$1; fi + if [ "$cmake_version" \< "3.19.0" ] ; then + cmake .. $opts + else + cmake .. --preset=default + fi + cmake_run_package $target +fi + + +# for debug +# cd build_out +# make +# cpack +# verbose append -v \ No newline at end of file diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/config.cmake b/atb_operator/AtbTrilSample/TrilOp/cmake/config.cmake new file mode 100755 index 000000000..886119daa --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/config.cmake @@ -0,0 +1,25 @@ + +set(CMAKE_CXX_FLAGS_DEBUG "") +set(CMAKE_CXX_FLAGS_RELEASE "") + +if (NOT DEFINED vendor_name) + set(vendor_name customize CACHE STRING "") +endif() +if (NOT DEFINED ASCEND_CANN_PACKAGE_PATH) + set(ASCEND_CANN_PACKAGE_PATH /usr/local/Ascend/latest CACHE PATH "") +endif() +if (NOT DEFINED ASCEND_PYTHON_EXECUTABLE) + set(ASCEND_PYTHON_EXECUTABLE python3 CACHE STRING "") +endif() +if (NOT DEFINED ASCEND_COMPUTE_UNIT) + message(FATAL_ERROR "ASCEND_COMPUTE_UNIT not set in CMakePreset.json ! +") +endif() +set(ASCEND_TENSOR_COMPILER_PATH ${ASCEND_CANN_PACKAGE_PATH}/compiler) +set(ASCEND_CCEC_COMPILER_PATH ${ASCEND_TENSOR_COMPILER_PATH}/ccec_compiler/bin) +set(ASCEND_AUTOGEN_PATH ${CMAKE_BINARY_DIR}/autogen) +set(ASCEND_FRAMEWORK_TYPE tensorflow) +file(MAKE_DIRECTORY ${ASCEND_AUTOGEN_PATH}) +set(CUSTOM_COMPILE_OPTIONS "custom_compile_options.ini") +execute_process(COMMAND rm -rf ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} + COMMAND touch ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS}) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/func.cmake b/atb_operator/AtbTrilSample/TrilOp/cmake/func.cmake new file mode 100755 index 000000000..4179dfd25 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/func.cmake @@ -0,0 +1,192 @@ + +function(get_system_info SYSTEM_INFO) + if (UNIX) + execute_process(COMMAND grep -i ^id= /etc/os-release OUTPUT_VARIABLE TEMP) + string(REGEX REPLACE "\n|id=|ID=|\"" "" SYSTEM_NAME ${TEMP}) + set(${SYSTEM_INFO} ${SYSTEM_NAME}_${CMAKE_SYSTEM_PROCESSOR} PARENT_SCOPE) + elseif (WIN32) + message(STATUS "System is Windows. Only for pre-build.") + else () + message(FATAL_ERROR "${CMAKE_SYSTEM_NAME} not support.") + endif () +endfunction() + +function(opbuild) + message(STATUS "Opbuild generating sources") + cmake_parse_arguments(OPBUILD "" "OUT_DIR;PROJECT_NAME;ACCESS_PREFIX" "OPS_SRC" ${ARGN}) + execute_process(COMMAND ${CMAKE_COMPILE} -g -fPIC -shared -std=c++11 ${OPBUILD_OPS_SRC} -D_GLIBCXX_USE_CXX11_ABI=0 + -I ${ASCEND_CANN_PACKAGE_PATH}/include -L ${ASCEND_CANN_PACKAGE_PATH}/lib64 -lexe_graph -lregister -ltiling_api + -o ${OPBUILD_OUT_DIR}/libascend_all_ops.so + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("build ops lib info: ${EXEC_INFO}") + message("build ops lib error: ${EXEC_ERROR}") + message(FATAL_ERROR "opbuild run failed!") + endif() + set(proj_env "") + set(prefix_env "") + if (NOT "${OPBUILD_PROJECT_NAME}x" STREQUAL "x") + set(proj_env "OPS_PROJECT_NAME=${OPBUILD_PROJECT_NAME}") + endif() + if (NOT "${OPBUILD_ACCESS_PREFIX}x" STREQUAL "x") + set(prefix_env "OPS_DIRECT_ACCESS_PREFIX=${OPBUILD_ACCESS_PREFIX}") + endif() + execute_process(COMMAND ${proj_env} ${prefix_env} ${ASCEND_CANN_PACKAGE_PATH}/toolkit/tools/opbuild/op_build + ${OPBUILD_OUT_DIR}/libascend_all_ops.so ${OPBUILD_OUT_DIR} + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("opbuild ops info: ${EXEC_INFO}") + message("opbuild ops error: ${EXEC_ERROR}") + endif() + message(STATUS "Opbuild generating sources - done") +endfunction() + +function(add_ops_info_target) + cmake_parse_arguments(OPINFO "" "TARGET;OPS_INFO;OUTPUT;INSTALL_DIR" "" ${ARGN}) + get_filename_component(opinfo_file_path "${OPINFO_OUTPUT}" DIRECTORY) + add_custom_command(OUTPUT ${OPINFO_OUTPUT} + COMMAND mkdir -p ${opinfo_file_path} + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/parse_ini_to_json.py + ${OPINFO_OPS_INFO} ${OPINFO_OUTPUT} + ) + add_custom_target(${OPINFO_TARGET} ALL + DEPENDS ${OPINFO_OUTPUT} + ) + install(FILES ${OPINFO_OUTPUT} + DESTINATION ${OPINFO_INSTALL_DIR} + ) +endfunction() + +function(add_ops_compile_options OP_TYPE) + cmake_parse_arguments(OP_COMPILE "" "OP_TYPE" "COMPUTE_UNIT;OPTIONS" ${ARGN}) + file(APPEND ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} + "${OP_TYPE},${OP_COMPILE_COMPUTE_UNIT},${OP_COMPILE_OPTIONS}\n") +endfunction() + +function(add_ops_impl_target) + cmake_parse_arguments(OPIMPL "" "TARGET;OPS_INFO;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) + add_custom_command(OUTPUT ${OPIMPL_OUT_DIR}/.impl_timestamp + COMMAND mkdir -m 700 -p ${OPIMPL_OUT_DIR}/dynamic + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py + ${OPIMPL_OPS_INFO} + \"${OPIMPL_OPS_BATCH}\" \"${OPIMPL_OPS_ITERATE}\" + ${OPIMPL_IMPL_DIR} + ${OPIMPL_OUT_DIR}/dynamic + ${ASCEND_AUTOGEN_PATH} + + COMMAND rm -rf ${OPIMPL_OUT_DIR}/.impl_timestamp + COMMAND touch ${OPIMPL_OUT_DIR}/.impl_timestamp + DEPENDS ${OPIMPL_OPS_INFO} + ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py + ) + add_custom_target(${OPIMPL_TARGET} ALL + DEPENDS ${OPIMPL_OUT_DIR}/.impl_timestamp) + if (${ENABLE_SOURCE_PACKAGE}) + install(DIRECTORY ${OPIMPL_OUT_DIR}/dynamic + DESTINATION ${OPIMPL_INSTALL_DIR} + ) + endif() +endfunction() + +function(add_npu_support_target) + cmake_parse_arguments(NPUSUP "" "TARGET;OPS_INFO_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) + get_filename_component(npu_sup_file_path "${NPUSUP_OUT_DIR}" DIRECTORY) + add_custom_command(OUTPUT ${NPUSUP_OUT_DIR}/npu_supported_ops.json + COMMAND mkdir -p ${NPUSUP_OUT_DIR} + COMMAND ${CMAKE_SOURCE_DIR}/cmake/util/gen_ops_filter.sh + ${NPUSUP_OPS_INFO_DIR} + ${NPUSUP_OUT_DIR} + ) + add_custom_target(npu_supported_ops ALL + DEPENDS ${NPUSUP_OUT_DIR}/npu_supported_ops.json + ) + install(FILES ${NPUSUP_OUT_DIR}/npu_supported_ops.json + DESTINATION ${NPUSUP_INSTALL_DIR} + ) +endfunction() + +function(add_bin_compile_target) + cmake_parse_arguments(BINCMP "" "TARGET;OPS_INFO;COMPUTE_UNIT;IMPL_DIR;ADP_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/src) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/gen) + execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_bin_param_build.py + ${BINCMP_OPS_INFO} ${BINCMP_OUT_DIR}/gen ${BINCMP_COMPUTE_UNIT} + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("ops binary compile scripts gen info: ${EXEC_INFO}") + message("ops binary compile scripts gen error: ${EXEC_ERROR}") + message(FATAL_ERROR "ops binary compile scripts gen failed!") + endif() + if (NOT TARGET binary) + add_custom_target(binary) + endif() + add_custom_target(${BINCMP_TARGET} + COMMAND cp -r ${BINCMP_IMPL_DIR}/*.* ${BINCMP_OUT_DIR}/src + ) + add_custom_target(${BINCMP_TARGET}_gen_ops_config + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/insert_simplified_keys.py -p ${BINCMP_OUT_DIR}/bin + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_ops_config.py -p ${BINCMP_OUT_DIR}/bin + -s ${BINCMP_COMPUTE_UNIT} + ) + add_dependencies(binary ${BINCMP_TARGET}_gen_ops_config) + file(GLOB bin_scripts ${BINCMP_OUT_DIR}/gen/*.sh) + foreach(bin_script ${bin_scripts}) + get_filename_component(bin_file ${bin_script} NAME_WE) + string(REPLACE "-" ";" bin_sep ${bin_file}) + list(GET bin_sep 0 op_type) + list(GET bin_sep 1 op_file) + list(GET bin_sep 2 op_index) + if (NOT TARGET ${BINCMP_TARGET}_${op_file}_copy) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file}) + add_custom_target(${BINCMP_TARGET}_${op_file}_copy + COMMAND cp ${BINCMP_ADP_DIR}/${op_file}.py ${BINCMP_OUT_DIR}/src/${op_type}.py + ) + install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} + DESTINATION ${BINCMP_INSTALL_DIR}/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) + install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json + DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT}/ OPTIONAL + ) + endif() + add_custom_target(${BINCMP_TARGET}_${op_file}_${op_index} + COMMAND export HI_PYTHON=${ASCEND_PYTHON_EXECUTABLE} && bash ${bin_script} ${BINCMP_OUT_DIR}/src/${op_type}.py ${BINCMP_OUT_DIR}/bin/${op_file} && echo $(MAKE) + WORKING_DIRECTORY ${BINCMP_OUT_DIR} + ) + add_dependencies(${BINCMP_TARGET}_${op_file}_${op_index} ${BINCMP_TARGET} ${BINCMP_TARGET}_${op_file}_copy) + add_dependencies(${BINCMP_TARGET}_gen_ops_config ${BINCMP_TARGET}_${op_file}_${op_index}) + endforeach() + install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json + DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) + + install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} + DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/../build_out/kernel/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) + install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json + DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/../build_out/kernel/config/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) + install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json + DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/../build_out/kernel/config/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) + +endfunction() + +function(add_cross_compile_target) + cmake_parse_arguments(CROSSMP "" "TARGET;OUT_DIR;INSTALL_DIR" "" ${ARGN}) + add_custom_target(${CROSSMP_TARGET} ALL + DEPENDS ${CROSSMP_OUT_DIR} + ) + install(DIRECTORY ${CROSSMP_OUT_DIR} + DESTINATION ${CROSSMP_INSTALL_DIR} + ) +endfunction() diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/intf.cmake b/atb_operator/AtbTrilSample/TrilOp/cmake/intf.cmake new file mode 100755 index 000000000..2f362c396 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/intf.cmake @@ -0,0 +1,26 @@ + +add_library(intf_pub INTERFACE) +target_compile_options(intf_pub INTERFACE + -fPIC + -fvisibility=hidden + -fvisibility-inlines-hidden + $<$:-O2> + $<$:-O0 -g> + $<$:-std=c++11> + $<$,$>:-ftrapv -fstack-check> + $<$:-pthread -Wfloat-equal -Wshadow -Wformat=2 -Wno-deprecated -Wextra> + $,-fstack-protector-strong,-fstack-protector-all> +) +target_compile_definitions(intf_pub INTERFACE + _GLIBCXX_USE_CXX11_ABI=0 + $<$:_FORTIFY_SOURCE=2> +) +target_include_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/include) +target_link_options(intf_pub INTERFACE + $<$,EXECUTABLE>:-pie> + $<$:-s> + -Wl,-z,relro + -Wl,-z,now + -Wl,-z,noexecstack +) +target_link_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/lib64) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/makeself.cmake b/atb_operator/AtbTrilSample/TrilOp/cmake/makeself.cmake new file mode 100755 index 000000000..48c565bfb --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/makeself.cmake @@ -0,0 +1,17 @@ +execute_process(COMMAND chmod +x ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh) +execute_process(COMMAND ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh + --header ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself-header.sh + --help-header ./help.info + --gzip --complevel 4 --nomd5 --sha256 + ./ ${CPACK_PACKAGE_FILE_NAME} "version:1.0" ./install.sh + WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} + RESULT_VARIABLE EXEC_RESULT + ERROR_VARIABLE EXEC_ERROR +) +if (NOT "${EXEC_RESULT}x" STREQUAL "0x") + message(FATAL_ERROR "CPack Command error: ${EXEC_RESULT}\n${EXEC_ERROR}") +endif() +execute_process(COMMAND cp ${CPACK_EXTERNAL_BUILT_PACKAGES} ${CPACK_PACKAGE_DIRECTORY}/ + COMMAND echo "Copy ${CPACK_EXTERNAL_BUILT_PACKAGES} to ${CPACK_PACKAGE_DIRECTORY}/" + WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} +) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/__init__.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/__init__.py new file mode 100755 index 000000000..c4ddc893a --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/__init__.py @@ -0,0 +1,8 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- + +import sys +import os + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) +sys.path.append(PYF_PATH) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_bin_param_build.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_bin_param_build.py new file mode 100755 index 000000000..63b15c900 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_bin_param_build.py @@ -0,0 +1,127 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import json +import hashlib +import const_var +import opdesc_parser + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + + +class BinParamBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + super().__init__(op_type) + self.soc = '' + self.out_path = '' + + def set_soc_version(self: any, soc: str): + self.soc = soc + + def set_out_path(self: any, out_path: str): + self.out_path = out_path + + def gen_input_json(self: any): + key_map = {} + count = len(self.input_dtype[0].split(',')) + for i in range(0, count): + inputs = [] + outputs = [] + attrs = [] + op_node = {} + for idx in range(0, len(self.input_name)): + idtypes = self.input_dtype[idx].split(',') + ifmts = self.input_fmt[idx].split(',') + itype = self.input_type[idx] + para = {} + para['name'] = self.input_name[idx][:-5] + para['index'] = idx + para['dtype'] = idtypes[i] + para['format'] = ifmts[i] + para['paramType'] = itype + para['shape'] = [-2] + if itype == 'dynamic': + inputs.append([para]) + else: + inputs.append(para) + for idx in range(0, len(self.output_name)): + odtypes = self.output_dtype[idx].split(',') + ofmts = self.output_fmt[idx].split(',') + otype = self.output_type[idx] + para = {} + para['name'] = self.output_name[idx][:-5] + para['index'] = idx + para['dtype'] = odtypes[i] + para['format'] = ofmts[i] + para['paramType'] = otype + para['shape'] = [-2] + if otype == 'dynamic': + outputs.append([para]) + else: + outputs.append(para) + for attr in self.attr_list: + att = {} + att['name'] = attr + atype = self.attr_val.get(attr).get('type').lower() + att['dtype'] = atype + att['value'] = const_var.ATTR_DEF_VAL.get(atype) + attrs.append(att) + op_node['bin_filename'] = '' + op_node['inputs'] = inputs + op_node['outputs'] = outputs + if len(attrs) > 0: + op_node['attrs'] = attrs + param = {} + param['op_type'] = self.op_type + param['op_list'] = [op_node] + objstr = json.dumps(param, indent=' ') + md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() + while key_map.get(md5sum) is not None: + objstr += '1' + md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() + key_map[md5sum] = md5sum + bin_file = self.op_type + '_' + md5sum + op_node['bin_filename'] = bin_file + param_file = os.path.join(self.out_path, bin_file + '_param.json') + param_file = os.path.realpath(param_file) + with os.fdopen(os.open(param_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + json.dump(param, fd, indent=' ') + self._write_buld_cmd(param_file, bin_file, i) + + def _write_buld_cmd(self: any, param_file: str, bin_file: str, index: int): + hard_soc = const_var.SOC_MAP_EXT.get(self.soc) + if not hard_soc: + hard_soc = self.soc.capitalize() + name_com = [self.op_type, self.op_file, str(index)] + compile_file = os.path.join(self.out_path, '-'.join(name_com) + '.sh') + compile_file = os.path.realpath(compile_file) + with os.fdopen(os.open(compile_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + fd.write('#!/bin/bash\n') + fd.write('echo "[{}] Generating {} ..."\n'.format(hard_soc, bin_file)) + cmd = const_var.BIN_CMD.format(fun=self.op_intf, soc=hard_soc, param=param_file, impl='""') + fd.write(cmd) + chk = const_var.CHK_CMD.format(res_file=bin_file + '.json') + fd.write(chk) + chk = const_var.CHK_CMD.format(res_file=bin_file + '.o') + fd.write(chk) + fd.write('echo "[{}] Generating {} Done"\n'.format(hard_soc, bin_file)) + + +def gen_bin_param_file(cfgfile: str, out_dir: str, soc: str): + op_descs = opdesc_parser.get_op_desc(cfgfile, [], [], BinParamBuilder, None) + for op_desc in op_descs: + op_desc.set_soc_version(soc) + op_desc.set_out_path(out_dir) + op_desc.gen_input_json() + + +if __name__ == '__main__': + if len(sys.argv) <= 3: + raise RuntimeError('arguments must greater than 3') + gen_bin_param_file(sys.argv[1], sys.argv[2], sys.argv[3]) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_impl_build.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_impl_build.py new file mode 100755 index 000000000..e8cf2b9ab --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_impl_build.py @@ -0,0 +1,496 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import re +import stat +import opdesc_parser +import const_var + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + +IMPL_HEAD = ''' +import os, sys +import ctypes +import json +import shutil +from tbe.common.platform import get_soc_spec +from tbe.common.utils import para_check +from tbe.tikcpp import compile_op, replay_op, check_op_cap, generalize_op_params, get_code_channel, OpInfo +from tbe.common.buildcfg import get_default_build_config +from impl.util.platform_adapter import tbe_register +from tbe.common.buildcfg import get_current_build_config +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + +DTYPE_MAP = {"float32": ["DT_FLOAT", "float"], + "float16": ["DT_FLOAT16", "half"], + "int8": ["DT_INT8", "int8_t"], + "int16": ["DT_INT16", "int16_t"], + "int32": ["DT_INT32", "int32_t"], + "int64": ["DT_INT64", "int64_t"], + "uint1": ["DT_UINT1", "uint8_t"], + "uint8": ["DT_UINT8", "uint8_t"], + "uint16": ["DT_UINT16", "uint16_t"], + "uint32": ["DT_UINT32", "uint32_t"], + "uint64": ["DT_UINT64", "uint64_t"], + "bool": ["DT_BOOL", "bool"], + "double": ["DT_DOUBLE", "double"], + "dual": ["DT_DUAL", "unknown"], + "dual_sub_int8": ["DT_DUAL_SUB_INT8", "unknown"], + "dual_sub_uint8": ["DT_DUAL_SUB_UINT8", "unknown"], + "string": ["DT_STRING", "unknown"], + "complex64": ["DT_COMPLEX64", "unknown"], + "complex128": ["DT_COMPLEX128", "unknown"], + "qint8": ["DT_QINT8", "unknown"], + "qint16": ["DT_QINT16", "unknown"], + "qint32": ["DT_QINT32", "unknown"], + "quint8": ["DT_QUINT8", "unknown"], + "quint16": ["DT_QUINT16", "unknown"], + "resource": ["DT_RESOURCE", "unknown"], + "string_ref": ["DT_STRING_REF", "unknown"], + "int4": ["DT_INT4", "int4b_t"], + "bfloat16": ["DT_BF16", "bfloat16_t"]} + +def add_dtype_fmt_option_single(x, x_n, is_ref: bool = False): + options = [] + x_fmt = x.get("format") + x_dtype = x.get("dtype") + x_n_in_kernel = x_n + '_REF' if is_ref else x_n + options.append("-DDTYPE_{n}={t}".format(n=x_n_in_kernel, t=DTYPE_MAP.get(x_dtype)[1])) + options.append("-DORIG_DTYPE_{n}={ot}".format(n=x_n_in_kernel, ot=DTYPE_MAP.get(x_dtype)[0])) + options.append("-DFORMAT_{n}=FORMAT_{f}".format(n=x_n_in_kernel, f=x_fmt)) + return options + +def get_dtype_fmt_options(__inputs__, __outputs__): + options = [] + unique_param_name_set = set() + for x in __inputs__: + if x is None: + continue + x_n = x.get("param_name")[:-5].upper() + unique_param_name_set.add(x_n) + options += add_dtype_fmt_option_single(x, x_n) + + for x in __outputs__: + if x is None: + continue + x_n = x.get("param_name")[:-5].upper() + if x_n in unique_param_name_set: + options += add_dtype_fmt_option_single(x, x_n, True) + else: + options += add_dtype_fmt_option_single(x, x_n) + return options + +def load_dso(so_path): + try: + ctypes.CDLL(so_path) + except OSError as error : + print(error) + raise RuntimeError("cannot open %s" %(so_path)) + else: + print("load so succ ", so_path) + +def get_shortsoc_compile_option(compile_option_list: list, shortsoc:str): + compile_options = [] + if shortsoc in compile_option_list: + compile_options = compile_option_list[shortsoc] + elif '__ALLSOC__' in compile_option_list: + compile_options = compile_option_list['__ALLSOC__'] + return compile_options + +def get_kernel_source(src_file, dir_snake, dir_ex): + src_ex = os.path.join(PYF_PATH, "..", "ascendc", dir_ex, src_file) + if os.path.exists(src_ex): + return src_ex + src = os.path.join(PYF_PATH, "..", "ascendc", dir_snake, src_file) + if os.path.exists(src): + return src + src = os.path.join(PYF_PATH, src_file) + if os.path.exists(src): + return src + return src_ex + +''' + +IMPL_API = ''' +@tbe_register.register_operator("{}", trans_bool_to_s8=False) +@para_check.check_op_params({}) +def {}({}, kernel_name="{}", impl_mode=""): + if get_current_build_config("enable_op_prebuild"): + return + __inputs__, __outputs__, __attrs__ = _build_args({}) + options = get_dtype_fmt_options(__inputs__, __outputs__) + options += ["-x", "cce"] + bisheng = shutil.which("bisheng") + if bisheng != None: + bisheng_path = os.path.dirname(bisheng) + tikcpp_path = os.path.realpath(os.path.join(bisheng_path, "..", "..", "tikcpp")) + else: + tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") + options.append("-I" + tikcpp_path) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) + options.append("-I" + os.path.join(PYF_PATH, "..", "ascendc", "common")) + if impl_mode == "high_performance": + options.append("-DHIGH_PERFORMANCE=1") + elif impl_mode == "high_precision": + options.append("-DHIGH_PRECISION=1") + if get_default_build_config("enable_deterministic_mode") == 1: + options.append("-DDETEMINISTIC_MODE=1") + + custom_compile_options = {}, + custom_all_compile_options = {}, + soc_version = get_soc_spec("SOC_VERSION") + soc_short = get_soc_spec("SHORT_SOC_VERSION").lower() + custom_compile_options_soc = get_shortsoc_compile_option(custom_compile_options[0], soc_short) + custom_all_compile_options_soc = get_shortsoc_compile_option(custom_all_compile_options[0], soc_short) + options += custom_all_compile_options_soc + options += custom_compile_options_soc + + origin_func_name = "{}" + ascendc_src_dir_ex = "{}" + ascendc_src_dir = "{}" + ascendc_src_file = "{}" + src = get_kernel_source(ascendc_src_file, ascendc_src_dir, ascendc_src_dir_ex) +''' + +REPLAY_OP_API = ''' + print("start replay Acend C Operator {}, kernel name is {}") + tikreplay_codegen_path = tikcpp_path + "/tikreplaylib/lib" + tikreplay_stub_path = tikcpp_path + "/tikreplaylib/lib/" + soc_version + print("start load libtikreplaylib_codegen.so and libtikreplaylib_stub.so") + codegen_so_path = tikreplay_codegen_path + "/libtikreplaylib_codegen.so" + replaystub_so_path = tikreplay_stub_path + "/libtikreplaylib_stub.so" + if PYF_PATH.endswith("dynamic"): + op_replay_path = os.path.join(PYF_PATH, "..", "..", "op_replay") + else: + op_replay_path = os.path.join(PYF_PATH, "..", "op_replay") + replayapi_so_path = os.path.join(op_replay_path, "libreplay_{}_" + soc_short + ".so") + load_dso(codegen_so_path) + load_dso(replaystub_so_path) + load_dso(replayapi_so_path) + op_type = "{}" + entry_obj = os.path.join(op_replay_path, "{}_entry_" + soc_short + ".o") + code_channel = get_code_channel(src, kernel_name, op_type, options) + op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ + attrs = __attrs__, impl_mode = impl_mode) + res, msg = replay_op(op_info, entry_obj, code_channel, src, options) + if not res: + print("call replay op failed for %s and get into call compile op" %(msg)) + compile_op(src, origin_func_name, op_info, options, code_channel, '{}') +''' + +COMPILE_OP_API = ''' + print("start compile Ascend C operator {}. kernel name is " + kernel_name) + op_type = "{}" + code_channel = get_code_channel(src, kernel_name, op_type, options) + op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ + attrs = __attrs__, impl_mode = impl_mode, origin_inputs=[{}], origin_outputs = [{}]) + compile_op(src, origin_func_name, op_info, options, code_channel, '{}') +''' + +SUP_API = ''' +def {}({}, impl_mode=""): + __inputs__, __outputs__, __attrs__ = _build_args({}) + ret_str = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) + ret_dict = json.loads(ret_str) + err_code = ret_dict.get("ret_code") + sup = "Unknown" + reason = "Unknown reason" + if err_code is not None: + if err_code is 0: + sup = "True" + reason = "" + elif err_code is 1: + sup = "False" + reason = ret_dict.get("reason") + else: + sup = "Unknown" + reason = ret_dict.get("reason") + return sup, reason +''' +CAP_API = ''' +def {}({}, impl_mode=""): + __inputs__, __outputs__, __attrs__ = _build_args({}) + result = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) + return result.decode("utf-8") +''' +GLZ_API = ''' +@tbe_register.register_param_generalization("{}") +def {}_generalization({}, generalize_config=None): + __inputs__, __outputs__, __attrs__ = _build_args({}) + ret_str = generalize_op_params("{}", __inputs__, __outputs__, __attrs__, generalize_config) + return [json.loads(ret_str)] +''' + +ATTR_DEFAULT = {'bool': 'False', 'int': '0', 'float': '0.0', 'list_int': '[]', + 'list_float': '[]', 'list_bool': '[]', 'list_list_int': '[[]]', 'str': ''} + + +def _get_snake_str(s, i): + if s[i - 1] != '_': + if not s[i - 1].isupper(): + return "_" + elif s[i - 1].isupper() and (i + 1) < len(s) and s[i + 1].islower(): + return "_" + return "" + return "" + + +def optype_snake(origin_str): + temp_str = origin_str[0].lower() + origin_str[1:] + new_str = re.sub(r'([A-Z])', r'_\1', temp_str).lower() + return new_str + + +def optype_snake_ex(s): + snake_case = "" + for i, c in enumerate(s): + if i == 0: + snake_case += c.lower() + elif c.isupper(): + snake_case += _get_snake_str(s, i) + snake_case += c.lower() + else: + snake_case += c + return snake_case + + +class AdpBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + self.argsname = [] + self.argsdefv = [] + self.op_compile_option:str = '{}' + super().__init__(op_type) + + + def write_adapt(self: any, impl_path, path: str, op_compile_option_all:list = None): + self._build_paradefault() + if impl_path != "": + src_file = os.path.join(impl_path, self.op_file + '.cpp') + if not os.path.exists(src_file): + return + out_path = os.path.abspath(path) + if self.dynamic_shape and not out_path.endswith('dynamic'): + out_path = os.path.join(path, 'dynamic') + os.makedirs(out_path, mode=0o700, exist_ok=True) + adpfile = os.path.join(out_path, self.op_file + '.py') + self._gen_op_compile_option(op_compile_option_all) + with os.fdopen(os.open(adpfile, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + self._write_head(fd) + self._write_argparse(fd) + self._write_impl(fd) + if self.op_chk_support: + self._write_cap('check_supported', fd) + self._write_cap('get_op_support_info', fd) + if self.op_fmt_sel: + self._write_cap('op_select_format', fd) + self._write_cap('get_op_specific_info', fd) + if self.op_range_limit == 'limited' or self.op_range_limit == 'dynamic': + self._write_glz(fd) + + + def _gen_op_compile_option(self:any, op_compile_option_all:list =None): + if op_compile_option_all is not None: + if self.op_type in op_compile_option_all: + self.op_compile_option = op_compile_option_all[self.op_type] + elif "__all__" in op_compile_option_all: + self.op_compile_option = op_compile_option_all["__all__"] + + + def _ip_argpack(self: any, default: bool = True) -> list: + args = [] + for i in range(len(self.input_name)): + arg = self.input_name[i] + if default and self.argsdefv[i] is not None: + arg += '=' + self.argsdefv[i] + args.append(arg) + return args + + def _op_argpack(self: any, default: bool = True) -> list: + args = [] + argidx = len(self.input_name) + for i in range(len(self.output_name)): + arg = self.output_name[i] + if default and self.argsdefv[i + argidx] is not None: + arg += '=' + self.argsdefv[i + argidx] + args.append(arg) + return args + + def _attr_argpack(self: any, default: bool = True) -> list: + args = [] + argidx = len(self.input_name) + len(self.output_name) + for i in range(len(self.attr_list)): + att = self.attr_list[i] + arg = att + if default and self.argsdefv[i + argidx] is not None: + if self.attr_val.get(att).get('type') == 'str': + arg += '="' + self.argsdefv[i + argidx] + '"' + elif self.attr_val.get(att).get('type') == 'bool': + arg += '=' + self.argsdefv[i + argidx].capitalize() + else: + arg += '=' + self.argsdefv[i + argidx] + args.append(arg) + return args + + def _build_paralist(self: any, default: bool = True) -> str: + args = [] + args.extend(self._ip_argpack(default)) + args.extend(self._op_argpack(default)) + args.extend(self._attr_argpack(default)) + return ', '.join(args) + + def _io_parachk(self: any, types: list, type_name: str) -> list: + chk = [] + for iot in types: + if iot == 'optional': + ptype = 'OPTION' + else: + ptype = iot.upper() + chk.append('para_check.{}_{}'.format(ptype, type_name)) + return chk + + def _attr_parachk(self: any) -> list: + chk = [] + for att in self.attr_list: + if self.attr_val.get(att).get('paramType') == 'optional': + pt = 'OPTION' + else: + pt = self.attr_val.get(att).get('paramType').upper() + att_type = self.attr_val.get(att).get('type').upper() + chk.append('para_check.{}_ATTR_{}'.format(pt, att_type)) + return chk + + def _build_parachk(self: any) -> str: + chk = [] + chk.extend(self._io_parachk(self.input_type, 'INPUT')) + chk.extend(self._io_parachk(self.output_type, 'OUTPUT')) + chk.extend(self._attr_parachk()) + chk.append('para_check.KERNEL_NAME') + return ', '.join(chk) + + def _build_paradefault(self: any): + optional = False + argtypes = [] + argtypes.extend(self.input_type) + argtypes.extend(self.output_type) + for atype in argtypes: + if atype == 'optional': + optional = True + if optional: + self.argsdefv.append('None') + else: + self.argsdefv.append(None) + for attr in self.attr_list: + atype = self.attr_val.get(attr).get('paramType') + if atype == 'optional': + optional = True + attrval = self.attr_val.get(attr).get('defaultValue') + if attrval is not None: + optional = True + if type == "bool": + attrval = attrval.capitalize() + elif type == "str": + attrval = "\"" + attrval + "\"" + self.argsdefv.append(attrval) + continue + if optional: + self.argsdefv.append(ATTR_DEFAULT.get(self.attr_val.get(attr).get('type'))) + else: + self.argsdefv.append(None) + + def _write_head(self: any, fd: object): + fd.write(IMPL_HEAD) + + def _write_argparse(self: any, fd: object): + args = self._build_paralist(False) + fd.write('def _build_args({}):\n'.format(args)) + fd.write(' __inputs__ = []\n') + fd.write(' for arg in [{}]:\n'.format(', '.join(self.input_name))) + fd.write(' if arg != None:\n') + fd.write(' if isinstance(arg, (list, tuple)):\n') + fd.write(' if len(arg) == 0:\n') + fd.write(' continue\n') + fd.write(' __inputs__.append(arg[0])\n') + fd.write(' else:\n') + fd.write(' __inputs__.append(arg)\n') + fd.write(' __outputs__ = []\n') + fd.write(' for arg in [{}]:\n'.format(', '.join(self.output_name))) + fd.write(' if arg != None:\n') + fd.write(' if isinstance(arg, (list, tuple)):\n') + fd.write(' if len(arg) == 0:\n') + fd.write(' continue\n') + fd.write(' __outputs__.append(arg[0])\n') + fd.write(' else:\n') + fd.write(' __outputs__.append(arg)\n') + fd.write(' __attrs__ = []\n') + for attr in self.attr_list: + fd.write(' if {} != None:\n'.format(attr)) + fd.write(' attr = {}\n') + fd.write(' attr["name"] = "{}"\n'.format(attr)) + fd.write(' attr["dtype"] = "{}"\n'.format(self.attr_val.get(attr).get('type'))) + fd.write(' attr["value"] = {}\n'.format(attr)) + fd.write(' __attrs__.append(attr)\n') + fd.write(' return __inputs__, __outputs__, __attrs__\n') + + def _write_impl(self: any, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + pchk = self._build_parachk() + if len(self.kern_name) > 0: + kern_name = self.kern_name + else: + kern_name = self.op_intf + src = self.op_file + '.cpp' + fd.write(IMPL_API.format(self.op_type, pchk, self.op_intf, argsdef, kern_name, argsval,\ + self.custom_compile_options, self.custom_all_compile_options, self.op_intf,\ + optype_snake_ex(self.op_type), optype_snake(self.op_type), src)) + if self.op_replay_flag: + fd.write(REPLAY_OP_API.format(self.op_type, kern_name, self.op_file, self.op_type, self.op_file,\ + self.op_compile_option)) + else: + fd.write(COMPILE_OP_API.format(self.op_type, self.op_type, ', '.join(self.input_name),\ + ', '.join(self.output_name), self.op_compile_option)) + + def _write_cap(self: any, cap_name: str, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + if cap_name == 'check_supported': + fd.write(SUP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) + else: + fd.write(CAP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) + + def _write_glz(self: any, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + fd.write(GLZ_API.format(self.op_type, self.op_intf, argsdef, argsval, self.op_type)) + + +def write_scripts(cfgfile: str, cfgs: dict, dirs: dict, ops: list = None, op_compile_option:list = None): + batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') + iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') + file_map = {} + op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, AdpBuilder,\ + ops, dirs.get(const_var.AUTO_GEN_DIR)) + for op_desc in op_descs: + op_desc.write_adapt(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), op_compile_option) + file_map[op_desc.op_type] = op_desc.op_file + return file_map + +if __name__ == '__main__': + if len(sys.argv) <= 6: + raise RuntimeError('arguments must greater equal than 6') + rep_cfg = {} + rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] + rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] + cfg_dir = {} + cfg_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] + cfg_dir[const_var.CFG_OUT_DIR] = sys.argv[5] + cfg_dir[const_var.AUTO_GEN_DIR] = sys.argv[6] + write_scripts(cfgfile=sys.argv[1], cfgs=rep_cfg, dirs=cfg_dir) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_ops_config.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_ops_config.py new file mode 100755 index 000000000..7a97180be --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_ops_config.py @@ -0,0 +1,114 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import glob +import json +import argparse +import const_var + + +def load_json(json_file: str): + with open(json_file, encoding='utf-8') as file: + json_content = json.load(file) + return json_content + + +def get_specified_suffix_file(root_dir, suffix): + specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) + all_suffix_files = glob.glob(specified_suffix, recursive=True) + return all_suffix_files + + +def add_simplified_config(op_type, key, core_type, objfile, config): + simple_cfg = config.get('binary_info_config.json') + op_cfg = simple_cfg.get(op_type) + if not op_cfg: + op_cfg = {} + op_cfg['dynamicRankSupport'] = True + op_cfg['simplifiedKeyMode'] = 0 + op_cfg['binaryList'] = [] + simple_cfg[op_type] = op_cfg + bin_list = op_cfg.get('binaryList') + bin_list.append({'coreType': core_type, 'simplifiedKey': key, 'binPath': objfile}) + + +def add_op_config(op_file, bin_info, config): + op_cfg = config.get(op_file) + if not op_cfg: + op_cfg = {} + op_cfg['binList'] = [] + config[op_file] = op_cfg + op_cfg.get('binList').append(bin_info) + + +def gen_ops_config(json_file, soc, config): + core_type_map = {"MIX": 0, "AiCore": 1, "VectorCore": 2} + contents = load_json(json_file) + if ('binFileName' not in contents) or ('supportInfo' not in contents): + return + json_base_name = os.path.basename(json_file) + op_dir = os.path.basename(os.path.dirname(json_file)) + support_info = contents.get('supportInfo') + bin_name = contents.get('binFileName') + bin_suffix = contents.get('binFileSuffix') + core_type = core_type_map.get(contents.get("coreType")) + bin_file_name = bin_name + bin_suffix + op_type = bin_name.split('_')[0] + op_file = op_dir + '.json' + bin_info = {} + keys = support_info.get('simplifiedKey') + if keys: + bin_info['simplifiedKey'] = keys + for key in keys: + add_simplified_config(op_type, key, core_type, os.path.join(soc, op_dir, bin_file_name), config) + bin_info['staticKey'] = support_info.get('staticKey') + bin_info['int64Mode'] = support_info.get('int64Mode') + bin_info['inputs'] = support_info.get('inputs') + bin_info['outputs'] = support_info.get('outputs') + if support_info.get('attrs'): + bin_info['attrs'] = support_info.get('attrs') + bin_info['binInfo'] = {'jsonFilePath': os.path.join(soc, op_dir, json_base_name)} + add_op_config(op_file, bin_info, config) + + +def gen_all_config(root_dir, soc): + suffix = 'json' + config = {} + config['binary_info_config.json'] = {} + all_json_files = get_specified_suffix_file(root_dir, suffix) + for _json in all_json_files: + gen_ops_config(_json, soc, config) + for cfg_key in config.keys(): + cfg_file = os.path.join(root_dir, cfg_key) + with os.fdopen(os.open(cfg_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + json.dump(config.get(cfg_key), fd, indent=' ') + + +def args_prase(): + parser = argparse.ArgumentParser() + parser.add_argument('-p', + '--path', + nargs='?', + required=True, + help='Parse the path of the json file.') + parser.add_argument('-s', + '--soc', + nargs='?', + required=True, + help='Parse the soc_version of ops.') + return parser.parse_args() + + +def main(): + args = args_prase() + gen_all_config(args.path, args.soc) + + +if __name__ == '__main__': + main() diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_replay_build.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_replay_build.py new file mode 100755 index 000000000..1cac7d911 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_replay_build.py @@ -0,0 +1,65 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import opdesc_parser +import replay_codegen +import const_var +from replay_codegen import ReplayCodeGenParams + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + + +class ReplayBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + super().__init__(op_type) + + def gen_replay_source(self: any, impl_path: str, out_path: str, ops_product: str): + if not self.op_replay_flag: + print('{} replay not enabled'.format(self.op_type)) + return + argn = len(self.input_name) + len(self.output_name) + 1 + if self.op_replay_batch: + print('{} replay in batch mode'.format(self.op_type)) + else: + print('{} replay in normal mode'.format(self.op_type)) + if impl_path.endswith('op_kernel'): + implf = os.path.join(impl_path, self.op_file + '.cpp') + tiling_file = os.path.join(impl_path, "../op_host", self.op_file + '_tiling.h') + else: + if self.dynamic_shape: + dyn_path = 'dynamic' + else: + dyn_path = '' + implf = os.path.join(impl_path, dyn_path, self.op_file + '.cpp') + tiling_file = os.path.join(impl_path, "../../op_tiling", self.op_file + '_tiling.h') + rep_conf = replay_codegen.ReplayCodeGen(ReplayCodeGenParams(self.op_type, implf, tiling_file, self.op_file, \ + self.op_intf, argn, self.op_replay_batch, self.max_block_dim, self.max_shape_size)) + rep_conf.set_batch(self.op_replay_batch) + rep_conf.set_outdir(out_path) + rep_conf.gen_replay(ops_product) + + +def gen_replay(cfgfile: str, cfgs: dict, dirs: dict, ops_product: str, ops: list = None): + batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') + iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') + op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, ReplayBuilder, ops) + for op_desc in op_descs: + op_desc.gen_replay_source(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), ops_product) + + +if __name__ == '__main__': + if len(sys.argv) <= 6: + raise RuntimeError('arguments must greater than 6') + rep_cfg = {} + rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] + rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] + rep_dir = {} + rep_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] + rep_dir[const_var.CFG_OUT_DIR] = sys.argv[5] + gen_replay(sys.argv[1], rep_cfg, rep_dir, sys.argv[6]) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/batch_replay_impl.temp b/atb_operator/AtbTrilSample/TrilOp/cmake/util/batch_replay_impl.temp new file mode 100755 index 000000000..0e8834664 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/batch_replay_impl.temp @@ -0,0 +1,117 @@ +#include +#include +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#include "register/op_check.h" +#define __ASCENDC_REPLAY_CODE__ +#include + +using namespace std; +using namespace optiling; +using namespace AscendCReplay; + +extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); +extern "C" int elf_batch_append(char *elf, uint32_t elfSize, char *jit, int kernum, char *atext[], int alen[], + int atlen, const char* kernelname[]); + +#define KERNEL_N 1 +#define ARG_N (__ARG_NUM__) +#define MAX_L (1024 * 1024 * 100) +#define MAX_E (1024 * 1024) + +int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) +{ + // gen type 1 : direct call codes 0: load .o file + if (param.gentype < 0 || param.gentype > 1) { + printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); + return 0; + } else if (param.gentype == 1 && param.objptr == nullptr) { + printf("Error: call replay with direct call mode, but code obj addr is null\n"); + return 0; + } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { + printf("Error: call replay with object file mode, but object file path is null\n"); + return 0; + } + // core_type 0:MIX 1:CUBE 2:VEC + if (core_type < 0 || core_type > 2) { + printf("Error: call replay core type is %d !\n", core_type); + return 0; + } + g_coreType = __CORE_TYPE__; + g_taskRation = param.task_ration; + g_tilingKey = param.tiling_key; + + unsigned char *buf, *jit; + char *kernel[KERNEL_N]; + int len[KERNEL_N]; + block_idx = 0; + block_num = param.block_dim; + g_ubBase = block_num; + uint8_t *code = (uint8_t *)malloc(MAX_L); + uint8_t *pos = code; + struct timespec tp1, tp2; + + clock_gettime(CLOCK_MONOTONIC, &tp1); + if (block_num > 32) { + printf("Error: block_num > 32\n"); + return 0; + } + //__OP_FOPEN__ + for (int i = 0; i < KERNEL_N; i++) { + //__OP_SET_KERNEL__ + for (int j = 0; j < ARG_N; j++) + AddArg(j, ARG_STEP * (j + 1)); +#ifdef FP_CEILING + SetCtrlFloatEnable(); +#else + SetCtrlFloatDisable(); +#endif + CodeInit(pos, true); + __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); + CodeEnd(); + kernel[i] = (char *)pos; + len[i] = CodeLen(); + pos += len[i]; + } + //__OP_FCLOSE__ + clock_gettime(CLOCK_MONOTONIC, &tp2); + buf = (unsigned char *)malloc(MAX_E); + int fd = open(param.entry_file, O_RDONLY); + if (fd < 0) { + printf("[error]: cannot find entry.o : %s\n", param.entry_file); + return 0; + } + uint32_t bufSize = read(fd, buf, MAX_E); + if (bufSize <= 0) { + printf("[error]: entry.o : %s is too small ! \n", param.entry_file); + } + close(fd); + jit = (unsigned char *)malloc(MAX_L); + printf("total code generated %ld\n", pos - code); + int sz = elf_batch_append((char *)buf, bufSize, (char *)jit, KERNEL_N, kernel, len, pos - code, ¶m.kernel_name); + if (tp1.tv_sec != tp2.tv_sec) { + printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); + } else { + printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); + } + printf("new elf size %d\n", sz); + if (param.gentype == 0) { + fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); + (void)write(fd, jit, sz); + close(fd); + free(jit); + } else if (param.gentype == 1) { + *param.objptr = (char*)jit; + } + free(buf); + free(code); + return sz; +} + +REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/code_channel_infer.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/code_channel_infer.py new file mode 100755 index 000000000..d4e6a540c --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/code_channel_infer.py @@ -0,0 +1,58 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" +import os +import stat +import ctypes +import collections +import shutil +import subprocess +import copy + +"""CODE_* is used to cube/vector api is called in operator code +CODE_MIX means both cube and vector api is called +CODE_CUBE means only cube api is called +CODE_VEC means only vector api is called +""" +CODE_MIX = 0 +CODE_CUBE = 1 +CODE_VEC = 2 + + +def _is_v220(op_product: str): + """return if current soc version is V220 + + Returns: + res: True means V220 + """ + if op_product in ["ascend910b", "ascend910c"]: + return True + return False + + +InfoCodeChanelParams = collections.namedtuple('InfoCodeChanelParams',\ +['src_file', 'tiling_header', 'kernel_name', 'outdir', 'op_product', 'compile_options']) + + +def infer_code_channel(params: InfoCodeChanelParams): + """get code channel for v220, return CODE_MIX if soc version is not V220 + + Args: + src_file (str): AscendC operator code file + src_file (str): AscendC operator tiling header file + kernel_name (str): kernel function name + optype (str): operator type + compile_options (list): compile options for bisheng cmd + + Raises: + Exception: if not exist L1/L0/UB if code, it's not a aicore code + + Returns: + res (int): CODE_MIX/CODE_CUBE/CODE_VEC + """ + if not _is_v220(params.op_product): + return CODE_MIX + return CODE_VEC diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/const_var.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/const_var.py new file mode 100755 index 000000000..3e71f5ab8 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/const_var.py @@ -0,0 +1,33 @@ + +#!/usr/bin/env python +# coding=utf-8 +""" +Function: +The replay funtion entry +Copyright Information: +Huawei Technologies Co., Ltd. All Rights Reserved © 2020 +""" + +import os +import stat + + +REPLAY_BATCH = 'batch' +REPLAY_ITERATE = 'iterate' +CFG_IMPL_DIR = 'impl_dir' +CFG_OUT_DIR = 'out_dir' +AUTO_GEN_DIR = 'auto_gen_dir' +WFLAGS = os.O_WRONLY | os.O_CREAT | os.O_TRUNC +WMODES = stat.S_IWUSR | stat.S_IRUSR +SOC_MAP_EXT = {'ascend310p': 'Ascend310P3', 'ascend310b': 'Ascend310B1', + 'ascend910': 'Ascend910A', 'ascend910b': 'Ascend910B1', 'ascend910c': 'Ascend910C1'} +BIN_CMD = 'opc $1 --main_func={fun} --input_param={param} --soc_version={soc} \ +--output=$2 --impl_mode={impl} --simplified_key_mode=0 --op_mode=dynamic\n' +CHK_CMD = ''' +if ! test -f $2/{res_file} ; then + echo "$2/{res_file} not generated!" + exit 1 +fi +''' +ATTR_DEF_VAL = {'str' : '', 'int': 0, 'float': 0.0, 'bool': False, 'list_bool': [], + 'list_int': [], 'list_float': [], 'list_list_int': [[]]} diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_impl_and_mrege_json.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_impl_and_mrege_json.sh new file mode 100755 index 000000000..492034e49 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_impl_and_mrege_json.sh @@ -0,0 +1,20 @@ +#!/usr/bin/bash + +project_path=$1 +build_path=$2 +vendor_name=customize +if [[ ! -d "$project_path" ]]; then + echo "[ERROR] No projcet path is provided" + exit 1 +fi + +if [[ ! -d "$build_path" ]]; then + echo "[ERROR] No build path is provided" + exit 1 +fi + +# copy aicpu kernel so operators +if [[ -d "${project_path}/cpukernel/aicpu_kernel_lib" ]]; then + cp -f ${project_path}/cpukernel/aicpu_kernel_lib/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/cpu/aicpu_kernel/impl + rm -rf ${project_path}/cpukernel/aicpu_kernel_lib +fi diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_ops_filter.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_ops_filter.sh new file mode 100755 index 000000000..d4c27d17f --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_ops_filter.sh @@ -0,0 +1,62 @@ +#!/bin/bash +# Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +# Description: Generate npu_supported_ops.json +# ============================================================================== + +if [[ -z "$1" ]]; then + echo -e "[ERROR] No source dir provided" + exit 1 +fi + +if [[ -z "$2" ]]; then + echo -e "[ERROR] No destination dir provided" + exit 1 +fi + +src=$1 +dest_file=$2/npu_supported_ops.json + +if [ -f "$dest_file" ];then + chmod u+w $dest_file +fi + +echo $* + +add_ops() { + name=$1 + isHeavy=$2 + file=$3 + grep -w "\"$name\"" ${file} >/dev/null + if [ $? == 0 ];then + return + fi + echo " \"${name}\": {" >> ${file} + echo " \"isGray\": false," >> ${file} + echo " \"isHeavy\": ${isHeavy}" >> ${file} + echo " }," >> ${file} +} + +echo "{" > ${dest_file} +ini_files=$(find ${src} -name "*.ini") +for file in ${ini_files} ; do + name=$(grep '^\[' ${file} | sed 's/\[//g' | sed 's/]//g' | sed 's/\r//g') + grep 'heavyOp.flag' ${file} >/dev/null + if [ $? == 0 ];then + isHeavy=$(grep 'heavyOp.flag' ${file} | awk -F= '{print $2}') + else + isHeavy="false" + fi + for op in ${name} ; do + add_ops ${op} "false" ${dest_file} + done +done +echo "}" >> ${dest_file} +file_count=$(cat ${dest_file} | wc -l) +line=$(($file_count-1)) +sed -i "${line}{s/,//g}" ${dest_file} + +chmod 640 "${dest_file}" +echo -e "[INFO] Succed generated ${dest_file}" + +exit 0 + diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_version_info.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_version_info.sh new file mode 100755 index 000000000..a06cfc78d --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_version_info.sh @@ -0,0 +1,6 @@ +ascend_install_dir=$1 +gen_file_dir=$2 + +# create version.info +compiler_version=$(grep "Version" -w ${ascend_install_dir}/compiler/version.info | awk -F = '{print $2}') +echo "custom_opp_compiler_version=${compiler_version}" > ${gen_file_dir}/version.info \ No newline at end of file diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_op_info.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_op_info.py new file mode 100755 index 000000000..28ba08757 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_op_info.py @@ -0,0 +1,36 @@ +# -*- coding: utf-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" +import json +import os +import sys +import stat +import const_var + + +if __name__ == '__main__': + if len(sys.argv) != 3: + print(sys.argv) + print('argv error, inert_op_info.py your_op_file lib_op_file') + sys.exit(2) + + with open(sys.argv[1], 'r') as load_f: + insert_operator = json.load(load_f) + + all_operators = {} + if os.path.exists(sys.argv[2]): + if os.path.getsize(sys.argv[2]) != 0: + with open(sys.argv[2], 'r') as load_f: + all_operators = json.load(load_f) + + for k in insert_operator.keys(): + if k in all_operators.keys(): + print('replace op:[', k, '] success') + else: + print('insert op:[', k, '] success') + all_operators[k] = insert_operator[k] + + with os.fdopen(os.open(sys.argv[2], const_var.WFLAGS, const_var.WMODES), 'w') as json_file: + json_file.write(json.dumps(all_operators, indent=4)) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_simplified_keys.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_simplified_keys.py new file mode 100755 index 000000000..7e8dc2c55 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_simplified_keys.py @@ -0,0 +1,242 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import os +import re +import glob +import json +import argparse + + +DATA_TPYE_DICT = { + 'float32': 0, + 'float16': 1, + 'int8': 2, + 'int16': 6, + 'uint16': 7, + 'uint8': 4, + 'int32': 3, + 'int64': 9, + 'uint32': 8, + 'uint64': 10, + 'bool': 12, + 'double': 11, + 'complex64': 16, + 'complex128': 17, + 'qint8': 18, + 'qint16': 19, + 'qint32': 20, + 'quint8': 21, + 'quint16': 22, + 'resource': 23, + 'string': 24, + 'dual': 25, + 'variant': 26, + 'bf16': 27, + 'bfloat16': 27, + 'undefined': 28, + 'int4': 29, + 'uint1': 30, + 'int2': 31 +} + +FORMAT_DICT = { + 'NCHW': 0, + 'NHWC': 1, + 'ND': 2, + 'NC1HWC0': 3, + 'FRACTAL_Z': 4, + 'NC1C0HWPAD': 5, + 'NHWC1C0': 6, + 'FSR_NCHW': 7, + 'FRACTAL_DECONV': 8, + 'C1HWNC0': 9, + 'FRACTAL_DECONV_TRANSPOSE': 10, + 'FRACTAL_DECONV_SP_STRIDE_TRANS': 11, + 'NC1HWC0_C04': 12, + 'FRACTAL_Z_C04': 13, + 'CHWN': 14, + 'FRACTAL_DECONV_SP_STRIDE8_TRANS': 15, + 'HWCN': 16, + 'NC1KHKWHWC0': 17, + 'BN_WEIGHT': 18, + 'FILTER_HWCK': 19, + 'HASHTABLE_LOOKUP_LOOKUPS': 20, + 'HASHTABLE_LOOKUP_KEYS': 21, + 'HASHTABLE_LOOKUP_VALUE': 22, + 'HASHTABLE_LOOKUP_OUTPUT': 23, + 'HASHTABLE_LOOKUP_HITS': 24, + 'C1HWNCoC0': 25, + 'MD': 26, + 'NDHWC': 27, + 'FRACTAL_ZZ': 28, + 'FRACTAL_NZ': 29, + 'NCDHW': 30, + 'DHWCN': 31, + 'NDC1HWC0': 32, + 'FRACTAL_Z_3D': 33, + 'CN': 34, + 'NC': 35, + 'DHWNC': 36, + 'FRACTAL_Z_3D_TRANSPOSE': 37, + 'FRACTAL_ZN_LSTM': 38, + 'FRACTAL_Z_G': 39, + 'RESERVED': 40, + 'ALL': 41, + 'NULL': 42, + 'ND_RNN_BIAS': 43, + 'FRACTAL_ZN_RNN': 44, + 'NYUV': 45, + 'NYUV_A': 46 +} + + +def load_json(json_file: str): + with open(json_file, encoding='utf-8') as file: + json_content = json.load(file) + return json_content + + +def get_specified_suffix_file(root_dir, suffix): + specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) + all_suffix_files = glob.glob(specified_suffix, recursive=True) + return all_suffix_files + + +def get_deterministic_value(support_info): + deterministic_key = 'deterministic' + if deterministic_key not in support_info: + return 0 + deterministic_value = support_info.get(deterministic_key) + if deterministic_value == 'true': + return 1 + else: + return 0 + + +def get_precision_value(support_info): + precision_key = 'implMode' + precision_value = support_info.get(precision_key) + if precision_value == 'high_performance': + _value = 1 + elif precision_value == 'high_precision': + _value = 2 + else: + _value = 0 + return _value + + +def get_overflow_value(support_info): + return 0 + + +def get_parameters(info): + if info: + if 'dtype' in info: + data_type = info['dtype'] + data_type_value = DATA_TPYE_DICT.get(data_type) + else: + data_type_value = 0 + if 'format' in info: + _format = info['format'] + _format_value = FORMAT_DICT.get(_format) + else: + _format_value = 0 + else: + data_type_value = 0 + _format_value = 0 + return str(data_type_value), str(_format_value) + + +def get_dynamic_parameters(info): + # 动态输入时只需获取第一个参数 + return get_parameters(info[0]) + + +def get_all_parameters(support_info, _type): + result_list = list() + info_lists = support_info.get(_type) + if info_lists: + for _info in info_lists: + # 输入为列表时是动态输入 + if isinstance(_info, (list, tuple)): + data_type_value, _format_value = get_dynamic_parameters(_info) + else: + data_type_value, _format_value = get_parameters(_info) + result_list.append("{},{}".format(data_type_value, _format_value)) + return result_list + + +def get_all_input_parameters(support_info): + result = get_all_parameters(support_info, 'inputs') + return '/'.join(result) + + +def insert_content_into_file(input_file, content): + with open(input_file, 'r+') as file: + lines = file.readlines() + for index, line in enumerate(lines): + match_result = re.search(r'"staticKey":', line) + if match_result: + count = len(line) - len(line.lstrip()) + new_content = "{}{}".format(' ' * count, content) + # 插入到前一行,防止插入最后时还需要考虑是否添加逗号 + lines.insert(index, new_content) + break + file.seek(0) + file.write(''.join(lines)) + + +def insert_simplified_keys(json_file): + contents = load_json(json_file) + # 不存在'binFileName'或者'supportInfo'字段时,非需要替换的解析json文件 + if ('binFileName' not in contents) or ('supportInfo' not in contents): + return + support_info = contents.get('supportInfo') + bin_file_name = contents.get('binFileName') + # 'simplifiedKey'字段已经存在时,直接返回,不重复生成 + if 'simplifiedKey' in support_info: + return + op_type = bin_file_name.split('_')[0] + deterministic = str(get_deterministic_value(support_info)) + precision = str(get_precision_value(support_info)) + overflow = str(get_overflow_value(support_info)) + input_parameters = get_all_input_parameters(support_info) + key = '{}/d={},p={},o={}/{}/'.format( + op_type, + deterministic, + precision, + overflow, + input_parameters) + result = '"simplifiedKey": "' + key + '",\n' + insert_content_into_file(json_file, result) + + +def insert_all_simplified_keys(root_dir): + suffix = 'json' + all_json_files = get_specified_suffix_file(root_dir, suffix) + for _json in all_json_files: + insert_simplified_keys(_json) + + +def args_prase(): + parser = argparse.ArgumentParser() + parser.add_argument('-p', + '--path', + nargs='?', + required=True, + help='Parse the path of the json file.') + return parser.parse_args() + + +def main(): + args = args_prase() + insert_all_simplified_keys(args.path) + + +if __name__ == '__main__': + main() diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_entry.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_entry.py new file mode 100755 index 000000000..2b77c970d --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_entry.py @@ -0,0 +1,115 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + + +def gen_fun_def(title, kernel, argn, arg_type, arg_name): + entry = [] + entry.append(title) + entry.append(kernel) + entry.append('(') + args = [] + for i in range(0, argn): + args.append(arg_type + ' ' + arg_name + str(i)) + entry.append(', '.join(args)) + entry.append(')') + return ' '.join(entry) + + +def gen_batch_kernel_body(fname, argn, arg_name): + body = [] + body.append('{') + fun = [] + fun.append(fname) + fun.append('(') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + fun.append(', '.join(args)) + fun.append(');') + body.append(' '.join(fun)) + body.append('}') + return '\n'.join(body) + + +def gen_mc_kernel_body(kn, argn, arg_name, blknum): + body = [] + body.append('{') + body.append(' switch(block_idx) {') + for blk in range(0, blknum): + fun = [] + fun.append('{}_blk{:02d}'.format(kn, blk)) + fun.append('(') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + fun.append(', '.join(args)) + fun.append(')') + body.append(' case {}: {}; break;'.format(blk, ' '.join(fun))) + body.append(' default: break;') + body.append(' }') + body.append('}') + return '\n'.join(body) + + +def gen_proc_body(argn, arg_name): + body = [] + body.append('{') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + body.append('uint64_t __x = (uint64_t)' + ' + (uint64_t)'.join(args) + ';') + body.append('__asm__ ("NOP");') + body.append('__asm__ ("NOP");') + body.append('__asm__ ("NOP");') + body.append('}') + return '\n'.join(body) + + +def batch_code_gen(kn, argn, argt): + codes = [] + kernel_name = kn + proc_name = kernel_name + '_percore' + arg_num = int(argn) + data_type = argt + arg_type = '__gm__ ' + data_type + '* __restrict__' + arg_name = 'arg' + kernel_title = 'extern \"C\" __global__ __aicore__ void' + proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' + codes.append('#ifndef __aicore__') + codes.append('#define __aicore__ [aicore]') + codes.append('#endif') + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') + codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) + codes.append(gen_batch_kernel_body(proc_name, arg_num, arg_name)) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) + codes.append(gen_proc_body(arg_num, arg_name)) + return '\n'.join(codes) + '\n' + + +def mc_code_gen(kn, argn, argt, blknum): + codes = [] + kernel_name = kn + core_num = int(blknum) + arg_num = int(argn) + data_type = argt + arg_type = '__gm__ ' + data_type + '* __restrict__' + arg_name = 'arg' + kernel_title = 'extern \"C\" __global__ __aicore__ void' + proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' + codes.append('#ifndef __aicore__') + codes.append('#define __aicore__ [aicore]') + codes.append('#endif') + for i in range(0, core_num): + proc_name = '{}_blk{:02d}'.format(kernel_name, i) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') + codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) + codes.append(gen_mc_kernel_body(kernel_name, arg_num, arg_name, core_num)) + for i in range(0, core_num): + proc_name = '{}_blk{:02d}'.format(kernel_name, i) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) + codes.append(gen_proc_body(arg_num, arg_name)) + return '\n'.join(codes) + '\n' diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_impl.temp b/atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_impl.temp new file mode 100755 index 000000000..5079a1043 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_impl.temp @@ -0,0 +1,10 @@ +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#define __ASCENDC_REPLAY_CODE__ +#include "__CCE_FILE__" diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/COPYING b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/COPYING new file mode 100755 index 000000000..d159169d1 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/COPYING @@ -0,0 +1,339 @@ + GNU GENERAL PUBLIC LICENSE + Version 2, June 1991 + + Copyright (C) 1989, 1991 Free Software Foundation, Inc., + 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + Everyone is permitted to copy and distribute verbatim copies + of this license document, but changing it is not allowed. + + Preamble + + The licenses for most software are designed to take away your +freedom to share and change it. By contrast, the GNU General Public +License is intended to guarantee your freedom to share and change free +software--to make sure the software is free for all its users. This +General Public License applies to most of the Free Software +Foundation's software and to any other program whose authors commit to +using it. (Some other Free Software Foundation software is covered by +the GNU Lesser General Public License instead.) You can apply it to +your programs, too. + + When we speak of free software, we are referring to freedom, not +price. Our General Public Licenses are designed to make sure that you +have the freedom to distribute copies of free software (and charge for +this service if you wish), that you receive source code or can get it +if you want it, that you can change the software or use pieces of it +in new free programs; and that you know you can do these things. + + To protect your rights, we need to make restrictions that forbid +anyone to deny you these rights or to ask you to surrender the rights. +These restrictions translate to certain responsibilities for you if you +distribute copies of the software, or if you modify it. + + For example, if you distribute copies of such a program, whether +gratis or for a fee, you must give the recipients all the rights that +you have. You must make sure that they, too, receive or can get the +source code. And you must show them these terms so they know their +rights. + + We protect your rights with two steps: (1) copyright the software, and +(2) offer you this license which gives you legal permission to copy, +distribute and/or modify the software. + + Also, for each author's protection and ours, we want to make certain +that everyone understands that there is no warranty for this free +software. If the software is modified by someone else and passed on, we +want its recipients to know that what they have is not the original, so +that any problems introduced by others will not reflect on the original +authors' reputations. + + Finally, any free program is threatened constantly by software +patents. We wish to avoid the danger that redistributors of a free +program will individually obtain patent licenses, in effect making the +program proprietary. To prevent this, we have made it clear that any +patent must be licensed for everyone's free use or not licensed at all. + + The precise terms and conditions for copying, distribution and +modification follow. + + GNU GENERAL PUBLIC LICENSE + TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION + + 0. This License applies to any program or other work which contains +a notice placed by the copyright holder saying it may be distributed +under the terms of this General Public License. The "Program", below, +refers to any such program or work, and a "work based on the Program" +means either the Program or any derivative work under copyright law: +that is to say, a work containing the Program or a portion of it, +either verbatim or with modifications and/or translated into another +language. (Hereinafter, translation is included without limitation in +the term "modification".) Each licensee is addressed as "you". + +Activities other than copying, distribution and modification are not +covered by this License; they are outside its scope. The act of +running the Program is not restricted, and the output from the Program +is covered only if its contents constitute a work based on the +Program (independent of having been made by running the Program). +Whether that is true depends on what the Program does. + + 1. You may copy and distribute verbatim copies of the Program's +source code as you receive it, in any medium, provided that you +conspicuously and appropriately publish on each copy an appropriate +copyright notice and disclaimer of warranty; keep intact all the +notices that refer to this License and to the absence of any warranty; +and give any other recipients of the Program a copy of this License +along with the Program. + +You may charge a fee for the physical act of transferring a copy, and +you may at your option offer warranty protection in exchange for a fee. + + 2. You may modify your copy or copies of the Program or any portion +of it, thus forming a work based on the Program, and copy and +distribute such modifications or work under the terms of Section 1 +above, provided that you also meet all of these conditions: + + a) You must cause the modified files to carry prominent notices + stating that you changed the files and the date of any change. + + b) You must cause any work that you distribute or publish, that in + whole or in part contains or is derived from the Program or any + part thereof, to be licensed as a whole at no charge to all third + parties under the terms of this License. + + c) If the modified program normally reads commands interactively + when run, you must cause it, when started running for such + interactive use in the most ordinary way, to print or display an + announcement including an appropriate copyright notice and a + notice that there is no warranty (or else, saying that you provide + a warranty) and that users may redistribute the program under + these conditions, and telling the user how to view a copy of this + License. (Exception: if the Program itself is interactive but + does not normally print such an announcement, your work based on + the Program is not required to print an announcement.) + +These requirements apply to the modified work as a whole. If +identifiable sections of that work are not derived from the Program, +and can be reasonably considered independent and separate works in +themselves, then this License, and its terms, do not apply to those +sections when you distribute them as separate works. But when you +distribute the same sections as part of a whole which is a work based +on the Program, the distribution of the whole must be on the terms of +this License, whose permissions for other licensees extend to the +entire whole, and thus to each and every part regardless of who wrote it. + +Thus, it is not the intent of this section to claim rights or contest +your rights to work written entirely by you; rather, the intent is to +exercise the right to control the distribution of derivative or +collective works based on the Program. + +In addition, mere aggregation of another work not based on the Program +with the Program (or with a work based on the Program) on a volume of +a storage or distribution medium does not bring the other work under +the scope of this License. + + 3. You may copy and distribute the Program (or a work based on it, +under Section 2) in object code or executable form under the terms of +Sections 1 and 2 above provided that you also do one of the following: + + a) Accompany it with the complete corresponding machine-readable + source code, which must be distributed under the terms of Sections + 1 and 2 above on a medium customarily used for software interchange; or, + + b) Accompany it with a written offer, valid for at least three + years, to give any third party, for a charge no more than your + cost of physically performing source distribution, a complete + machine-readable copy of the corresponding source code, to be + distributed under the terms of Sections 1 and 2 above on a medium + customarily used for software interchange; or, + + c) Accompany it with the information you received as to the offer + to distribute corresponding source code. (This alternative is + allowed only for noncommercial distribution and only if you + received the program in object code or executable form with such + an offer, in accord with Subsection b above.) + +The source code for a work means the preferred form of the work for +making modifications to it. For an executable work, complete source +code means all the source code for all modules it contains, plus any +associated interface definition files, plus the scripts used to +control compilation and installation of the executable. However, as a +special exception, the source code distributed need not include +anything that is normally distributed (in either source or binary +form) with the major components (compiler, kernel, and so on) of the +operating system on which the executable runs, unless that component +itself accompanies the executable. + +If distribution of executable or object code is made by offering +access to copy from a designated place, then offering equivalent +access to copy the source code from the same place counts as +distribution of the source code, even though third parties are not +compelled to copy the source along with the object code. + + 4. You may not copy, modify, sublicense, or distribute the Program +except as expressly provided under this License. Any attempt +otherwise to copy, modify, sublicense or distribute the Program is +void, and will automatically terminate your rights under this License. +However, parties who have received copies, or rights, from you under +this License will not have their licenses terminated so long as such +parties remain in full compliance. + + 5. You are not required to accept this License, since you have not +signed it. However, nothing else grants you permission to modify or +distribute the Program or its derivative works. These actions are +prohibited by law if you do not accept this License. Therefore, by +modifying or distributing the Program (or any work based on the +Program), you indicate your acceptance of this License to do so, and +all its terms and conditions for copying, distributing or modifying +the Program or works based on it. + + 6. Each time you redistribute the Program (or any work based on the +Program), the recipient automatically receives a license from the +original licensor to copy, distribute or modify the Program subject to +these terms and conditions. You may not impose any further +restrictions on the recipients' exercise of the rights granted herein. +You are not responsible for enforcing compliance by third parties to +this License. + + 7. If, as a consequence of a court judgment or allegation of patent +infringement or for any other reason (not limited to patent issues), +conditions are imposed on you (whether by court order, agreement or +otherwise) that contradict the conditions of this License, they do not +excuse you from the conditions of this License. If you cannot +distribute so as to satisfy simultaneously your obligations under this +License and any other pertinent obligations, then as a consequence you +may not distribute the Program at all. For example, if a patent +license would not permit royalty-free redistribution of the Program by +all those who receive copies directly or indirectly through you, then +the only way you could satisfy both it and this License would be to +refrain entirely from distribution of the Program. + +If any portion of this section is held invalid or unenforceable under +any particular circumstance, the balance of the section is intended to +apply and the section as a whole is intended to apply in other +circumstances. + +It is not the purpose of this section to induce you to infringe any +patents or other property right claims or to contest validity of any +such claims; this section has the sole purpose of protecting the +integrity of the free software distribution system, which is +implemented by public license practices. Many people have made +generous contributions to the wide range of software distributed +through that system in reliance on consistent application of that +system; it is up to the author/donor to decide if he or she is willing +to distribute software through any other system and a licensee cannot +impose that choice. + +This section is intended to make thoroughly clear what is believed to +be a consequence of the rest of this License. + + 8. If the distribution and/or use of the Program is restricted in +certain countries either by patents or by copyrighted interfaces, the +original copyright holder who places the Program under this License +may add an explicit geographical distribution limitation excluding +those countries, so that distribution is permitted only in or among +countries not thus excluded. In such case, this License incorporates +the limitation as if written in the body of this License. + + 9. The Free Software Foundation may publish revised and/or new versions +of the General Public License from time to time. Such new versions will +be similar in spirit to the present version, but may differ in detail to +address new problems or concerns. + +Each version is given a distinguishing version number. If the Program +specifies a version number of this License which applies to it and "any +later version", you have the option of following the terms and conditions +either of that version or of any later version published by the Free +Software Foundation. If the Program does not specify a version number of +this License, you may choose any version ever published by the Free Software +Foundation. + + 10. If you wish to incorporate parts of the Program into other free +programs whose distribution conditions are different, write to the author +to ask for permission. For software which is copyrighted by the Free +Software Foundation, write to the Free Software Foundation; we sometimes +make exceptions for this. Our decision will be guided by the two goals +of preserving the free status of all derivatives of our free software and +of promoting the sharing and reuse of software generally. + + NO WARRANTY + + 11. BECAUSE THE PROGRAM IS LICENSED FREE OF CHARGE, THERE IS NO WARRANTY +FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN +OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES +PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED +OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF +MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS +TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE +PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, +REPAIR OR CORRECTION. + + 12. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING +WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MAY MODIFY AND/OR +REDISTRIBUTE THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, +INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING +OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED +TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY +YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER +PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE +POSSIBILITY OF SUCH DAMAGES. + + END OF TERMS AND CONDITIONS + + How to Apply These Terms to Your New Programs + + If you develop a new program, and you want it to be of the greatest +possible use to the public, the best way to achieve this is to make it +free software which everyone can redistribute and change under these terms. + + To do so, attach the following notices to the program. It is safest +to attach them to the start of each source file to most effectively +convey the exclusion of warranty; and each file should have at least +the "copyright" line and a pointer to where the full notice is found. + + + Copyright (C) + + This program is free software; you can redistribute it and/or modify + it under the terms of the GNU General Public License as published by + the Free Software Foundation; either version 2 of the License, or + (at your option) any later version. + + This program is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + GNU General Public License for more details. + + You should have received a copy of the GNU General Public License along + with this program; if not, write to the Free Software Foundation, Inc., + 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. + +Also add information on how to contact you by electronic and paper mail. + +If the program is interactive, make it output a short notice like this +when it starts in an interactive mode: + + Gnomovision version 69, Copyright (C) year name of author + Gnomovision comes with ABSOLUTELY NO WARRANTY; for details type `show w'. + This is free software, and you are welcome to redistribute it + under certain conditions; type `show c' for details. + +The hypothetical commands `show w' and `show c' should show the appropriate +parts of the General Public License. Of course, the commands you use may +be called something other than `show w' and `show c'; they could even be +mouse-clicks or menu items--whatever suits your program. + +You should also get your employer (if you work as a programmer) or your +school, if any, to sign a "copyright disclaimer" for the program, if +necessary. Here is a sample; alter the names: + + Yoyodyne, Inc., hereby disclaims all copyright interest in the program + `Gnomovision' (which makes passes at compilers) written by James Hacker. + + , 1 April 1989 + Ty Coon, President of Vice + +This General Public License does not permit incorporating your program into +proprietary programs. If your program is a subroutine library, you may +consider it more useful to permit linking proprietary applications with the +library. If this is what you want to do, use the GNU Lesser General +Public License instead of this License. diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/README.md b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/README.md new file mode 100755 index 000000000..b41f01682 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/README.md @@ -0,0 +1,246 @@ +[![License: GPL v2](https://img.shields.io/badge/License-GPL%20v2-blue.svg)](https://www.gnu.org/licenses/old-licenses/gpl-2.0.en.html) +![Build Status](https://github.com/megastep/makeself/workflows/CI/badge.svg) + +# makeself - Make self-extractable archives on Unix + +[makeself.sh][1] is a small shell script that generates a self-extractable +compressed tar archive from a directory. The resulting file appears as a shell script +(many of those have a **.run** suffix), and can be launched as is. The archive +will then uncompress itself to a temporary directory and an optional arbitrary +command will be executed (for example an installation script). This is pretty +similar to archives generated with WinZip Self-Extractor in the Windows world. +Makeself archives also include checksums for integrity self-validation (CRC +and/or MD5/SHA256 checksums). + +The makeself.sh script itself is used only to create the archives from a +directory of files. The resultant archive is actually a compressed (using +gzip, bzip2, or compress) TAR archive, with a small shell script stub at the +beginning. This small stub performs all the steps of extracting the files, +running the embedded command, and removing the temporary files when done. +All the user has to do to install the software contained in such an +archive is to "run" the archive, i.e **sh nice-software.run**. I recommend +using the ".run" (which was introduced by some Makeself archives released by +Loki Software) or ".sh" suffix for such archives not to confuse the users, +so that they will know they are actually shell scripts (with quite a lot of binary data +attached to them though!). + +I am trying to keep the code of this script as portable as possible, i.e it is +not relying on any bash-specific features and only calls commands that are +installed on any functioning UNIX-compatible system. This script as well as +the archives it generates should run on any Unix flavor, with any compatible +Bourne shell, provided of course that the compression programs are available. + +As of version 2.1, Makeself has been rewritten and tested on the following +platforms : + + * Linux (all distributions) + * Sun Solaris (8 and above) + * HP-UX (tested on 11.0 and 11i on HPPA RISC) + * SCO OpenUnix and OpenServer + * IBM AIX 5.1L + * macOS (Darwin) + * SGI IRIX 6.5 + * FreeBSD + * UnicOS / Cray + * Cygwin (Windows) + +If you successfully run Makeself and/or archives created with it on another +system, then please [let me know][2]! + +Examples of publicly available archives made using makeself are : + + * Game patches and installers for [Id Software][3] games like Quake 3 for Linux or Return To Castle Wolfenstein ; + * All game patches released by [Loki Software][4] for the Linux version of popular games ; + * The [nVidia drivers][5] for Linux + * The installer for the Linux version of [Google Earth][6] + * The [VirtualBox][7] installers for Linux + * The [Makeself][1] distribution itself ;-) + * and countless others... + +**Important note for Apache users:** By default, most Web servers will think that Makeself archives are regular text files and thus they may show up as text in a Web browser. The correct way to prevent this is to add a MIME type for this file format, like so (in httpd.conf) : + +`AddType application/x-makeself .run` + +**Important note for certain GNU/Linux distributions:** Archives created with Makeself prior to v2.1.2 were using an old syntax for the _head_ and _tail_ Unix commands that is being progressively obsoleted in their GNU forms. Therefore you may have problems uncompressing some of these archives. A workaround for this is to set the environment variable $_POSIX2_VERSION to enable the old syntax, i.e. : + +`export _POSIX2_VERSION=199209` + +## Usage + +The syntax of makeself is the following: + +``` +makeself.sh [args] archive_dir file_name label startup_script [script_args] +``` + + * _args_ are optional options for Makeself. The available ones are : + + * **`--version`** : Prints the version number on stdout, then exits immediately + * **`--gzip`** : Use gzip for compression (the default on platforms on which gzip is commonly available, like Linux) + * **`--bzip2`** : Use bzip2 instead of gzip for better compression. The bzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. + * **`--pbzip2`** : Use pbzip2 instead of gzip for better and faster compression on machines having multiple CPUs. The pbzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. + * **`--xz`** : Use xz instead of gzip for better compression. The xz command must be available in the command path. It is recommended that the archive prefix be set to something like '.xz.run' for the archive, so that potential users know that they'll need xz to extract it. + * **`--lzo`** : Use lzop instead of gzip for better compression. The lzop command must be available in the command path. It is recommended that the archive prefix be set to something like `.lzo.run` for the archive, so that potential users know that they'll need lzop to extract it. + * **`--lz4`** : Use lz4 instead of gzip for better compression. The lz4 command must be available in the command path. It is recommended that the archive prefix be set to something like '.lz4.run' for the archive, so that potential users know that they'll need lz4 to extract it. + * **`--zstd`** : Use zstd instead of gzip for better compression. The zstd command must be available in the command path. It is recommended that the archive prefix be set to something like '.zstd.run' for the archive, so that potential users know that they'll need zstd to extract it. + * **`--pigz`** : Use pigz for compression. + * **`--base64`** : Encode the archive to ASCII in Base64 format instead of compressing (base64 command required). + * **`--gpg-encrypt`** : Encrypt the archive using `gpg -ac -z $COMPRESS_LEVEL`. This will prompt for a password to encrypt with. Assumes that potential users have `gpg` installed. + * **`--ssl-encrypt`** : Encrypt the archive using `openssl aes-256-cbc -a -salt`. This will prompt for a password to encrypt with. Assumes that the potential users have the OpenSSL tools installed. + * **`--compress`** : Use the UNIX `compress` command to compress the data. This should be the default on all platforms that don't have gzip available. + * **`--nocomp`** : Do not use any compression for the archive, which will then be an uncompressed TAR. + * **`--complevel`** : Specify the compression level for gzip, bzip2, pbzip2, zstd, xz, lzo or lz4. (defaults to 9) + * **`--threads`** : Specify the number of threads to be used by compressors that support parallelization. Omit to use compressor's default. Most useful (and required) for opting into xz's threading, usually with `--threads=0` for all available cores. pbzip2 and pigz are parallel by default, and setting this value allows limiting the number of threads they use. + * **`--notemp`** : The generated archive will not extract the files to a temporary directory, but in a new directory created in the current directory. This is better to distribute software packages that may extract and compile by themselves (i.e. launch the compilation through the embedded script). + * **`--current`** : Files will be extracted to the current directory, instead of in a subdirectory. This option implies `--notemp` above. + * **`--follow`** : Follow the symbolic links inside of the archive directory, i.e. store the files that are being pointed to instead of the links themselves. + * **`--append`** _(new in 2.1.x)_: Append data to an existing archive, instead of creating a new one. In this mode, the settings from the original archive are reused (compression type, label, embedded script), and thus don't need to be specified again on the command line. + * **`--header`** : Makeself uses a separate file to store the header stub, called `makeself-header.sh`. By default, it is assumed that it is stored in the same location as makeself.sh. This option can be used to specify its actual location if it is stored someplace else. + * **`--cleanup`** : Specify a script that is run when execution is interrupted or finishes successfully. The script is executed with the same environment and initial `script_args` as `startup_script`. + * **`--copy`** : Upon extraction, the archive will first extract itself to a temporary directory. The main application of this is to allow self-contained installers stored in a Makeself archive on a CD, when the installer program will later need to unmount the CD and allow a new one to be inserted. This prevents "Filesystem busy" errors for installers that span multiple CDs. + * **`--nox11`** : Disable the automatic spawning of a new terminal in X11. + * **`--nowait`** : When executed from a new X11 terminal, disable the user prompt at the end of the script execution. + * **`--nomd5`** and **`--nocrc`** : Disable the creation of a MD5 / CRC checksum for the archive. This speeds up the extraction process if integrity checking is not necessary. + * **`--sha256`** : Adds a SHA256 checksum for the archive. This is in addition to the MD5 / CRC checksums unless `--nomd5` is also used. + * **`--lsm` _file_** : Provide and LSM file to makeself, that will be embedded in the generated archive. LSM files are describing a software package in a way that is easily parseable. The LSM entry can then be later retrieved using the `--lsm` argument to the archive. An example of a LSM file is provided with Makeself. + * **`--tar-format opt`** : Specify the tar archive format (default is ustar); you may use any value accepted by your tar command (such as posix, v7, etc). + * **`--tar-extra opt`** : Append more options to the tar command line. + + For instance, in order to exclude the `.git` directory from the packaged archive directory using the GNU `tar`, one can use `makeself.sh --tar-extra "--exclude=.git" ...` + + * **`--keep-umask`** : Keep the umask set to shell default, rather than overriding when executing self-extracting archive. + * **`--packaging-date date`** : Use provided string as the packaging date instead of the current date. + * **`--license`** : Append a license file. + * **`--nooverwrite`** : Do not extract the archive if the specified target directory already exists. + * **`--help-header file`** : Add a header to the archive's `--help` output. + * `archive_dir` is the name of the directory that contains the files to be archived + * `file_name` is the name of the archive to be created + * `label` is an arbitrary text string describing the package. It will be displayed while extracting the files. + * `startup_script` is the command to be executed _from within_ the directory of extracted files. Thus, if you wish to execute a program contained in this directory, you must prefix your command with `./`. For example, `./program` will be fine. The `script_args` are additional arguments for this command. + +Here is an example, assuming the user has a package image stored in a **/home/joe/mysoft**, and he wants to generate a self-extracting package named +**mysoft.sh**, which will launch the "setup" script initially stored in /home/joe/mysoft : + +`makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup +` + +Here is also how I created the [makeself.run][9] archive which contains the Makeself distribution : + +`makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" ` + +Archives generated with Makeself can be passed the following arguments: + + * **`--keep`** : Prevent the files to be extracted in a temporary directory that will be removed after the embedded script's execution. The files will then be extracted in the current working directory and will stay here until you remove them. + * **`--verbose`** : Will prompt the user before executing the embedded command + * **`--target dir`** : Allows to extract the archive in an arbitrary place. + * **`--nox11`** : Do not spawn a X11 terminal. + * **`--confirm`** : Prompt the user for confirmation before running the embedded command. + * **`--info`** : Print out general information about the archive (does not extract). + * **`--lsm`** : Print out the LSM entry, if it is present. + * **`--list`** : List the files in the archive. + * **`--check`** : Check the archive for integrity using the embedded checksums. Does not extract the archive. + * **`--nochown`** : By default, a `chown -R` command is run on the target directory after extraction, so that all files belong to the current user. This is mostly needed if you are running as root, as tar will then try to recreate the initial user ownerships. You may disable this behavior with this flag. + * **`--tar`** : Run the tar command on the contents of the archive, using the following arguments as parameter for the command. + * **`--noexec`** : Do not run the embedded script after extraction. + * **`--noexec-cleanup`** : Do not run the embedded cleanup script. + * **`--nodiskspace`** : Do not check for available disk space before attempting to extract. + * **`--cleanup-args`** : Specify arguments to be passed to the cleanup script. Wrap value in quotes to specify multiple arguments. + +Any subsequent arguments to the archive will be passed as additional arguments to the embedded command. You must explicitly use the `--` special command-line construct before any such options to make sure that Makeself will not try to interpret them. + +## Startup Script + +The startup script must be a regular Shell script. + +Within the startup script, you can use the `$USER_PWD` variable to get the path of the folder from which the self-extracting script is executed. This is especially useful to access files that are located in the same folder as the script, as shown in the example below. + +`my-self-extracting-script.sh --fooBarFileParameter foo.bar` + +## Building and Testing + +Clone the git repo and execute `git submodule update --init --recursive` to obtain all submodules. + +* To make a release: `make` +* To run all tests: `make test` + +## Maven Usage + +Makeself is now supported by the following maven plugin [makeself-maven-plugin](https://github.com/hazendaz/makeself-maven-plugin). Please refer to project for usage and report any bugs in regards to maven plugin on that project. + +## License + +Makeself itself is covered by the [GNU General Public License][8] (GPL) version 2 and above. Archives generated by Makeself don't have to be placed under this license (although I encourage it ;-)), since the archive itself is merely data for Makeself. + +## Contributing + +I will gladly consider merging your pull requests on the [GitHub][10] repository. However, please keep the following in mind: + + * One of the main purposes of Makeself is portability. Do not submit patches that will break supported platforms. The more platform-agnostic, the better. + * Please explain clearly what the purpose of the patch is, and how you achieved it. + +## Download + +Get the latest official distribution [here][9] (version 2.4.2). + +The latest development version can be grabbed from [GitHub][10]. Feel free to submit any patches there through the fork and pull request process. + +## Version history + + * **v1.0:** Initial public release + * **v1.1:** The archive can be passed parameters that will be passed on to the embedded script, thanks to John C. Quillan + * **v1.2:** Cosmetic updates, support for bzip2 compression and non-temporary archives. Many ideas thanks to Francois Petitjean. + * **v1.3:** More patches from Bjarni R. Einarsson and Francois Petitjean: Support for no compression (`--nocomp`), script is no longer mandatory, automatic launch in an xterm, optional verbose output, and -target archive option to indicate where to extract the files. + * **v1.4:** Many patches from Francois Petitjean: improved UNIX compatibility, automatic integrity checking, support of LSM files to get info on the package at run time.. + * **v1.5.x:** A lot of bugfixes, and many other patches, including automatic verification through the usage of checksums. Version 1.5.5 was the stable release for a long time, even though the Web page didn't get updated ;-). Makeself was also officially made a part of the [Loki Setup installer][11], and its source is being maintained as part of this package. + * **v2.0:** Complete internal rewrite of Makeself. The command-line parsing was vastly improved, the overall maintenance of the package was greatly improved by separating the stub from makeself.sh. Also Makeself was ported and tested to a variety of Unix platforms. + * **v2.0.1:** First public release of the new 2.0 branch. Prior versions are officially obsoleted. This release introduced the `--copy` argument that was introduced in response to a need for the [UT2K3][12] Linux installer. + * **v2.1.0:** Big change : Makeself can now support multiple embedded tarballs, each stored separately with their own checksums. An existing archive can be updated with the `--append` flag. Checksums are also better managed, and the `--nochown` option for archives appeared. + * **v2.1.1:** Fixes related to the Unix compression (compress command). Some Linux distributions made the insane choice to make it unavailable, even though gzip is capable of uncompressing these files, plus some more bugfixes in the extraction and checksum code. + * **v2.1.2:** Some bug fixes. Use head -n to avoid problems with POSIX conformance. + * **v2.1.3:** Bug fixes with the command line when spawning terminals. Added `--tar`, `--noexec` for archives. Added `--nomd5` and `--nocrc` to avoid creating checksums in archives. The embedded script is now run through "eval". The `--info` output now includes the command used to create the archive. A man page was contributed by Bartosz Fenski. + * **v2.1.4:** Fixed `--info` output. Generate random directory name when extracting files to . to avoid problems. Better handling of errors with wrong permissions for the directory containing the files. Avoid some race conditions, Unset the $CDPATH variable to avoid problems if it is set. Better handling of dot files in the archive directory. + * **v2.1.5:** Made the md5sum detection consistent with the header code. Check for the presence of the archive directory. Added `--encrypt` for symmetric encryption through gpg (Eric Windisch). Added support for the digest command on Solaris 10 for MD5 checksums. Check for available disk space before extracting to the target directory (Andreas Schweitzer). Allow extraction to run asynchronously (patch by Peter Hatch). Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo). + * **v2.1.6:** Replaced one dot per file progress with a realtime progress percentage and a spinning cursor. Added `--noprogress` to prevent showing the progress during the decompression. Added `--target` dir to allow extracting directly to a target directory. (Guy Baconniere) + * **v2.2.0:** First major new release in years! Includes many bugfixes and user contributions. Please look at the [project page on Github][10] for all the details. + * **v2.3.0:** Support for archive encryption via GPG or OpenSSL. Added LZO and LZ4 compression support. Options to set the packaging date and stop the umask from being overriden. Optionally ignore check for available disk space when extracting. New option to check for root permissions before extracting. + * **v2.3.1:** Various compatibility updates. Added unit tests for Travis CI in the GitHub repo. New `--tar-extra`, `--untar-extra`, `--gpg-extra`, `--gpg-asymmetric-encrypt-sign` options. + * **v2.4.0:** Added optional support for SHA256 archive integrity checksums. + * **v2.4.2:** New --cleanup and --cleanup-args arguments for cleanup scripts. Added threading support for supported compressors. Now supports zstd compression. + * **v2.4.3:** Make explicit POSIX tar archives for increased compatibility. + * **v2.4.4:** Fixed various compatibility issues (no longer use POSIX tar archives), Github Actions to check on Solaris and FreeBSD. + * **v2.4.5:** Added `--tar-format` option to set the tar archive format (default is ustar) + +## Links + + * Check out the ["Loki Setup"][11] installer, used to install many Linux games and other applications, and of which I am the co-author. Since the demise of Loki, I am now the official maintainer of the project, and it is now being hosted here on GitHub. + * Bjarni R. Einarsson also wrote the **setup.sh** installer script, inspired by Makeself. [Check it out !][14] + +## Contact + +This script was written by [Stéphane Peter][15] (megastep at megastep.org). Any enhancements and suggestions are welcome. + +Contributions were included from John C. Quillan, Bjarni R. Einarsson, +Francois Petitjean, Ryan C. Gordon, and many contributors on GitHub. If you think I forgot +your name, don't hesitate to contact me. + +This project is now hosted on GitHub. Feel free to submit patches and bug reports on the [project page][10]. + +* * * + +[Stephane Peter][2] + + [1]: http://makeself.io/ + [2]: mailto:megastep@megastep.org + [3]: http://www.idsoftware.com/ + [4]: http://www.lokigames.com/products/myth2/updates.php3 + [5]: http://www.nvidia.com/ + [6]: http://earth.google.com/ + [7]: http://www.virtualbox.org/ + [8]: http://www.gnu.org/copyleft/gpl.html + [9]: https://github.com/megastep/makeself/releases/download/release-2.4.5/makeself-2.4.5.run + [10]: https://github.com/megastep/makeself + [11]: https://github.com/megastep/loki_setup/ + [12]: http://www.unrealtournament2003.com/ + [13]: http://www.icculus.org/ + [14]: http://bre.klaki.net/programs/setup.sh/ + [15]: https://stephanepeter.com/ diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/VERSION b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/VERSION new file mode 100755 index 000000000..59aa62c1f --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/VERSION @@ -0,0 +1 @@ +2.4.5 diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/make-release.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/make-release.sh new file mode 100755 index 000000000..b5692d490 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/make-release.sh @@ -0,0 +1,9 @@ +#!/bin/sh +# +# Create a distributable archive of the current version of Makeself + +VER=`cat VERSION` +mkdir -p /tmp/makeself-$VER release +cp -pPR makeself* test README.md COPYING VERSION .gitmodules /tmp/makeself-$VER/ +./makeself.sh --notemp /tmp/makeself-$VER release/makeself-$VER.run "Makeself v$VER" echo "Makeself has extracted itself" + diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself-header.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself-header.sh new file mode 100755 index 000000000..940903148 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself-header.sh @@ -0,0 +1,660 @@ +cat << EOF > "$archname" +#!/bin/bash +# This script was generated using Makeself $MS_VERSION +# The license covering this archive and its contents, if any, is wholly independent of the Makeself license (GPL) +# 2022.3.19-Modified the MS_Help function and some options +# Huawei Technologies Co., Ltd. + +ORIG_UMASK=\`umask\` + +CRCsum="$CRCsum" +MD5="$MD5sum" +SHA="$SHAsum" +SIGNATURE="$Signature" +TMPROOT=\${TMPDIR:="\$HOME"} +if ! test -d "\$TMPROOT"; then + TMPROOT="\$PWD" +fi +export TMPDIR="\$TMPROOT" +USER_PWD="\$PWD" +if ! test -d "\$USER_PWD"; then + exit 1 +fi +export USER_PWD +ARCHIVE_DIR=\`dirname "\$0"\` +export ARCHIVE_DIR + +name_of_file="\$0 " +pwd_of_file="\$PWD" +label="$LABEL" +script="$SCRIPT" +scriptargs="$SCRIPTARGS" +cleanup_script="${CLEANUP_SCRIPT}" +licensetxt="$LICENSE" +helpheader='$HELPHEADER' +targetdir="$archdirname" +filesizes="$filesizes" +totalsize="$totalsize" +keep="$KEEP" +nooverwrite="$NOOVERWRITE" +quiet="n" +accept="n" +nodiskspace="n" +export_conf="$EXPORT_CONF" +decrypt_cmd="$DECRYPT_CMD" +skip="$SKIP" + +print_cmd_arg="" +if type printf > /dev/null; then + print_cmd="printf" +elif test -x /usr/ucb/echo; then + print_cmd="/usr/ucb/echo" +else + print_cmd="echo" +fi + +if test -d /usr/xpg4/bin; then + PATH=/usr/xpg4/bin:\$PATH + export PATH +fi + +if test -d /usr/sfw/bin; then + PATH=\$PATH:/usr/sfw/bin + export PATH +fi + +unset CDPATH + +MS_Printf() +{ + \$print_cmd \$print_cmd_arg "\$1" +} + +MS_PrintLicense() +{ + PAGER=\${PAGER:=more} + if test x"\$licensetxt" != x; then + PAGER_PATH=\`exec <&- 2>&-; which \$PAGER || command -v \$PAGER || type \$PAGER\` + if test -x "\$PAGER_PATH"; then + echo "\$licensetxt" | \$PAGER + else + echo "\$licensetxt" + fi + if test x"\$accept" != xy; then + while true + do + MS_Printf "Please type y to accept, n otherwise: " + read yn + if test x"\$yn" = xn; then + keep=n + eval \$finish; exit 1 + break; + elif test x"\$yn" = xy; then + break; + fi + done + fi + fi +} + +MS_diskspace() +{ + ( + df -kP "\$1" | tail -1 | awk '{ if (\$4 ~ /%/) {print \$3} else {print \$4} }' + ) +} + +MS_dd() +{ + blocks=\`expr \$3 / 1024\` + bytes=\`expr \$3 % 1024\` + # Test for ibs, obs and conv feature + if dd if=/dev/zero of=/dev/null count=1 ibs=512 obs=512 conv=sync 2> /dev/null; then + dd if="\$1" ibs=\$2 skip=1 obs=1024 conv=sync 2> /dev/null | \\ + { test \$blocks -gt 0 && dd ibs=1024 obs=1024 count=\$blocks ; \\ + test \$bytes -gt 0 && dd ibs=1 obs=1024 count=\$bytes ; } 2> /dev/null + else + dd if="\$1" bs=\$2 skip=1 2> /dev/null + fi +} + +MS_dd_Progress() +{ + if test x"\$noprogress" = xy; then + MS_dd "\$@" + return \$? + fi + file="\$1" + offset=\$2 + length=\$3 + pos=0 + bsize=4194304 + while test \$bsize -gt \$length; do + bsize=\`expr \$bsize / 4\` + done + blocks=\`expr \$length / \$bsize\` + bytes=\`expr \$length % \$bsize\` + ( + dd ibs=\$offset skip=1 2>/dev/null + pos=\`expr \$pos \+ \$bsize\` + MS_Printf " 0%% " 1>&2 + if test \$blocks -gt 0; then + while test \$pos -le \$length; do + dd bs=\$bsize count=1 2>/dev/null + pcent=\`expr \$length / 100\` + pcent=\`expr \$pos / \$pcent\` + if test \$pcent -lt 100; then + MS_Printf "\b\b\b\b\b\b\b" 1>&2 + if test \$pcent -lt 10; then + MS_Printf " \$pcent%% " 1>&2 + else + MS_Printf " \$pcent%% " 1>&2 + fi + fi + pos=\`expr \$pos \+ \$bsize\` + done + fi + if test \$bytes -gt 0; then + dd bs=\$bytes count=1 2>/dev/null + fi + MS_Printf "\b\b\b\b\b\b\b" 1>&2 + MS_Printf " 100%% " 1>&2 + ) < "\$file" +} + +MS_Help() +{ + cat << EOH >&2 +Usage: \$0 [options] +Options: + --help | -h Print this message + --info Print embedded info : title, default target directory, embedded script ... + --list Print the list of files in the archive + --check Checks integrity and version dependency of the archive + --quiet Quiet install mode, skip human-computer interactions + --nox11 Do not spawn an xterm + --noexec Do not run embedded script + --extract= Extract directly to a target directory (absolute or relative) + Usually used with --noexec to just extract files without running + --tar arg1 [arg2 ...] Access the contents of the archive through the tar command +\${helpheader} +EOH +} + +MS_Verify_Sig() +{ + GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` + MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` + test -x "\$GPG_PATH" || GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` + test -x "\$MKTEMP_PATH" || MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` + offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` + temp_sig=\`mktemp -t XXXXX\` + echo \$SIGNATURE | base64 --decode > "\$temp_sig" + gpg_output=\`MS_dd "\$1" \$offset \$totalsize | LC_ALL=C "\$GPG_PATH" --verify "\$temp_sig" - 2>&1\` + gpg_res=\$? + rm -f "\$temp_sig" + if test \$gpg_res -eq 0 && test \`echo \$gpg_output | grep -c Good\` -eq 1; then + if test \`echo \$gpg_output | grep -c \$sig_key\` -eq 1; then + test x"\$quiet" = xn && echo "GPG signature is good" >&2 + else + echo "GPG Signature key does not match" >&2 + exit 2 + fi + else + test x"\$quiet" = xn && echo "GPG signature failed to verify" >&2 + exit 2 + fi +} + +MS_Check() +{ + OLD_PATH="\$PATH" + PATH=\${GUESS_MD5_PATH:-"\$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} + MD5_ARG="" + MD5_PATH=\`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum\` + test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which md5 || command -v md5 || type md5\` + test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which digest || command -v digest || type digest\` + PATH="\$OLD_PATH" + + SHA_PATH=\`exec <&- 2>&-; which shasum || command -v shasum || type shasum\` + test -x "\$SHA_PATH" || SHA_PATH=\`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum\` + + if test x"\$quiet" = xn; then + MS_Printf "Verifying archive integrity..." + fi + offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` + fsize=\`cat "\$1" | wc -c | tr -d " "\` + if test \$totalsize -ne \`expr \$fsize - \$offset\`; then + echo " Unexpected archive size." >&2 + exit 2 + fi + verb=\$2 + i=1 + for s in \$filesizes + do + crc=\`echo \$CRCsum | cut -d" " -f\$i\` + if test -x "\$SHA_PATH"; then + if test x"\`basename \$SHA_PATH\`" = xshasum; then + SHA_ARG="-a 256" + fi + sha=\`echo \$SHA | cut -d" " -f\$i\` + if test x"\$sha" = x0000000000000000000000000000000000000000000000000000000000000000; then + test x"\$verb" = xy && echo " \$1 does not contain an embedded SHA256 checksum." >&2 + else + shasum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$SHA_PATH \$SHA_ARG" | cut -b-64\`; + if test x"\$shasum" != x"\$sha"; then + echo "Error in SHA256 checksums: \$shasum is different from \$sha" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " SHA256 checksums are OK." >&2 + fi + crc="0000000000"; + fi + fi + if test -x "\$MD5_PATH"; then + if test x"\`basename \$MD5_PATH\`" = xdigest; then + MD5_ARG="-a md5" + fi + md5=\`echo \$MD5 | cut -d" " -f\$i\` + if test x"\$md5" = x00000000000000000000000000000000; then + test x"\$verb" = xy && echo " \$1 does not contain an embedded MD5 checksum." >&2 + else + md5sum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$MD5_PATH \$MD5_ARG" | cut -b-32\`; + if test x"\$md5sum" != x"\$md5"; then + echo "Error in MD5 checksums: \$md5sum is different from \$md5" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " MD5 checksums are OK." >&2 + fi + crc="0000000000"; verb=n + fi + fi + if test x"\$crc" = x0000000000; then + test x"\$verb" = xy && echo " \$1 does not contain a CRC checksum." >&2 + else + sum1=\`MS_dd_Progress "\$1" \$offset \$s | CMD_ENV=xpg4 cksum | awk '{print \$1}'\` + if test x"\$sum1" != x"\$crc"; then + echo "Error in checksums: \$sum1 is different from \$crc" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " CRC checksums are OK." >&2 + fi + fi + i=\`expr \$i + 1\` + offset=\`expr \$offset + \$s\` + done + if test x"\$quiet" = xn; then + echo " All good." + fi +} + +MS_Decompress() +{ + if test x"\$decrypt_cmd" != x""; then + { eval "\$decrypt_cmd" || echo " ... Decryption failed." >&2; } | eval "$GUNZIP_CMD" + else + eval "$GUNZIP_CMD" + fi + + if test \$? -ne 0; then + echo " ... Decompression failed." >&2 + fi +} + +UnTAR() +{ + if test x"\$quiet" = xn; then + tar \$1vf - $UNTAR_EXTRA 2>&1 || { echo " ... Extraction failed." >&2; kill -15 \$$; } + else + tar \$1f - $UNTAR_EXTRA 2>&1 || { echo Extraction failed. >&2; kill -15 \$$; } + fi +} + +MS_exec_cleanup() { + if test x"\$cleanup" = xy && test x"\$cleanup_script" != x""; then + cleanup=n + cd "\$tmpdir" + eval "\"\$cleanup_script\" \$scriptargs \$cleanupargs" + fi +} + +MS_cleanup() +{ + echo 'Signal caught, cleaning up' >&2 + MS_exec_cleanup + cd "\$TMPROOT" + rm -rf "\$tmpdir" + eval \$finish; exit 15 +} + +Script_Args_Check() +{ + script_supported_args=\$(echo \${helpheader} | grep -o -E "\-\-[^ ]+" | awk -F"=" {'print \$1'}) + arg_to_test=\$(echo \$1|awk -F"=" {'print \$1'}) + + for arg in \${script_supported_args}; + do + if test x"\$arg_to_test" = x"\$arg" ;then + return + fi + done + + MS_Help + exit 1 +} + +finish=true +xterm_loop= +noprogress=$NOPROGRESS +nox11=$NOX11 +copy=$COPY +ownership=$OWNERSHIP +verbose=n +cleanup=y +cleanupargs= +sig_key= + +initargs="\$@" + +while [ -n "\$*" ] +do + case "\$1" in + -h | --help) + MS_Help + exit 0 + ;; + -q | --quiet) + quiet=y + noprogress=y + shift + ;; + --info) + echo Identification: "\$label" + echo Target directory: "\$targetdir" + echo Uncompressed size: $USIZE KB + echo Compression: $COMPRESS + if test x"$ENCRYPT" != x""; then + echo Encryption: $ENCRYPT + fi + echo Date of packaging: $DATE + echo Built with Makeself version $MS_VERSION + echo Build command was: "$MS_COMMAND" + if test x"\$script" != x; then + echo Script run after extraction: + echo " " \$script \$scriptargs + fi + if test x"$copy" = xcopy; then + echo "Archive will copy itself to a temporary location" + fi + if test x"$NEED_ROOT" = xy; then + echo "Root permissions required for extraction" + fi + if test x"$KEEP" = xy; then + echo "directory \$targetdir is permanent" + else + echo "\$targetdir will be removed after extraction" + fi + exit 0 + ;; + --list) + echo Target directory: \$targetdir + offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + for s in \$filesizes + do + MS_dd "\$0" \$offset \$s | MS_Decompress | UnTAR t + offset=\`expr \$offset + \$s\` + done + exit 0 + ;; + --tar) + offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + arg1="\$2" + shift 2 || { MS_Help; exit 1; } + for s in \$filesizes + do + MS_dd "\$0" \$offset \$s | MS_Decompress | tar "\$arg1" - "\$@" + offset=\`expr \$offset + \$s\` + done + exit 0 + ;; + --check) + MS_Check "\$0" y + scriptargs="\$scriptargs \$1" + shift + ;; + --noexec) + script="" + cleanup_script="" + shift + ;; + --extract=*) + keep=y + targetdir=\`echo \$1 | cut -d"=" -f2 \` + if ! shift; then MS_Help; exit 1; fi + ;; + --nox11) + nox11=y + shift + ;; + --xwin) + if test "$NOWAIT" = n; then + finish="echo Press Return to close this window...; read junk" + fi + xterm_loop=1 + shift + ;; + --phase2) + copy=phase2 + shift + ;; + --repack | --repack-path=*) + Script_Args_Check \$1 + scriptargs="\$scriptargs '\$1'" + shift + if [[ ! "\$1" =~ ^-.* ]]; then + scriptargs="\$scriptargs '\$1'" + shift + fi + ;; + *) + Script_Args_Check \$1 + scriptargs="\$scriptargs '\$1'" + shift + ;; + esac +done + +quiet_para="" +if test x"\$quiet" = xy; then + quiet_para="--quiet " +fi +scriptargs="--\$name_of_file""--\"\$pwd_of_file\""" \$quiet_para""\$scriptargs" + +if test x"\$quiet" = xy -a x"\$verbose" = xy; then + echo Cannot be verbose and quiet at the same time. >&2 + exit 1 +fi + +if test x"$NEED_ROOT" = xy -a \`id -u\` -ne 0; then + echo "Administrative privileges required for this archive (use su or sudo)" >&2 + exit 1 +fi + +if test x"\$copy" \!= xphase2; then + MS_PrintLicense +fi + +case "\$copy" in +copy) + tmpdir="\$TMPROOT"/makeself.\$RANDOM.\`date +"%y%m%d%H%M%S"\`.\$\$ + mkdir "\$tmpdir" || { + echo "Could not create temporary directory \$tmpdir" >&2 + exit 1 + } + SCRIPT_COPY="\$tmpdir/makeself" + echo "Copying to a temporary location..." >&2 + cp "\$0" "\$SCRIPT_COPY" + chmod +x "\$SCRIPT_COPY" + cd "\$TMPROOT" + exec "\$SCRIPT_COPY" --phase2 -- \$initargs + ;; +phase2) + finish="\$finish ; rm -rf \`dirname \$0\`" + ;; +esac + +if test x"\$nox11" = xn; then + if tty -s; then # Do we have a terminal? + : + else + if test x"\$DISPLAY" != x -a x"\$xterm_loop" = x; then # No, but do we have X? + if xset q > /dev/null 2>&1; then # Check for valid DISPLAY variable + GUESS_XTERMS="xterm gnome-terminal rxvt dtterm eterm Eterm xfce4-terminal lxterminal kvt konsole aterm terminology" + for a in \$GUESS_XTERMS; do + if type \$a >/dev/null 2>&1; then + XTERM=\$a + break + fi + done + chmod a+x \$0 || echo Please add execution rights on \$0 + if test \`echo "\$0" | cut -c1\` = "/"; then # Spawn a terminal! + exec \$XTERM -e "\$0 --xwin \$initargs" + else + exec \$XTERM -e "./\$0 --xwin \$initargs" + fi + fi + fi + fi +fi + +if test x"\$targetdir" = x.; then + tmpdir="." +else + if test x"\$keep" = xy; then + if test x"\$nooverwrite" = xy && test -d "\$targetdir"; then + echo "Target directory \$targetdir already exists, aborting." >&2 + exit 1 + fi + if test x"\$quiet" = xn; then + echo "Creating directory \$targetdir" >&2 + fi + tmpdir="\$targetdir" + dashp="-p" + else + tmpdir="\$TMPROOT/selfgz\$\$\$RANDOM" + dashp="" + fi + mkdir \$dashp "\$tmpdir" || { + echo 'Cannot create target directory' \$tmpdir >&2 + echo 'You should try option --extract=' >&2 + eval \$finish + exit 1 + } +fi + +location="\`pwd\`" +if test x"\$SETUP_NOCHECK" != x1; then + MS_Check "\$0" +fi +offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + +if test x"\$verbose" = xy; then + MS_Printf "About to extract $USIZE KB in \$tmpdir ... Proceed ? [Y/n] " + read yn + if test x"\$yn" = xn; then + eval \$finish; exit 1 + fi +fi + +if test x"\$quiet" = xn; then + # Decrypting with openssl will ask for password, + # the prompt needs to start on new line + if test x"$ENCRYPT" = x"openssl"; then + echo "Decrypting and uncompressing \$label..." + else + MS_Printf "Uncompressing \$label" + fi +fi +res=3 +if test x"\$keep" = xn; then + trap MS_cleanup 1 2 3 15 +fi + +if test x"\$nodiskspace" = xn; then + leftspace=\`MS_diskspace "\$tmpdir"\` + if test -n "\$leftspace"; then + if test "\$leftspace" -lt $USIZE; then + echo + echo "Not enough space left in "\`dirname \$tmpdir\`" (\$leftspace KB) to decompress \$0 ($USIZE KB)" >&2 + if test x"\$keep" = xn; then + echo "Consider setting TMPDIR to a directory with more free space." + fi + eval \$finish; exit 1 + fi + fi +fi + +for s in \$filesizes +do + if MS_dd_Progress "\$0" \$offset \$s | MS_Decompress | ( cd "\$tmpdir"; umask \$ORIG_UMASK ; UnTAR xp ) 1>/dev/null; then + if test x"\$ownership" = xy; then + (cd "\$tmpdir"; chown -R \`id -u\` .; chgrp -R \`id -g\` .) + fi + else + echo >&2 + echo "Unable to decompress \$0" >&2 + eval \$finish; exit 1 + fi + offset=\`expr \$offset + \$s\` +done +if test x"\$quiet" = xn; then + echo +fi + +cd "\$tmpdir" +res=0 +if test x"\$script" != x; then + if test x"\$export_conf" = x"y"; then + MS_BUNDLE="\$0" + MS_LABEL="\$label" + MS_SCRIPT="\$script" + MS_SCRIPTARGS="\$scriptargs" + MS_ARCHDIRNAME="\$archdirname" + MS_KEEP="\$KEEP" + MS_NOOVERWRITE="\$NOOVERWRITE" + MS_COMPRESS="\$COMPRESS" + MS_CLEANUP="\$cleanup" + export MS_BUNDLE MS_LABEL MS_SCRIPT MS_SCRIPTARGS + export MS_ARCHDIRNAME MS_KEEP MS_NOOVERWRITE MS_COMPRESS + fi + + if test x"\$verbose" = x"y"; then + yn="x" + while test x"\$yn" != x -a x"\$yn" != xy -a x"\$yn" != xY -a x"\$yn" != xn -a x"\$yn" != xN + do + MS_Printf "OK to execute: \$script \$scriptargs \$* ? [Y/n] " + read yn + if test x"\$yn" = x -o x"\$yn" = xy -o x"\$yn" = xY; then + eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$?; + elif test x"\$yn" = xn -o x"\$yn" = xN; then + echo "Unable to decompress \$script ,because of aborting! ";res=\$? + else + echo "Input value is unacceptable,please try again." + fi + done + else + eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$? + fi + if test "\$res" -ne 0; then + test x"\$verbose" = xy && echo "The program '\$script' returned an error code (\$res)" >&2 + fi +fi + +MS_exec_cleanup + +if test x"\$keep" = xn; then + cd "\$TMPROOT" + rm -rf "\$tmpdir" +fi +eval \$finish; exit \$res +EOF diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.1 b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.1 new file mode 100755 index 000000000..81bf6e4ff --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.1 @@ -0,0 +1,110 @@ +.TH "MAKESELF" "1" "2.4.5" +.SH "NAME" +makeself \- An utility to generate self-extractable archives. +.SH "SYNTAX" +.B makeself [\fIoptions\fP] archive_dir file_name label +.B [\fIstartup_script\fP] [\fIargs\fP] +.SH "DESCRIPTION" +This program is a free (GPL) utility designed to create self-extractable +archives from a directory. +.SH "OPTIONS" +The following options are supported. +.TP 15 +.B -v, --version +Prints out the makeself version number and exits. +.TP +.B -h, --help +Print out help information. +.TP +.B --tar-quietly +Suppress verbose output from the tar command +.TP +.B --quiet +Do not print any messages other than errors +.TP +.B --gzip +Compress using gzip (default if detected). +.TP +.B --bzip2 +Compress using bzip2. +.TP +.B --pbzip2 +Compress using pbzip2. +.TP +.B --xz +Compress using xz. +.TP +.B --lzo +Compress using lzop. +.TP +.B --lz4 +Compress using lz4. +.TP +.B --compress +Compress using the UNIX 'compress' command. +.TP +.B --nocomp +Do not compress the data. +.TP +.B --complevel lvl +Specify the compression level for gzip,bzip2,pbzui2,xz,lzo or lz4 +.TP +.B --notemp +The archive will create archive_dir in the current directory and +uncompress in ./archive_dir. +.TP +.B --copy +Upon extraction, the archive will first copy itself to a temporary directory. +.TP +.B --append +Append more files to an existing makeself archive. The label and startup scripts will then be ignored. +.TP +.B --current +Files will be extracted to the current directory. Both --current and --target dir imply --notemp. +.TP +.B --target dir +Extract directly to a target directory. Directory path can be either absolute or relative. +.TP +.B --header file +Specify location of the header script. +.TP +.B --cleanup file +Specify a cleanup script that executes on interrupt and when finished successfully. +.TP +.B --follow +Follow the symlinks in the archive. +.TP +.B --noprogress +Do not show the progress during the decompression. +.TP +.B --nox11 +Disable automatic spawn of an xterm if running in X11. +.TP +.B --nowait +Do not wait for user input after executing embedded program from an xterm. +.TP +.B --nomd5 +Do not create a MD5 checksum for the archive. +.TP +.B --nocrc +Do not create a CRC32 checksum for the archive. +.TP +.B --lsm file +LSM file describing the package. +.B --packaging-date date +Use provided string as the packaging date instead of the current date. +.SH "EXAMPLES" +Here is an example, assuming the user has a package image stored in a /home/joe/mysoft, +and he wants to generate a self-extracting package named mysoft.sh, which will launch +the "setup" script initially stored in /home/joe/mysoft: +.TP +makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup +.TP +Here is also how I created the makeself.run archive which contains the Makeself distribution: +.TP +makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" +.SH "AUTHORS" +Makeself has been written by Stéphane Peter . +.BR +This man page was originally written by Bartosz Fenski for the +Debian GNU/Linux distribution (but it may be used by others). diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.lsm b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.lsm new file mode 100755 index 000000000..3c4cea8c1 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.lsm @@ -0,0 +1,16 @@ +Begin3 +Title: makeself.sh +Version: 2.4.5 +Description: makeself.sh is a shell script that generates a self-extractable + tar.gz archive from a directory. The resulting file appears as a shell + script, and can be launched as is. The archive will then uncompress + itself to a temporary directory and an arbitrary command will be + executed (for example an installation script). This is pretty similar + to archives generated with WinZip Self-Extractor in the Windows world. +Keywords: Installation archive tar winzip +Author: Stephane Peter (megastep@megastep.org) +Maintained-by: Stephane Peter (megastep@megastep.org) +Original-site: https://makeself.io/ +Platform: Unix +Copying-policy: GPL +End diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.sh new file mode 100755 index 000000000..c8ea56597 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.sh @@ -0,0 +1,822 @@ +#!/bin/sh +# +# Makeself version 2.4.x +# by Stephane Peter +# +# Utility to create self-extracting tar.gz archives. +# The resulting archive is a file holding the tar.gz archive with +# a small Shell script stub that uncompresses the archive to a temporary +# directory and then executes a given script from withing that directory. +# +# Makeself home page: https://makeself.io/ +# +# Version 2.0 is a rewrite of version 1.0 to make the code easier to read and maintain. +# +# Version history : +# - 1.0 : Initial public release +# - 1.1 : The archive can be passed parameters that will be passed on to +# the embedded script, thanks to John C. Quillan +# - 1.2 : Package distribution, bzip2 compression, more command line options, +# support for non-temporary archives. Ideas thanks to Francois Petitjean +# - 1.3 : More patches from Bjarni R. Einarsson and Francois Petitjean: +# Support for no compression (--nocomp), script is no longer mandatory, +# automatic launch in an xterm, optional verbose output, and -target +# archive option to indicate where to extract the files. +# - 1.4 : Improved UNIX compatibility (Francois Petitjean) +# Automatic integrity checking, support of LSM files (Francois Petitjean) +# - 1.5 : Many bugfixes. Optionally disable xterm spawning. +# - 1.5.1 : More bugfixes, added archive options -list and -check. +# - 1.5.2 : Cosmetic changes to inform the user of what's going on with big +# archives (Quake III demo) +# - 1.5.3 : Check for validity of the DISPLAY variable before launching an xterm. +# More verbosity in xterms and check for embedded command's return value. +# Bugfix for Debian 2.0 systems that have a different "print" command. +# - 1.5.4 : Many bugfixes. Print out a message if the extraction failed. +# - 1.5.5 : More bugfixes. Added support for SETUP_NOCHECK environment variable to +# bypass checksum verification of archives. +# - 1.6.0 : Compute MD5 checksums with the md5sum command (patch from Ryan Gordon) +# - 2.0 : Brand new rewrite, cleaner architecture, separated header and UNIX ports. +# - 2.0.1 : Added --copy +# - 2.1.0 : Allow multiple tarballs to be stored in one archive, and incremental updates. +# Added --nochown for archives +# Stopped doing redundant checksums when not necesary +# - 2.1.1 : Work around insane behavior from certain Linux distros with no 'uncompress' command +# Cleaned up the code to handle error codes from compress. Simplified the extraction code. +# - 2.1.2 : Some bug fixes. Use head -n to avoid problems. +# - 2.1.3 : Bug fixes with command line when spawning terminals. +# Added --tar for archives, allowing to give arbitrary arguments to tar on the contents of the archive. +# Added --noexec to prevent execution of embedded scripts. +# Added --nomd5 and --nocrc to avoid creating checksums in archives. +# Added command used to create the archive in --info output. +# Run the embedded script through eval. +# - 2.1.4 : Fixed --info output. +# Generate random directory name when extracting files to . to avoid problems. (Jason Trent) +# Better handling of errors with wrong permissions for the directory containing the files. (Jason Trent) +# Avoid some race conditions (Ludwig Nussel) +# Unset the $CDPATH variable to avoid problems if it is set. (Debian) +# Better handling of dot files in the archive directory. +# - 2.1.5 : Made the md5sum detection consistent with the header code. +# Check for the presence of the archive directory +# Added --encrypt for symmetric encryption through gpg (Eric Windisch) +# Added support for the digest command on Solaris 10 for MD5 checksums +# Check for available disk space before extracting to the target directory (Andreas Schweitzer) +# Allow extraction to run asynchronously (patch by Peter Hatch) +# Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo) +# - 2.1.6 : Replaced one dot per file progress with a realtime progress percentage and a spining cursor (Guy Baconniere) +# Added --noprogress to prevent showing the progress during the decompression (Guy Baconniere) +# Added --target dir to allow extracting directly to a target directory (Guy Baconniere) +# - 2.2.0 : Many bugfixes, updates and contributions from users. Check out the project page on Github for the details. +# - 2.3.0 : Option to specify packaging date to enable byte-for-byte reproducibility. (Marc Pawlowsky) +# - 2.4.0 : Optional support for SHA256 checksums in archives. +# - 2.4.2 : Add support for threads for several compressors. (M. Limber) +# Added zstd support. +# - 2.4.3 : Make explicit POSIX tar archives for increased compatibility. +# - 2.4.5 : Added --tar-format to override ustar tar archive format +# +# (C) 1998-2021 by Stephane Peter +# +# This software is released under the terms of the GNU GPL version 2 and above +# Please read the license at http://www.gnu.org/copyleft/gpl.html +# Self-extracting archives created with this script are explictly NOT released under the term of the GPL +# + +MS_VERSION=2.4.5 +MS_COMMAND="$0" +unset CDPATH + +for f in ${1+"$@"}; do + MS_COMMAND="$MS_COMMAND \\\\ + \\\"$f\\\"" +done + +# For Solaris systems +if test -d /usr/xpg4/bin; then + PATH=/usr/xpg4/bin:$PATH + export PATH +fi + +# Procedures + +MS_Usage() +{ + echo "Usage: $0 [args] archive_dir file_name label startup_script [script_args]" + echo "args can be one or more of the following :" + echo " --version | -v : Print out Makeself version number and exit" + echo " --help | -h : Print out this help message" + echo " --tar-quietly : Suppress verbose output from the tar command" + echo " --quiet | -q : Do not print any messages other than errors." + echo " --gzip : Compress using gzip (default if detected)" + echo " --pigz : Compress with pigz" + echo " --zstd : Compress with zstd" + echo " --bzip2 : Compress using bzip2 instead of gzip" + echo " --pbzip2 : Compress using pbzip2 instead of gzip" + echo " --xz : Compress using xz instead of gzip" + echo " --lzo : Compress using lzop instead of gzip" + echo " --lz4 : Compress using lz4 instead of gzip" + echo " --compress : Compress using the UNIX 'compress' command" + echo " --complevel lvl : Compression level for gzip pigz zstd xz lzo lz4 bzip2 and pbzip2 (default 9)" + echo " --threads thds : Number of threads to be used by compressors that support parallelization." + echo " Omit to use compressor's default. Most useful (and required) for opting" + echo " into xz's threading, usually with '--threads=0' for all available cores." + echo " pbzip2 and pigz are parallel by default, and setting this value allows" + echo " limiting the number of threads they use." + echo " --base64 : Instead of compressing, encode the data using base64" + echo " --gpg-encrypt : Instead of compressing, encrypt the data using GPG" + echo " --gpg-asymmetric-encrypt-sign" + echo " : Instead of compressing, asymmetrically encrypt and sign the data using GPG" + echo " --gpg-extra opt : Append more options to the gpg command line" + echo " --ssl-encrypt : Instead of compressing, encrypt the data using OpenSSL" + echo " --ssl-passwd pass : Use the given password to encrypt the data using OpenSSL" + echo " --ssl-pass-src src : Use the given src as the source of password to encrypt the data" + echo " using OpenSSL. See \"PASS PHRASE ARGUMENTS\" in man openssl." + echo " If this option is not supplied, the user will be asked to enter" + echo " encryption password on the current terminal." + echo " --ssl-no-md : Do not use \"-md\" option not supported by older OpenSSL." + echo " --nochown : Do not give the target folder to the current user (default)" + echo " --chown : Give the target folder to the current user recursively" + echo " --nocomp : Do not compress the data" + echo " --notemp : The archive will create archive_dir in the" + echo " current directory and uncompress in ./archive_dir" + echo " --needroot : Check that the root user is extracting the archive before proceeding" + echo " --copy : Upon extraction, the archive will first copy itself to" + echo " a temporary directory" + echo " --append : Append more files to an existing Makeself archive" + echo " The label and startup scripts will then be ignored" + echo " --target dir : Extract directly to a target directory" + echo " directory path can be either absolute or relative" + echo " --nooverwrite : Do not extract the archive if the specified target directory exists" + echo " --current : Files will be extracted to the current directory" + echo " Both --current and --target imply --notemp" + echo " --tar-format opt : Specify a tar archive format (default is ustar)" + echo " --tar-extra opt : Append more options to the tar command line" + echo " --untar-extra opt : Append more options to the during the extraction of the tar archive" + echo " --nomd5 : Don't calculate an MD5 for archive" + echo " --nocrc : Don't calculate a CRC for archive" + echo " --sha256 : Compute a SHA256 checksum for the archive" + echo " --header file : Specify location of the header script" + echo " --cleanup file : Specify a cleanup script that executes on interrupt and when finished successfully." + echo " --follow : Follow the symlinks in the archive" + echo " --noprogress : Do not show the progress during the decompression" + echo " --nox11 : Disable automatic spawn of a xterm" + echo " --nowait : Do not wait for user input after executing embedded" + echo " program from an xterm" + echo " --sign passphrase : Signature private key to sign the package with" + echo " --lsm file : LSM file describing the package" + echo " --license file : Append a license file" + echo " --help-header file : Add a header to the archive's --help output" + echo " --packaging-date date" + echo " : Use provided string as the packaging date" + echo " instead of the current date." + echo + echo " --keep-umask : Keep the umask set to shell default, rather than overriding when executing self-extracting archive." + echo " --export-conf : Export configuration variables to startup_script" + echo + echo "Do not forget to give a fully qualified startup script name" + echo "(i.e. with a ./ prefix if inside the archive)." + exit 1 +} + +# Default settings +if type gzip >/dev/null 2>&1; then + COMPRESS=gzip +elif type compress >/dev/null 2>&1; then + COMPRESS=compress +else + echo "ERROR: missing commands: gzip, compress" >&2 + MS_Usage +fi +ENCRYPT=n +PASSWD="" +PASSWD_SRC="" +OPENSSL_NO_MD=n +COMPRESS_LEVEL=9 +DEFAULT_THREADS=123456 # Sentinel value +THREADS=$DEFAULT_THREADS +KEEP=n +CURRENT=n +NOX11=n +NOWAIT=n +APPEND=n +TAR_QUIETLY=n +KEEP_UMASK=n +QUIET=n +NOPROGRESS=n +COPY=none +NEED_ROOT=n +TAR_ARGS=rvf +TAR_FORMAT=ustar +TAR_EXTRA="" +GPG_EXTRA="" +DU_ARGS=-ks +HEADER=`dirname "$0"`/makeself-header.sh +SIGNATURE="" +TARGETDIR="" +NOOVERWRITE=n +DATE=`LC_ALL=C date` +EXPORT_CONF=n +SHA256=n +OWNERSHIP=n +SIGN=n +GPG_PASSPHRASE="" + +# LSM file stuff +LSM_CMD="echo No LSM. >> \"\$archname\"" + +while true +do + case "$1" in + --version | -v) + echo Makeself version $MS_VERSION + exit 0 + ;; + --pbzip2) + COMPRESS=pbzip2 + shift + ;; + --bzip2) + COMPRESS=bzip2 + shift + ;; + --gzip) + COMPRESS=gzip + shift + ;; + --pigz) + COMPRESS=pigz + shift + ;; + --zstd) + COMPRESS=zstd + shift + ;; + --xz) + COMPRESS=xz + shift + ;; + --lzo) + COMPRESS=lzo + shift + ;; + --lz4) + COMPRESS=lz4 + shift + ;; + --compress) + COMPRESS=compress + shift + ;; + --base64) + COMPRESS=base64 + shift + ;; + --gpg-encrypt) + COMPRESS=gpg + shift + ;; + --gpg-asymmetric-encrypt-sign) + COMPRESS=gpg-asymmetric + shift + ;; + --gpg-extra) + GPG_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-encrypt) + ENCRYPT=openssl + shift + ;; + --ssl-passwd) + PASSWD=$2 + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-pass-src) + PASSWD_SRC=$2 + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-no-md) + OPENSSL_NO_MD=y + shift + ;; + --nocomp) + COMPRESS=none + shift + ;; + --complevel) + COMPRESS_LEVEL="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --threads) + THREADS="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --nochown) + OWNERSHIP=n + shift + ;; + --chown) + OWNERSHIP=y + shift + ;; + --notemp) + KEEP=y + shift + ;; + --copy) + COPY=copy + shift + ;; + --current) + CURRENT=y + KEEP=y + shift + ;; + --tar-format) + TAR_FORMAT="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --tar-extra) + TAR_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --untar-extra) + UNTAR_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --target) + TARGETDIR="$2" + KEEP=y + shift 2 || { MS_Usage; exit 1; } + ;; + --sign) + SIGN=y + GPG_PASSPHRASE="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --nooverwrite) + NOOVERWRITE=y + shift + ;; + --needroot) + NEED_ROOT=y + shift + ;; + --header) + HEADER="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --cleanup) + CLEANUP_SCRIPT="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --license) + # We need to escape all characters having a special meaning in double quotes + LICENSE=$(sed 's/\\/\\\\/g; s/"/\\\"/g; s/`/\\\`/g; s/\$/\\\$/g' "$2") + shift 2 || { MS_Usage; exit 1; } + ;; + --follow) + TAR_ARGS=rvhf + DU_ARGS=-ksL + shift + ;; + --noprogress) + NOPROGRESS=y + shift + ;; + --nox11) + NOX11=y + shift + ;; + --nowait) + NOWAIT=y + shift + ;; + --nomd5) + NOMD5=y + shift + ;; + --sha256) + SHA256=y + shift + ;; + --nocrc) + NOCRC=y + shift + ;; + --append) + APPEND=y + shift + ;; + --lsm) + LSM_CMD="cat \"$2\" >> \"\$archname\"" + shift 2 || { MS_Usage; exit 1; } + ;; + --packaging-date) + DATE="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --help-header) + HELPHEADER=`sed -e "s/'/'\\\\\''/g" $2` + shift 2 || { MS_Usage; exit 1; } + [ -n "$HELPHEADER" ] && HELPHEADER="$HELPHEADER +" + ;; + --tar-quietly) + TAR_QUIETLY=y + shift + ;; + --keep-umask) + KEEP_UMASK=y + shift + ;; + --export-conf) + EXPORT_CONF=y + shift + ;; + -q | --quiet) + QUIET=y + shift + ;; + -h | --help) + MS_Usage + ;; + -*) + echo Unrecognized flag : "$1" + MS_Usage + ;; + *) + break + ;; + esac +done + +if test $# -lt 1; then + MS_Usage +else + if test -d "$1"; then + archdir="$1" + else + echo "Directory $1 does not exist." >&2 + exit 1 + fi +fi +archname="$2" + +if test "$QUIET" = "y" || test "$TAR_QUIETLY" = "y"; then + if test "$TAR_ARGS" = "rvf"; then + TAR_ARGS="rf" + elif test "$TAR_ARGS" = "rvhf"; then + TAR_ARGS="rhf" + fi +fi + +if test "$APPEND" = y; then + if test $# -lt 2; then + MS_Usage + fi + + # Gather the info from the original archive + OLDENV=`sh "$archname" --dumpconf` + if test $? -ne 0; then + echo "Unable to update archive: $archname" >&2 + exit 1 + else + eval "$OLDENV" + OLDSKIP=`expr $SKIP + 1` + fi +else + if test "$KEEP" = n -a $# = 3; then + echo "ERROR: Making a temporary archive with no embedded command does not make sense!" >&2 + echo >&2 + MS_Usage + fi + # We don't want to create an absolute directory unless a target directory is defined + if test "$CURRENT" = y; then + archdirname="." + elif test x"$TARGETDIR" != x; then + archdirname="$TARGETDIR" + else + archdirname=`basename "$1"` + fi + + if test $# -lt 3; then + MS_Usage + fi + + LABEL="$3" + SCRIPT="$4" + test "x$SCRIPT" = x || shift 1 + shift 3 + SCRIPTARGS="$*" +fi + +if test "$KEEP" = n -a "$CURRENT" = y; then + echo "ERROR: It is A VERY DANGEROUS IDEA to try to combine --notemp and --current." >&2 + exit 1 +fi + +case $COMPRESS in +gzip) + GZIP_CMD="gzip -c$COMPRESS_LEVEL" + GUNZIP_CMD="gzip -cd" + ;; +pigz) + GZIP_CMD="pigz -$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD --processes $THREADS" + fi + GUNZIP_CMD="gzip -cd" + ;; +zstd) + GZIP_CMD="zstd -$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD --threads=$THREADS" + fi + GUNZIP_CMD="zstd -cd" + ;; +pbzip2) + GZIP_CMD="pbzip2 -c$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD -p$THREADS" + fi + GUNZIP_CMD="bzip2 -d" + ;; +bzip2) + GZIP_CMD="bzip2 -$COMPRESS_LEVEL" + GUNZIP_CMD="bzip2 -d" + ;; +xz) + GZIP_CMD="xz -c$COMPRESS_LEVEL" + # Must opt-in by specifying a value since not all versions of xz support threads + if test $THREADS -ne $DEFAULT_THREADS; then + GZIP_CMD="$GZIP_CMD --threads=$THREADS" + fi + GUNZIP_CMD="xz -d" + ;; +lzo) + GZIP_CMD="lzop -c$COMPRESS_LEVEL" + GUNZIP_CMD="lzop -d" + ;; +lz4) + GZIP_CMD="lz4 -c$COMPRESS_LEVEL" + GUNZIP_CMD="lz4 -d" + ;; +base64) + GZIP_CMD="base64" + GUNZIP_CMD="base64 --decode -i -" + ;; +gpg) + GZIP_CMD="gpg $GPG_EXTRA -ac -z$COMPRESS_LEVEL" + GUNZIP_CMD="gpg -d" + ENCRYPT="gpg" + ;; +gpg-asymmetric) + GZIP_CMD="gpg $GPG_EXTRA -z$COMPRESS_LEVEL -es" + GUNZIP_CMD="gpg --yes -d" + ENCRYPT="gpg" + ;; +compress) + GZIP_CMD="compress -fc" + GUNZIP_CMD="(type compress >/dev/null 2>&1 && compress -fcd || gzip -cd)" + ;; +none) + GZIP_CMD="cat" + GUNZIP_CMD="cat" + ;; +esac + +if test x"$ENCRYPT" = x"openssl"; then + if test x"$APPEND" = x"y"; then + echo "Appending to existing archive is not compatible with OpenSSL encryption." >&2 + fi + + ENCRYPT_CMD="openssl enc -aes-256-cbc -salt" + DECRYPT_CMD="openssl enc -aes-256-cbc -d" + + if test x"$OPENSSL_NO_MD" != x"y"; then + ENCRYPT_CMD="$ENCRYPT_CMD -md sha256" + DECRYPT_CMD="$DECRYPT_CMD -md sha256" + fi + + if test -n "$PASSWD_SRC"; then + ENCRYPT_CMD="$ENCRYPT_CMD -pass $PASSWD_SRC" + elif test -n "$PASSWD"; then + ENCRYPT_CMD="$ENCRYPT_CMD -pass pass:$PASSWD" + fi +fi + +tmpfile="${TMPDIR:-/tmp}/mkself$$" + +if test -f "$HEADER"; then + oldarchname="$archname" + archname="$tmpfile" + # Generate a fake header to count its lines + SKIP=0 + . "$HEADER" + SKIP=`cat "$tmpfile" |wc -l` + # Get rid of any spaces + SKIP=`expr $SKIP` + rm -f "$tmpfile" + if test "$QUIET" = "n"; then + echo "Header is $SKIP lines long" >&2 + fi + archname="$oldarchname" +else + echo "Unable to open header file: $HEADER" >&2 + exit 1 +fi + +if test "$QUIET" = "n"; then + echo +fi + +if test "$APPEND" = n; then + if test -f "$archname"; then + echo "WARNING: Overwriting existing file: $archname" >&2 + fi +fi + +USIZE=`du $DU_ARGS "$archdir" | awk '{print $1}'` + +if test "." = "$archdirname"; then + if test "$KEEP" = n; then + archdirname="makeself-$$-`date +%Y%m%d%H%M%S`" + fi +fi + +test -d "$archdir" || { echo "Error: $archdir does not exist."; rm -f "$tmpfile"; exit 1; } +if test "$QUIET" = "n"; then + echo "About to compress $USIZE KB of data..." + echo "Adding files to archive named \"$archname\"..." +fi + +# See if we have GNU tar +TAR=`exec <&- 2>&-; which gtar || command -v gtar || type gtar` +test -x "$TAR" || TAR=tar + +tmparch="${TMPDIR:-/tmp}/mkself$$.tar" +( + if test "$APPEND" = "y"; then + tail -n "+$OLDSKIP" "$archname" | eval "$GUNZIP_CMD" > "$tmparch" + fi + cd "$archdir" + # "Determining if a directory is empty" + # https://www.etalabs.net/sh_tricks.html + find . \ + \( \ + ! -type d \ + -o \ + \( -links 2 -exec sh -c ' + is_empty () ( + cd "$1" + set -- .[!.]* ; test -f "$1" && return 1 + set -- ..?* ; test -f "$1" && return 1 + set -- * ; test -f "$1" && return 1 + return 0 + ) + is_empty "$0"' {} \; \ + \) \ + \) -print \ + | LC_ALL=C sort \ + | sed 's/./\\&/g' \ + | xargs $TAR $TAR_EXTRA --format $TAR_FORMAT -$TAR_ARGS "$tmparch" +) || { + echo "ERROR: failed to create temporary archive: $tmparch" + rm -f "$tmparch" "$tmpfile" + exit 1 +} + +USIZE=`du $DU_ARGS "$tmparch" | awk '{print $1}'` + +eval "$GZIP_CMD" <"$tmparch" >"$tmpfile" || { + echo "ERROR: failed to create temporary file: $tmpfile" + rm -f "$tmparch" "$tmpfile" + exit 1 +} +rm -f "$tmparch" + +if test x"$ENCRYPT" = x"openssl"; then + echo "About to encrypt archive \"$archname\"..." + { eval "$ENCRYPT_CMD -in $tmpfile -out ${tmpfile}.enc" && mv -f ${tmpfile}.enc $tmpfile; } || \ + { echo Aborting: could not encrypt temporary file: "$tmpfile".; rm -f "$tmpfile"; exit 1; } +fi + +fsize=`cat "$tmpfile" | wc -c | tr -d " "` + +# Compute the checksums + +shasum=0000000000000000000000000000000000000000000000000000000000000000 +md5sum=00000000000000000000000000000000 +crcsum=0000000000 + +if test "$NOCRC" = y; then + if test "$QUIET" = "n"; then + echo "skipping crc at user request" + fi +else + crcsum=`CMD_ENV=xpg4 cksum < "$tmpfile" | sed -e 's/ /Z/' -e 's/ /Z/' | cut -dZ -f1` + if test "$QUIET" = "n"; then + echo "CRC: $crcsum" + fi +fi + +if test "$SHA256" = y; then + SHA_PATH=`exec <&- 2>&-; which shasum || command -v shasum || type shasum` + if test -x "$SHA_PATH"; then + shasum=`eval "$SHA_PATH -a 256" < "$tmpfile" | cut -b-64` + else + SHA_PATH=`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum` + shasum=`eval "$SHA_PATH" < "$tmpfile" | cut -b-64` + fi + if test "$QUIET" = "n"; then + if test -x "$SHA_PATH"; then + echo "SHA256: $shasum" + else + echo "SHA256: none, SHA command not found" + fi + fi +fi +if test "$NOMD5" = y; then + if test "$QUIET" = "n"; then + echo "Skipping md5sum at user request" + fi +else + # Try to locate a MD5 binary + OLD_PATH=$PATH + PATH=${GUESS_MD5_PATH:-"$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} + MD5_ARG="" + MD5_PATH=`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum` + test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which md5 || command -v md5 || type md5` + test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which digest || command -v digest || type digest` + PATH=$OLD_PATH + if test -x "$MD5_PATH"; then + if test `basename ${MD5_PATH}`x = digestx; then + MD5_ARG="-a md5" + fi + md5sum=`eval "$MD5_PATH $MD5_ARG" < "$tmpfile" | cut -b-32` + if test "$QUIET" = "n"; then + echo "MD5: $md5sum" + fi + else + if test "$QUIET" = "n"; then + echo "MD5: none, MD5 command not found" + fi + fi +fi +if test "$SIGN" = y; then + GPG_PATH=`exec <&- 2>&-; which gpg || command -v gpg || type gpg` + if test -x "$GPG_PATH"; then + SIGNATURE=`$GPG_PATH --pinentry-mode=loopback --batch --yes --passphrase "$GPG_PASSPHRASE" --output - --detach-sig $tmpfile | base64 | tr -d \\\\n` + if test "$QUIET" = "n"; then + echo "Signature: $SIGNATURE" + fi + else + echo "Missing gpg command" >&2 + fi +fi + +totalsize=0 +for size in $fsize; +do + totalsize=`expr $totalsize + $size` +done + +if test "$APPEND" = y; then + mv "$archname" "$archname".bak || exit + + # Prepare entry for new archive + filesizes="$fsize" + CRCsum="$crcsum" + MD5sum="$md5sum" + SHAsum="$shasum" + Signature="$SIGNATURE" + # Generate the header + . "$HEADER" + # Append the new data + cat "$tmpfile" >> "$archname" + + chmod +x "$archname" + rm -f "$archname".bak + if test "$QUIET" = "n"; then + echo "Self-extractable archive \"$archname\" successfully updated." + fi +else + filesizes="$fsize" + CRCsum="$crcsum" + MD5sum="$md5sum" + SHAsum="$shasum" + Signature="$SIGNATURE" + + # Generate the header + . "$HEADER" + + # Append the compressed tar data after the stub + if test "$QUIET" = "n"; then + echo + fi + cat "$tmpfile" >> "$archname" + chmod +x "$archname" + if test "$QUIET" = "n"; then + echo Self-extractable archive \"$archname\" successfully created. + fi +fi +rm -f "$tmpfile" diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/run-tests.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/run-tests.sh new file mode 100755 index 000000000..31ee16511 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/run-tests.sh @@ -0,0 +1,8 @@ +#!/bin/sh +# Run every available test - Bash needed +cd test +for test in *test; +do + echo "Running test $test ..." + bash $test || { echo "*** ERROR: Test '$test' failed!"; exit 1; } +done diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/merge_aicpu_info_json.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/merge_aicpu_info_json.sh new file mode 100755 index 000000000..a977bd51d --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/merge_aicpu_info_json.sh @@ -0,0 +1,31 @@ +#!/bin/bash + +project_path=$1 +build_path=$2 +vendor_name=customize +echo $@ +if [[ ! -d "$project_path" ]]; then + echo "[ERROR] No projcet path is provided" + exit 1 +fi + +if [[ ! -d "$build_path" ]]; then + echo "[ERROR] No build path is provided" + exit 1 +fi + +if [[ ! -d "$ASCEND_OPP_PATH" ]]; then + echo "[ERROR] No opp install path is provided" + exit 1 +fi +custom_exist_info_json=$ASCEND_OPP_PATH/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json +custom_new_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json +temp_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/temp_cust_aicpu_kernel.json + +if [[ -f "$custom_exist_info_json" ]] && [[ -f "$custom_new_info_json" ]]; then + cp -f $custom_exist_info_json $temp_info_json + chmod +w $temp_info_json + python3 ${project_path}/cmake/util/insert_op_info.py ${custom_new_info_json} ${temp_info_json} + cp -f $temp_info_json $custom_new_info_json + rm -f $temp_info_json +fi diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/opdesc_parser.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/opdesc_parser.py new file mode 100755 index 000000000..bc39462b8 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/opdesc_parser.py @@ -0,0 +1,277 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os + + +OP_ALL = '__ALLOP__' +SOC_ALL = '__ALLSOC__' +SOC_TO_SHORT_SOC_MAP = { + "ascend910a": "ascend910", + "ascend910proa": "ascend910", + "ascend910b": "ascend910", + "ascend910prob": "ascend910", + "ascend910premiuma": "ascend910", + "ascend910b1": "ascend910b", + "ascend910b2": "ascend910b", + "ascend910b2c": "ascend910b", + "ascend910b3": "ascend910b", + "ascend910b4": "ascend910b", + "ascend910c1": "ascend910c", + "ascend910c2": "ascend910c", + "ascend910c3": "ascend910c", + "ascend910c4": "ascend910c", + "ascend310p1": "ascend310p", + "ascend310p3": "ascend310p", + "ascend310p3vir01": "ascend310p", + "ascend310p3vir02": "ascend310p", + "ascend310p3vir04": "ascend310p", + "ascend310p3vir08": "ascend310p", + "ascend310b1": "ascend310b", + "bs9sx1aa": "bs9sx1a" +} + + +class OpDesc: + def __init__(self: any, op_type: str): + self.op_type = op_type + self.attr_list = [] + self.attr_val = {} + self.input_name = [] + self.input_type = [] + self.input_dtype = [] + self.input_fmt = [] + self.output_name = [] + self.output_type = [] + self.output_dtype = [] + self.output_fmt = [] + self.op_fmt_sel = False + self.op_chk_support = False + self.op_intf = '' + self.kern_name = '' + self.op_file = '' + self.op_replay_flag = False + self.op_replay_batch = False + self.input_idx = -1 + self.output_idx = -1 + self.max_block_dim = 32 + self.max_shape_size = 268435456 + self.dynamic_shape = False + self.op_range_limit = '' + self.custom_compile_options = {} + self.custom_all_compile_options = {} + + @staticmethod + def _parse_digit(conf: str) -> int: + return int(conf.split('=')[1]) + + @staticmethod + def _parse_flag(conf: str) -> bool: + if 'true' == conf.split('=')[1]: + return True + return False + + @staticmethod + def _parse_str(conf: str) -> str: + return conf.split('=')[1] + + @staticmethod + def _parse_list(conf: str) -> list: + return conf.split('=')[1].split(',') + + def parse_input(self: any, conf: str): + if conf.startswith('input{}.name'.format(int(self.input_idx) + 1)): + self.input_idx += 1 + self.input_name.append(self._parse_str(conf) + '_in__') + elif conf.startswith('input{}.paramType'.format(int(self.input_idx))): + self.input_type.append(self._parse_str(conf)) + elif conf.startswith('input{}.dtype'.format(int(self.input_idx))): + self.input_dtype.append(self._parse_str(conf)) + elif conf.startswith('input{}.format'.format(int(self.input_idx))): + self.input_fmt.append(self._parse_str(conf)) + else: + return + + def parse_output(self: any, conf: str): + if conf.startswith('output{}.name'.format(int(self.output_idx) + 1)): + self.output_idx += 1 + self.output_name.append(self._parse_str(conf) + '_out_') + elif conf.startswith('output{}.paramType'.format(int(self.output_idx))): + self.output_type.append(self._parse_str(conf)) + elif conf.startswith('output{}.dtype'.format(int(self.output_idx))): + self.output_dtype.append(self._parse_str(conf)) + elif conf.startswith('output{}.format'.format(int(self.output_idx))): + self.output_fmt.append(self._parse_str(conf)) + else: + return + + def parse_op_format(self: any, conf: str): + self.op_fmt_sel = self._parse_flag(conf) + + def parse_check_support(self: any, conf: str): + self.op_chk_support = self._parse_flag(conf) + + def parse_range_limit(self: any, conf: str): + self.op_range_limit = self._parse_str(conf) + + def parse_kern_name(self: any, conf: str): + self.kern_name = self._parse_str(conf) + + def parse_op_intf(self: any, conf: str): + self.op_intf = self._parse_str(conf) + + def parse_op_file(self: any, conf: str): + self.op_file = self._parse_str(conf) + + def parse_dynamic_shape(self: any, conf: str): + self.dynamic_shape = self._parse_flag(conf) + + def parse_attr_list(self: any, conf: str): + self.attr_list = self._parse_list(conf) + + @staticmethod + def _camel_to_snake(camel_case_str: str): + snake_case_str = '' + for i, c in enumerate(camel_case_str): + if i == 0: + snake_case_str += c.lower() + elif c.isupper(): + snake_case_str += '_' + c.lower() + else: + snake_case_str += c + return snake_case_str + + def parse_attr_val(self: any, conf: str): + for attr in self.attr_list: + if self.attr_val.get(attr) is None: + self.attr_val[attr] = {} + if conf.startswith('attr_{}.type'.format(attr)): + self.attr_val.get(attr)['type'] = self._camel_to_snake(self._parse_str(conf)) + elif conf.startswith('attr_{}.paramType'.format(attr)): + self.attr_val.get(attr)['paramType'] = self._parse_str(conf) + elif conf.startswith('attr_{}.defaultValue'.format(attr)): + self.attr_val.get(attr)['defaultValue'] = self._parse_str(conf) + + def parse_replay_val(self: any, batch_list: list, iterator_list: list): + if self.op_type in batch_list: + self.op_replay_flag = True + self.op_replay_batch = True + elif self.op_type in iterator_list: + self.op_replay_flag = True + self.op_replay_batch = False + + +def _is_op_type_in_opdesc(op_descs: list, op_type: str): + for op in op_descs: + if op_type == op.op_type: + return True + return False + + +def _set_all_options_to_opdescs(op_descs, soc_ver_compile_options): + for op in op_descs: + op.custom_all_compile_options = soc_ver_compile_options + + +def _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options): + for op in op_descs: + if op.op_type != op_type: + continue + op.custom_compile_options = soc_ver_compile_options + + +def _trans_soc_ver_to_short(soc_ver: str): + low_soc_ver = soc_ver.lower() + if low_soc_ver not in SOC_TO_SHORT_SOC_MAP: + print(f'WARNING: caution: {soc_ver} will trans into ascend910, if not your intention,' + f'use ascend910b1~4 instead') + return SOC_TO_SHORT_SOC_MAP[low_soc_ver] + + +def _get_op_custom_options(op_descs: list, auto_gen_dir: str): + if auto_gen_dir is None: + return {} + file = os.path.join(auto_gen_dir, "custom_compile_options.ini") + if not os.path.exists(file): + print(f'WARNING: cannot find {auto_gen_dir}/custom_compile_options.ini') + return {} + with open (file, 'r') as fd: + lines = fd.readlines() + for line in lines: + param_list = str.split(line.rstrip('\n'), ',') + if len(param_list) != 3: + raise Exception(f'ERROR: custom compile option {param_list} len is not 3') + op_type = param_list[0] + if op_type.upper() == 'ALL': + op_type = OP_ALL + if op_type != OP_ALL and _is_op_type_in_opdesc(op_descs, op_type) == False: + print(f'WARNING: op: {op_type} are not exists in this project') + continue + soc_ver_compile_options = {} + soc_ver = param_list[1] + options_str = param_list[2] + options = str.split(options_str, ';') + if soc_ver == '': + soc_ver_compile_options[SOC_ALL] = options + else: + soc_ver_list = str.split(soc_ver, ';') + for ver in soc_ver_list: + short_ver = _trans_soc_ver_to_short(ver) + soc_ver_compile_options[short_ver] = options + if op_type == OP_ALL: + _set_all_options_to_opdescs(op_descs, soc_ver_compile_options) + else: + _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options) + + +def get_op_desc(file: str, batch_list: list, iterator_list: list, builder: any, + op_type: list, auto_gen_dir: str = None) -> list: + op_descs = [] + op_match = False + with open (file, 'r') as fd: + lines = fd.readlines() + for line in lines: + line = line.strip() + if line.startswith('['): + name = line[1:-1] + if op_type is None or name in op_type: + op_match = True + op_desc = builder(name) + op_desc.parse_replay_val(batch_list, iterator_list) + op_descs.append(op_desc) + else: + op_match = False + if op_type is not None and len(op_descs) == len(op_type): + return op_descs + continue + if not op_match: + continue + if line.startswith('input'): + op_desc.parse_input(line) + elif line.startswith('output'): + op_desc.parse_output(line) + elif line.startswith('dynamicFormat.flag'): + op_desc.parse_op_format(line) + elif line.startswith('needCheckSupport.flag'): + op_desc.parse_check_support(line) + elif line.startswith('rangeLimit.value'): + op_desc.parse_range_limit(line) + elif line.startswith('opInterface.value'): + op_desc.parse_op_intf(line) + elif line.startswith('kernel.name'): + op_desc.parse_kern_name(line) + elif line.startswith('opFile.value'): + op_desc.parse_op_file(line) + elif line.startswith('dynamicShapeSupport.flag'): + op_desc.parse_dynamic_shape(line) + elif line.startswith('attr.list'): + op_desc.parse_attr_list(line) + elif line.startswith('attr_'): + op_desc.parse_attr_val(line) + _get_op_custom_options(op_descs, auto_gen_dir) + return op_descs diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/parse_ini_to_json.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/parse_ini_to_json.py new file mode 100755 index 000000000..7a505aa38 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/parse_ini_to_json.py @@ -0,0 +1,338 @@ +# Copyright 2020-2021 Huawei Technologies Co., Ltd +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +""" +parser ini to json +""" + +import json +import os +import stat +import sys + + +ATTR_TYPE_LIST = ["int", "float", "bool", "str", "listInt", "listFloat", "listBool", "listStr", "listListInt", + "type", "listType", "tensor", "listTensor"] +ATTR_PARAMTYPE_LIST = ["optional", "required"] +BOOL_FLAG_KEY = ["dynamicFormat", "dynamicShapeSupport", "dynamicRankSupport", "precision_reduce", "heavyOp", + "needCheckSupport"] +BOOL_LIST = ["true", "false"] +DTYPE_LIST = ["float16", "float", "float32", "int8", "int16", "int32", "uint8", "uint16", "uint32", "bool", + "int64", "uint64", "qint8", "qint16", "qint32", "quint8", "quint16", "double", "complex64", + "complex128", "string", "resource", "dual", "dual_sub_int8", "dual_sub_uint8", "string_ref", + "int4", "bfloat16", "uint1"] +FORMAT_LIST = ["NCHW", "NHWC", "ND", "NC1HWC0", "FRACTAL_Z", "NC1C0HWPAD", "NHWC1C0", "FSR_NCHW", "FRACTAL_DECONV", + "C1HWNC0", "FRACTAL_DECONV_TRANSPOSE", "FRACTAL_DECONV_SP_STRIDE_TRANS", "NC1HWC0_C04", + "FRACTAL_Z_C04", "CHWN", "FRACTAL_DECONV_SP_STRIDE8_TRANS", "HWCN", "NC1KHKWHWC0", "BN_WEIGHT", + "FILTER_HWCK", "HASHTABLE_LOOKUP_LOOKUPS", "HASHTABLE_LOOKUP_KEYS", "HASHTABLE_LOOKUP_VALUE", + "HASHTABLE_LOOKUP_OUTPUT", "HASHTABLE_LOOKUP_HITS", "C1HWNCoC0", "MD", "NDHWC", "FRACTAL_ZZ", + "FRACTAL_NZ", "NCDHW", "DHWCN", "NDC1HWC0", "FRACTAL_Z_3D", "CN", "NC", "DHWNC", + "FRACTAL_Z_3D_TRANSPOSE", "FRACTAL_ZN_LSTM", "FRACTAL_ZN_RNN", "FRACTAL_Z_G", "NULL"] + + +def parse_ini_files(ini_files): + """ + parse ini files to json + Parameters: + ---------------- + ini_files:input file list + return:ops_info + ---------------- + """ + tbe_ops_info = {} + for ini_file in ini_files: + check_file_size(ini_file) + parse_ini_to_obj(ini_file, tbe_ops_info) + return tbe_ops_info + + +def check_file_size(input_file): + try: + file_size = os.path.getsize(input_file) + except OSError as os_error: + print('[ERROR] Failed to open "%s". %s' % (input_file, str(os_error))) + raise OSError from os_error + if file_size > 10*1024*1024: + print('[WARN] The size of %s exceeds 10MB, it may take more time to run, please wait.' % input_file) + + +def parse_ini_to_obj(ini_file, tbe_ops_info): + """ + parse ini file to json obj + Parameters: + ---------------- + ini_file:ini file path + tbe_ops_info:ops_info + ---------------- + """ + with open(ini_file) as ini_file: + lines = ini_file.readlines() + op_dict = {} + op_name = "" + find_op_type = False + for line in lines: + line = line.rstrip() + if line == "": + continue + if line.startswith("["): + if line.endswith("]"): + op_name = line[1:-1] + op_dict = {} + tbe_ops_info[op_name] = op_dict + find_op_type = True + elif "=" in line: + key1 = line[:line.index("=")] + key2 = line[line.index("=")+1:] + key1_0, key1_1 = key1.split(".") + if key1_0 not in op_dict: + op_dict[key1_0] = {} + if key1_1 in op_dict.get(key1_0): + raise RuntimeError("Op:" + op_name + " " + key1_0 + " " + + key1_1 + " is repeated!") + dic_key = op_dict.get(key1_0) + dic_key[key1_1] = key2 + else: + continue + if not find_op_type: + raise RuntimeError("Not find OpType in .ini file.") + + +def check_output_exist(op_dict, is_valid): + """ + Function Description: + Check output is exist + Parameter: op_dict + Parameter: is_valid + """ + if "output0" in op_dict: + output0_dict = op_dict.get("output0") + if output0_dict.get("name", None) is None: + is_valid = False + print("output0.name is required in .ini file!") + else: + is_valid = False + print("output0 is required in .ini file!") + return is_valid + + +def check_attr_dict(attr_dict, is_valid, attr): + """ + Function Description: + Check attr_dict + Parameter: attr_dict + Parameter: is_valid + Parameter: attr + """ + attr_type = attr_dict.get("type") + value = attr_dict.get("value") + param_type = attr_dict.get("paramType") + if attr_type is None or value is None: + is_valid = False + print("If attr.list is exist, {0}.type and {0}.value is required".format(attr)) + if param_type and param_type not in ATTR_PARAMTYPE_LIST: + is_valid = False + print("{0}.paramType only support {1}.".format(attr, ATTR_PARAMTYPE_LIST)) + if attr_type and attr_type not in ATTR_TYPE_LIST: + is_valid = False + print("{0}.type only support {1}.".format(attr, ATTR_TYPE_LIST)) + return is_valid + + +def check_attr(op_dict, is_valid): + """ + Function Description: + Check attr + Parameter: op_dict + Parameter: is_valid + """ + if "attr" in op_dict: + attr_dict = op_dict.get("attr") + attr_list_str = attr_dict.get("list", None) + if attr_list_str is None: + is_valid = False + print("attr.list is required in .ini file!") + else: + attr_list = attr_list_str.split(",") + for attr_name in attr_list: + attr = "attr_" + attr_name.strip() + attr_dict = op_dict.get(attr) + if attr_dict: + is_valid = check_attr_dict(attr_dict, is_valid, attr) + else: + is_valid = False + print("%s is required in .ini file, when attr.list is %s!" % (attr, attr_list_str)) + return is_valid + + +def check_bool_flag(op_dict, is_valid): + """ + Function Description: + check_bool_flag + Parameter: op_dict + Parameter: is_valid + """ + for key in BOOL_FLAG_KEY: + if key in op_dict: + op_bool_key = op_dict.get(key) + if op_bool_key.get("flag").strip() not in BOOL_LIST: + is_valid = False + print("{0}.flag only support {1}.".format(key, BOOL_LIST)) + return is_valid + + +def check_type_format(op_info, is_valid, op_info_key): + """ + Function Description: + Check type and format + Parameter: op_info + Parameter: is_valid + Parameter: op_info_key + """ + op_info_dtype_str = op_info.get("dtype") + op_info_dtype_num = 0 + op_info_format_num = 0 + if op_info_dtype_str: + op_info_dtype = op_info_dtype_str.split(",") + op_info_dtype_num = len(op_info_dtype) + for dtype in op_info_dtype: + if dtype.strip() not in DTYPE_LIST: + is_valid = False + print("{0}.dtype not support {1}.".format(op_info_key, dtype)) + op_info_format_str = op_info.get("format") + if op_info_format_str: + op_info_format = op_info_format_str.split(",") + op_info_format_num = len(op_info_format) + for op_format in op_info_format: + if op_format.strip() not in FORMAT_LIST: + is_valid = False + print("{0}.format not support {1}.".format(op_info_key, op_format)) + if op_info_dtype_num > 0 and op_info_format_num > 0: + if op_info_dtype_num != op_info_format_num: + is_valid = False + print("The number of {0}.dtype not match the number of {0}.format.".format(op_info_key)) + return is_valid + + +def check_op_info(tbe_ops): + """ + Function Description: + Check info. + Parameter: tbe_ops + Return Value: is_valid + """ + print("\n\n==============check valid for ops info start==============") + required_op_input_info_keys = ["paramType", "name"] + required_op_output_info_keys = ["paramType", "name"] + param_type_valid_value = ["dynamic", "optional", "required"] + is_valid = True + for op_key in tbe_ops: + op_dict = tbe_ops[op_key] + for op_info_key in op_dict: + if op_info_key.startswith("input"): + op_input_info = op_dict[op_info_key] + missing_keys = [] + for required_op_input_info_key in required_op_input_info_keys: + if required_op_input_info_key not in op_input_info: + missing_keys.append(required_op_input_info_key) + if len(missing_keys) > 0: + print("op: " + op_key + " " + op_info_key + " missing: " + + ",".join(missing_keys)) + is_valid = False + else: + if not op_input_info["paramType"] in param_type_valid_value: + print("op: " + op_key + " " + op_info_key + \ + " paramType not valid, valid key:[dynamic, " + "optional, required]") + is_valid = False + is_valid = check_type_format(op_input_info, is_valid, op_info_key) + if op_info_key.startswith("output"): + op_input_info = op_dict[op_info_key] + missing_keys = [] + for required_op_input_info_key in required_op_output_info_keys: + if required_op_input_info_key not in op_input_info: + missing_keys.append(required_op_input_info_key) + if len(missing_keys) > 0: + print("op: " + op_key + " " + op_info_key + " missing: " + + ",".join(missing_keys)) + is_valid = False + else: + if not op_input_info["paramType"] in param_type_valid_value: + print("op: " + op_key + " " + op_info_key + + " paramType not valid, valid key:[dynamic, " + "optional, required]") + is_valid = False + is_valid = check_type_format(op_input_info, is_valid, op_info_key) + is_valid = check_attr(op_dict, is_valid) + is_valid = check_bool_flag(op_dict, is_valid) + print("==============check valid for ops info end================\n\n") + return is_valid + + +def write_json_file(tbe_ops_info, json_file_path): + """ + Save info to json file + Parameters: + ---------------- + tbe_ops_info: ops_info + json_file_path: json file path + ---------------- + """ + json_file_real_path = os.path.realpath(json_file_path) + wr_flag = os.O_WRONLY | os.O_CREAT + wr_mode = stat.S_IWUSR | stat.S_IRUSR + with os.fdopen(os.open(json_file_real_path, wr_flag, wr_mode), 'w') as file_path: + # The owner have all rights,group only have read rights + os.chmod(json_file_real_path, stat.S_IWUSR + stat.S_IRGRP + + stat.S_IRUSR) + json.dump(tbe_ops_info, file_path, sort_keys=True, indent=4, + separators=(',', ':')) + print("Compile op info cfg successfully.") + + +def parse_ini_to_json(ini_file_paths, outfile_path): + """ + parse ini files to json file + Parameters: + ---------------- + ini_file_paths: list of ini file path + outfile_path: output file path + ---------------- + """ + tbe_ops_info = parse_ini_files(ini_file_paths) + if not check_op_info(tbe_ops_info): + print("Compile op info cfg failed.") + return False + write_json_file(tbe_ops_info, outfile_path) + return True + + +if __name__ == '__main__': + args = sys.argv + + OUTPUT_FILE_PATH = "tbe_ops_info.json" + ini_file_path_list = [] + + for arg in args: + if arg.endswith("ini"): + ini_file_path_list.append(arg) + OUTPUT_FILE_PATH = arg.replace(".ini", ".json") + if arg.endswith("json"): + OUTPUT_FILE_PATH = arg + + if len(ini_file_path_list) == 0: + ini_file_path_list.append("tbe_ops_info.ini") + + if not parse_ini_to_json(ini_file_path_list, OUTPUT_FILE_PATH): + sys.exit(1) + sys.exit(0) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/preset_parse.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/preset_parse.py new file mode 100755 index 000000000..8f1124b1d --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/preset_parse.py @@ -0,0 +1,23 @@ +import json +import sys +import os + + +def get_config_opts(file): + src_dir = os.path.abspath(os.path.dirname(file)) + opts = '' + with open(file, 'r') as fd: + config = json.load(fd) + for conf in config: + if conf == 'configurePresets': + for node in config[conf]: + macros = node.get('cacheVariables') + if macros is not None: + for key in macros: + opts += '-D{}={} '.format(key, macros[key]['value']) + opts = opts.replace('${sourceDir}', src_dir) + print(opts) + + +if __name__ == "__main__": + get_config_opts(sys.argv[1]) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_codegen.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_codegen.py new file mode 100755 index 000000000..1baa364ef --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_codegen.py @@ -0,0 +1,105 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import os +import stat +import collections +import kernel_entry as keb +from tiling_data_def_build import gen_tiling +import code_channel_infer +import const_var + +PYF_PATH = os.path.dirname(__file__) + +ReplayCodeGenParams = collections.namedtuple('ReplayCodeGenParams',\ +['op_type', 'impl', 'tiling_file', 'kernel', 'entry', 'argn', 'op_replay_batch', 'max_block_dim', 'max_shape_size']) + + +class ReplayCodeGen: + def __init__(self, replayCodeGenParams): + self.op_type = replayCodeGenParams.op_type + self.impl = replayCodeGenParams.impl + self.tiling_file = replayCodeGenParams.tiling_file + self.tiling_data_file = '' + self.kernel = replayCodeGenParams.kernel + self.entry = replayCodeGenParams.entry + self.argn = replayCodeGenParams.argn + self.batch = False + self.outdir = '' + self.data_type = 'uint8_t' + self.blknum = 32 + self.op_replay_batch = replayCodeGenParams.op_replay_batch + self.max_block_dim = replayCodeGenParams.max_block_dim + self.max_shape_size = replayCodeGenParams.max_shape_size + + def set_batch(self, is_batch): + self.batch = is_batch + + def set_outdir(self, outdir): + self.outdir = outdir + + def gen_replay(self, ops_product: str): + kerentry = os.path.join(self.outdir, self.kernel + '_entry.cce') + kerimpl = os.path.join(self.outdir, self.kernel + '_impl.cpp') + replayimpl = os.path.join(self.outdir, self.kernel + '_replay.cpp') + if self.batch: + reptmp = os.path.join(PYF_PATH, 'batch_replay_impl.temp') + else: + reptmp = os.path.join(PYF_PATH, 'replay_impl.temp') + kertmp = os.path.join(PYF_PATH, 'kernel_impl.temp') + self._gen_kentry(kerentry) + self._gen_kimpl_code(kerimpl, kertmp) + self._gen_tiling_data_header() + self._gen_replay_code(replayimpl, reptmp, ops_product) + + def _gen_tiling_data_header(self): + self.tiling_data_file = os.path.join(self.outdir, self.kernel + '_tiling_data.h') + gen_tiling(self.tiling_file, self.tiling_data_file) + + def _gen_kimpl_code(self, src, tmpfile): + with open(tmpfile, 'r') as fd: + temp = fd.read() + temp = temp.replace('__CCE_FILE__', self.impl) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(temp) + + def _gen_replay_code(self, src, tmpfile, ops_product: str): + with open(tmpfile, 'r') as fd: + temp = fd.read() + temp = temp.replace('__ARG_NUM__', str(self.argn)) + argdef = [] + kargs = [] + for i in range(0, self.argn): + argdef.append('{} *'.format(self.data_type)) + kargs.append('({} *)GetArg({})'.format(self.data_type, i)) + temp = temp.replace('__ARGS_DEF__', ', '.join(argdef)) + temp = temp.replace('__KERNEL_ARGS__', ', '.join(kargs)) + temp = temp.replace('__KERNEL_FUN__', self.entry) + core_type_infer = 'core_type' + code_channel = code_channel_infer.infer_code_channel(code_channel_infer.InfoCodeChanelParams(self.impl,\ + self.tiling_data_file, self.kernel, self.outdir, ops_product, None)) + if code_channel == code_channel_infer.CODE_VEC: + core_type_infer = '0' + elif code_channel == code_channel_infer.CODE_CUBE: + core_type_infer = '1' + temp = temp.replace('__CORE_TYPE__', core_type_infer) + # regist function + temp = temp.replace('__OPS_PRODUCT__', ops_product) + temp = temp.replace('__OPTYPE__', self.op_type) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(temp) + + def _gen_kentry(self, src): + kf = '' + pre_alloc_str = 'A' * 256 + if self.batch: + kf += keb.batch_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str), self.argn, self.data_type) + else: + kf += keb.mc_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str),\ + self.argn, self.data_type, self.blknum) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(kf) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_impl.temp b/atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_impl.temp new file mode 100755 index 000000000..1d30dd865 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_impl.temp @@ -0,0 +1,120 @@ +#include +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#include "register/op_check.h" +#define __ASCENDC_REPLAY_CODE__ +using namespace std; +using namespace optiling; +using namespace AscendCReplay; + +extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); +extern "C" int elf_append(char *elf, uint32_t elfSize, char *jit, int kernum, int blknum[], char *atext[], + int alen[], int atlen, const char* kernelname[]); + +#define KERNEL_N 1 +#define ARG_N (__ARG_NUM__) +#define MAX_L (1024 * 1024 * 100) +#define MAX_E (1024 * 1024) + +int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) +{ + // gen type 1 : direct call codes 0: load .o file + if (param.gentype < 0 || param.gentype > 1) { + printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); + return 0; + } else if (param.gentype == 1 && param.objptr == nullptr) { + printf("Error: call replay with direct call mode, but code obj addr is null\n"); + return 0; + } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { + printf("Error: call replay with object file mode, but object file path is null\n"); + return 0; + } + // core_type 0:MIX 1:CUBE 2:VEC + if (core_type < 0 || core_type > 2) { + printf("Error: call replay core type is %d !\n", core_type); + return 0; + } + g_coreType = __CORE_TYPE__; + g_taskRation = param.task_ration; + g_tilingKey = param.tiling_key; + + unsigned char *buf, *jit; + char *kernel[KERNEL_N * 32]; + int len[KERNEL_N * 32]; + int blknum[KERNEL_N]; + int max; + block_num = param.block_dim; + g_ubBase = block_num; + uint8_t *code = (uint8_t *)malloc(MAX_L); + uint8_t *pos = code; + struct timespec tp1, tp2; + + clock_gettime(CLOCK_MONOTONIC, &tp1); + if (block_num > 32) { + printf("Error: block_num > 32\n"); + return 0; + } + //__OP_FOPEN__ + for (int i = 0; i < KERNEL_N; i++) { + for (int j = 0; j < ARG_N; j++) + AddArg(j, ARG_STEP * (j + 1)); + for (block_idx = 0; block_idx < block_num; block_idx++) { + //__OP_SET_KERNEL__ + int code_idx = i * block_num + block_idx; +#ifdef FP_CEILING + SetCtrlFloatEnable(); +#else + SetCtrlFloatDisable(); +#endif + CodeInit(pos, false); + __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); + CodeEnd(); + kernel[code_idx] = (char *)pos; + len[code_idx] = CodeLen(); + pos += len[code_idx]; + printf("kernel %d core %ld code generated len %d\n", i, block_idx, len[code_idx]); + } + blknum[i] = block_num; + } + //__OP_FCLOSE__ + clock_gettime(CLOCK_MONOTONIC, &tp2); + buf = (unsigned char *)malloc(MAX_E); + int fd = open(param.entry_file, O_RDONLY); + if (fd < 0) { + printf("[error]: cannot find entry.o : %s\n", param.entry_file); + return 0; + } + uint32_t bufSize = read(fd, buf, MAX_E); + if (bufSize <= 0) { + printf("[error]: entry.o : %s is too small ! \n", param.entry_file); + } + close(fd); + jit = (unsigned char *)malloc(MAX_L); + printf("total code generated %ld\n", pos - code); + int sz = elf_append((char *)buf, bufSize, (char *)jit, KERNEL_N, blknum, kernel, len, pos - code, ¶m.kernel_name); + if (tp1.tv_sec != tp2.tv_sec) { + printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); + } else { + printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); + } + printf("new elf size %d\n", sz); + if (param.gentype == 0) { + fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); + (void)write(fd, jit, sz); + close(fd); + free(jit); + } else if (param.gentype == 1) { + *param.objptr = (char*)jit; + } + free(buf); + free(code); + return sz; +} + +REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/tiling_data_def_build.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/tiling_data_def_build.py new file mode 100755 index 000000000..4c5b099b7 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/cmake/util/tiling_data_def_build.py @@ -0,0 +1,87 @@ +#!/usr/bin/env python +# coding=utf-8 +""" +Function: +The replay funtion entry +Copyright Information: +Huawei Technologies Co., Ltd. All Rights Reserved © 2020 +""" + +import sys +import os +import stat +import re +import const_var + + +def gen_tiling(tiling_header_file: str, tiling_file_out: str): + if not os.path.exists(tiling_header_file): + print("warning: no userdef tiling header file: ", tiling_header_file) + return + print("generate tiling def header file: ", tiling_file_out) + tmp_name = os.path.splitext(os.path.basename(tiling_header_file))[0].upper() + tiling_source = '#ifndef __{}_H__\n'.format(tmp_name) + tiling_source += '#define __{}_H__\n\n'.format(tmp_name) + tiling_source += '#include \n' + tiling_source += '#include \n\n' + tiling_source += '#include "kernel_tiling/kernel_tiling.h"\n\n' + end_source = "" + pattern = re.compile(r'[(](.*)[)]', re.S) + with open(tiling_header_file, 'r') as fd: + lines = fd.readlines() + for line in lines: + line = line.strip() + if (line.startswith('BEGIN_TILING_DATA_DEF')): + tiling_source += '#pragma pack(1)\n' + tiling_source += 'struct ' + struct_def = re.findall(pattern, line)[0] + tiling_source += struct_def + ' {\n' + elif (line.startswith('TILING_DATA_FIELD_DEF_ARR')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {}[{}] = {{}};\n'.format(fds[0].strip(), fds[2].strip(), fds[1].strip()) + elif (line.startswith('TILING_DATA_FIELD_DEF_STRUCT')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {};\n'.format(fds[0].strip(), fds[1].strip()) + elif (line.startswith('TILING_DATA_FIELD_DEF')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {} = 0;\n'.format(fds[0].strip(), fds[1].strip()) + elif (line.startswith('END_TILING_DATA_DEF')): + tiling_source += '};\n' + tiling_source += '#pragma pack()\n\n' + tiling_source += '#ifdef __NPU_TILING__\n' + tiling_source += \ + 'inline [aicore] void Init{stru}(const __gm__ uint8_t* tiling, {stru}* const_data)\n'\ + .format(stru=struct_def) + tiling_source += '{\n' + tiling_source += ' const __gm__ uint32_t *src = (const __gm__ uint32_t *)tiling;\n' + tiling_source += ' uint32_t *dst = (uint32_t *)const_data;\n' + tiling_source += ' for (auto i = 0; i < sizeof({}) / 4; i++) *(dst + i) = *(src + i);\n'\ + .format(struct_def) + tiling_source += '}\n' + tiling_source += '#else\n' + tiling_source += 'inline void Init{stru}(uint8_t* tiling, {stru}* const_data)\n'.format(stru=struct_def) + tiling_source += '{\n' + tiling_source += ' uint64_t *src = (uint64_t *)tiling;\n' + tiling_source += ' uint64_t *dst = (uint64_t *)const_data;\n' + tiling_source += ' for (auto i = 0; i < sizeof({}) / 8; i++) *(dst + i) = *(src + i);\n'\ + .format(struct_def) + tiling_source += '}\n' + tiling_source += '#endif\n\n' + end_source = ''' +#define GET_TILING_DATA(tiling_data, tiling_arg) \\ +{stru} tiling_data; \\ +Init{stru}(tiling_arg, &tiling_data)\n +'''.format(stru=struct_def) + tiling_source += end_source + tiling_source += '#endif' + with os.fdopen(os.open(tiling_file_out, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(tiling_source) + + +if __name__ == '__main__': + if len(sys.argv) <= 2: + raise RuntimeError('arguments must greater than 2') + gen_tiling(sys.argv[1], sys.argv[2]) diff --git a/atb_operator/AtbTrilSample/TrilOp/framework/CMakeLists.txt b/atb_operator/AtbTrilSample/TrilOp/framework/CMakeLists.txt new file mode 100755 index 000000000..b6be9b492 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/framework/CMakeLists.txt @@ -0,0 +1,11 @@ +if(NOT EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/mindspore") + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/caffe_plugin") + add_subdirectory(caffe_plugin) + endif() + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/tf_plugin") + add_subdirectory(tf_plugin) + endif() + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/onnx_plugin") + add_subdirectory(onnx_plugin) + endif() +endif() diff --git a/atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/CMakeLists.txt b/atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/CMakeLists.txt new file mode 100755 index 000000000..a6aba5c20 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/CMakeLists.txt @@ -0,0 +1,14 @@ + +aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} plugin_srcs) +add_library(cust_tf_parsers SHARED ${plugin_srcs}) +target_compile_definitions(cust_tf_parsers PRIVATE google=ascend_private) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_tf_parsers PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_tf_parsers PRIVATE intf_pub graph) +install(TARGETS cust_tf_parsers + LIBRARY DESTINATION packages/vendors/${vendor_name}/framework/tensorflow +) diff --git a/atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/tensorflow_tril_plugin.cc b/atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/tensorflow_tril_plugin.cc new file mode 100755 index 000000000..9b71ea121 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/tensorflow_tril_plugin.cc @@ -0,0 +1,20 @@ +/* Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the Apache License Version 2.0. + * You may not use this file except in compliance with the License. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. + */ + +#include "register/register.h" + +namespace domi { +// register op info to GE +REGISTER_CUSTOM_OP("Tril") + .FrameworkType(TENSORFLOW) // type: CAFFE, TENSORFLOW + .OriginOpType("Tril") // name in tf module + .ParseParamsByOperatorFn(AutoMappingByOpFn); +} // namespace domi diff --git a/atb_operator/AtbTrilSample/TrilOp/op_host/CMakeLists.txt b/atb_operator/AtbTrilSample/TrilOp/op_host/CMakeLists.txt new file mode 100755 index 000000000..40dd51cfa --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/op_host/CMakeLists.txt @@ -0,0 +1,82 @@ + +aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} ops_srcs) + +opbuild(OPS_SRC ${ops_srcs} + OUT_DIR ${ASCEND_AUTOGEN_PATH} +) + +add_library(cust_op_proto SHARED ${ops_srcs} ${ASCEND_AUTOGEN_PATH}/op_proto.cc) +target_compile_definitions(cust_op_proto PRIVATE OP_PROTO_LIB) +target_compile_options(cust_op_proto PRIVATE + -fvisibility=hidden +) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_op_proto PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_op_proto PRIVATE + intf_pub + exe_graph + register + tiling_api + -Wl,--whole-archive + rt2_registry + -Wl,--no-whole-archive +) +set_target_properties(cust_op_proto PROPERTIES OUTPUT_NAME + cust_opsproto_rt2.0 +) +add_library(cust_optiling SHARED ${ops_srcs}) +target_compile_definitions(cust_optiling PRIVATE OP_TILING_LIB) +target_compile_options(cust_optiling PRIVATE + -fvisibility=hidden +) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_optiling PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_optiling PRIVATE + intf_pub + exe_graph + register + tiling_api + -Wl,--whole-archive + rt2_registry + -Wl,--no-whole-archive +) +set_target_properties(cust_optiling PROPERTIES OUTPUT_NAME + cust_opmaster_rt2.0 +) + +file(GLOB aclnn_src ${ASCEND_AUTOGEN_PATH}/aclnn_*.cpp) +file(GLOB aclnn_inc ${ASCEND_AUTOGEN_PATH}/aclnn_*.h) +add_library(cust_opapi SHARED ${aclnn_src}) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_opapi PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_opapi PRIVATE intf_pub ascendcl nnopbase) + +add_custom_target(optiling_compat ALL + COMMAND ln -sf lib/linux/${CMAKE_SYSTEM_PROCESSOR}/$ + ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so +) + +install(TARGETS cust_op_proto + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_proto/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) +install(FILES ${ASCEND_AUTOGEN_PATH}/op_proto.h + DESTINATION packages/vendors/${vendor_name}/op_proto/inc) +install(TARGETS cust_optiling + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) +install(FILES ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so + DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling) +install(TARGETS cust_opapi + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_api/lib) +install(FILES ${aclnn_inc} + DESTINATION packages/vendors/${vendor_name}/op_api/include) diff --git a/atb_operator/AtbTrilSample/TrilOp/op_host/tril.cpp b/atb_operator/AtbTrilSample/TrilOp/op_host/tril.cpp new file mode 100755 index 000000000..60b0bb3b0 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/op_host/tril.cpp @@ -0,0 +1,215 @@ +/* Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the Apache License Version 2.0. + * You may not use this file except in compliance with the License. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. + */ + +#include "tril_tiling.h" +#include "register/op_def_registry.h" +#include "tiling/platform/platform_ascendc.h" + +namespace optiling { + constexpr int minNum = 1; + + constexpr int keyOne = 1; + constexpr int keyTwo = 2; + constexpr int keyThree = 3; + constexpr int keyFour = 4; + + constexpr int bufferFour = 4; + constexpr int BlockSize = 32; + constexpr int computeBatchSize = 256; + constexpr int sizeHalf = 2; + constexpr int VAL_ZRRO = 0; + + uint32_t typeSize = VAL_ZRRO; + uint64_t key = keyOne; + // buffer for queue + uint64_t UB_SHARING_NUM = 2; + int64_t rowLength = VAL_ZRRO; + int64_t columnLength = VAL_ZRRO; + int64_t matrixNum = 1, matrixSize = 1; + int64_t diagVal = VAL_ZRRO; + + uint32_t ALIGN_NUM = VAL_ZRRO; + uint32_t totalLengthAligned = VAL_ZRRO; + uint64_t loopCnt = VAL_ZRRO, fullTileLength = VAL_ZRRO, lastTileLength = VAL_ZRRO; + int32_t fullCnt = VAL_ZRRO, lastCnt = VAL_ZRRO; + + static int setShapeInfo(gert::TilingContext *context){ + const auto inputDataType = context->GetInputTensor(0)->GetDataType(); + + switch (inputDataType){ + case ge::DT_FLOAT: + typeSize = sizeof(float); + break; + case ge::DT_FLOAT16: + typeSize = sizeHalf; + break; + default: + typeSize = sizeof(float); + break; + } + + const auto inputShape = context->GetInputTensor(0)->GetOriginShape(); + // class Shape: size_t dim_num_; int64_t dims_[]; + int64_t dimSize = inputShape.GetDimNum(), i = 0; + // The number 2 is to preserve the last two dimensions + for (i = 0; i < dimSize - 2; i++){ + matrixNum *= inputShape.GetDim(i); + } + rowLength = inputShape.GetDim(i); + i++; + columnLength = inputShape.GetDim(i); + matrixSize = rowLength * columnLength; + + const auto runtime_attrs = context->GetAttrs(); + const int64_t *diagPtr = runtime_attrs->GetInt(0); + diagVal = *diagPtr; + if (diagVal < columnLength - 1 && diagVal > -rowLength){ + // Regular + key = keyOne; + }else if (diagVal <= -rowLength){ + // The result is itself, TQueBind is enough + key = keyTwo; + }else{ + // All zero, just copyIn, Sub and copyOut + key = keyThree; + } + return 0; + } + + static int setTilingInfo(gert::TilingContext *context,uint64_t ub_size){ + loopCnt = VAL_ZRRO; + fullTileLength = VAL_ZRRO; + lastTileLength = VAL_ZRRO; + fullCnt = VAL_ZRRO; + lastCnt = VAL_ZRRO; + uint64_t ub_length = ((ub_size / typeSize / UB_SHARING_NUM) / ALIGN_NUM * ALIGN_NUM) - ALIGN_NUM; + if (key == keyOne && diagVal <= 0 && columnLength % (computeBatchSize / typeSize) == 0){ + // A faster method for aligned processing only + key = keyFour; + // Double buffer setting + UB_SHARING_NUM = bufferFour; + // The result would not be the expected + if (columnLength == 0){ + columnLength = minNum; + } + ub_length = ((ub_size) / typeSize / UB_SHARING_NUM) / columnLength * columnLength; + loopCnt = (matrixSize + ub_length - 1) / ub_length; + if (loopCnt == 1){ + fullCnt = 0; + lastCnt = rowLength; + }else{ + // The result would not be the expected + if (columnLength == 0){ + columnLength = minNum; + } + fullCnt = ub_length / columnLength; + lastCnt = rowLength - fullCnt * (loopCnt - 1); + } + // Already aligned + fullTileLength = fullCnt * columnLength; + lastTileLength = lastCnt * columnLength; + }else if (key == keyThree){ + loopCnt = (totalLengthAligned + ub_length - 1) / ub_length; + UB_SHARING_NUM = bufferFour; + ub_length = ((ub_size / typeSize / UB_SHARING_NUM) / ALIGN_NUM * ALIGN_NUM) - ALIGN_NUM; + fullTileLength = ub_length; + lastTileLength = (totalLengthAligned - fullTileLength * (loopCnt - 1) + ALIGN_NUM - 1) / ALIGN_NUM * ALIGN_NUM; + if (loopCnt == 1){ fullTileLength = 0; } + }else{ + loopCnt = (totalLengthAligned + ub_length - 1) / ub_length; + fullTileLength = ub_length; + lastTileLength = (totalLengthAligned - fullTileLength * (loopCnt - 1) + ALIGN_NUM - 1) / ALIGN_NUM * ALIGN_NUM; + if (loopCnt == 1){ fullTileLength = 0; } + } + return 0; + } + +static ge::graphStatus TilingFunc(gert::TilingContext* context) +{ + TrilTilingData tiling; + auto ascendcPlatform = platform_ascendc::PlatformAscendC(context->GetPlatformInfo()); + auto coreNum = ascendcPlatform.GetCoreNum(); + auto BLOCK_DIM = 1; + context->SetBlockDim(BLOCK_DIM); + + setShapeInfo(context); + + ALIGN_NUM = BlockSize / typeSize; + totalLengthAligned = (matrixNum * matrixSize + ALIGN_NUM - 1) / ALIGN_NUM * ALIGN_NUM; + uint64_t ub_size=0; + ascendcPlatform.GetCoreMemSize(platform_ascendc::CoreMemType::UB, ub_size); + + setTilingInfo(context,ub_size); + + tiling.set_totalLengthAligned(totalLengthAligned); + tiling.set_matrixNum(matrixNum); + tiling.set_matrixSize(matrixSize); + tiling.set_rowLength(rowLength); + tiling.set_columnLength(columnLength); + tiling.set_diagVal(diagVal); + + tiling.set_loopCnt(loopCnt); + tiling.set_fullTileLength(fullTileLength); + tiling.set_lastTileLength(lastTileLength); + tiling.set_fullCnt(fullCnt); + tiling.set_lastCnt(lastCnt); + + tiling.set_alignNum(ALIGN_NUM); + tiling.set_typeSize(typeSize); + + tiling.SaveToBuffer(context->GetRawTilingData()->GetData(), context->GetRawTilingData()->GetCapacity()); + context->GetRawTilingData()->SetDataSize(tiling.GetDataSize()); + context->SetTilingKey(key); + size_t *currentWorkspace = context->GetWorkspaceSizes(1); + currentWorkspace[0] = 0; + + return ge::GRAPH_SUCCESS; +} +} + +namespace ge { +static ge::graphStatus InferShape(gert::InferShapeContext* context) +{ + const gert::Shape* x1_shape = context->GetInputShape(0); + gert::Shape* y_shape = context->GetOutputShape(0); + *y_shape = *x1_shape; + return GRAPH_SUCCESS; +} +} + +namespace ops { +class Tril : public OpDef { +public: + explicit Tril(const char* name) : OpDef(name) + { + this->Input("x") + .ParamType(REQUIRED) + .DataType({ge::DT_FLOAT16, ge::DT_FLOAT}) + .Format({ge::FORMAT_ND, ge::FORMAT_ND}) + .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND}); + this->Output("y") + .ParamType(REQUIRED) + .DataType({ge::DT_FLOAT16, ge::DT_FLOAT}) + .Format({ge::FORMAT_ND, ge::FORMAT_ND}) + .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND}); + this->Attr("diagonal").AttrType(OPTIONAL).Int(0); + + this->SetInferShape(ge::InferShape); + + this->AICore() + .SetTiling(optiling::TilingFunc); + this->AICore().AddConfig("ascend310b") + .AddConfig("ascend910b"); + } +}; + +OP_ADD(Tril); +} diff --git a/atb_operator/AtbTrilSample/TrilOp/op_host/tril_tiling.h b/atb_operator/AtbTrilSample/TrilOp/op_host/tril_tiling.h new file mode 100755 index 000000000..fbf7749d2 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/op_host/tril_tiling.h @@ -0,0 +1,34 @@ +/* Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the Apache License Version 2.0. + * You may not use this file except in compliance with the License. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. + */ +#ifndef TRIL_TILING_H +#define TRIL_TILING_H +#include "register/tilingdata_base.h" + +namespace optiling { +BEGIN_TILING_DATA_DEF(TrilTilingData) + TILING_DATA_FIELD_DEF(uint32_t, totalLengthAligned); + TILING_DATA_FIELD_DEF(int32_t, matrixNum); + TILING_DATA_FIELD_DEF(int32_t, matrixSize); + TILING_DATA_FIELD_DEF(int32_t, rowLength); + TILING_DATA_FIELD_DEF(int32_t, columnLength); + TILING_DATA_FIELD_DEF(int32_t, diagVal); + TILING_DATA_FIELD_DEF(int32_t, loopCnt); + TILING_DATA_FIELD_DEF(uint32_t, fullTileLength); + TILING_DATA_FIELD_DEF(uint32_t, lastTileLength); + TILING_DATA_FIELD_DEF(int32_t, fullCnt); + TILING_DATA_FIELD_DEF(int32_t, lastCnt); + TILING_DATA_FIELD_DEF(uint32_t, alignNum); + TILING_DATA_FIELD_DEF(uint32_t, typeSize); +END_TILING_DATA_DEF; + +REGISTER_TILING_DATA_CLASS(Tril, TrilTilingData) +} +#endif \ No newline at end of file diff --git a/atb_operator/AtbTrilSample/TrilOp/op_kernel/CMakeLists.txt b/atb_operator/AtbTrilSample/TrilOp/op_kernel/CMakeLists.txt new file mode 100755 index 000000000..8c94a952d --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/op_kernel/CMakeLists.txt @@ -0,0 +1,68 @@ +# set custom compile options +if ("${CMAKE_BUILD_TYPE}x" STREQUAL "Debugx") + add_ops_compile_options(ALL OPTIONS -g -O0) +endif() + +foreach(compute_unit ${ASCEND_COMPUTE_UNIT}) + + # generate aic-${compute_unit}-ops-info.json + add_ops_info_target(TARGET ops_info_gen_${compute_unit} + OUTPUT ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core/${compute_unit}/aic-${compute_unit}-ops-info.json + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/config/${compute_unit} + ) + + # generate ascendc impl py once + if (NOT TARGET ascendc_impl_gen) + add_ops_impl_target(TARGET ascendc_impl_gen + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl + ) + endif() + + # dynamic shape binary compile + if (${ENABLE_BINARY_PACKAGE} AND NOT ${ENABLE_CROSS_COMPILE}) + add_bin_compile_target(TARGET ascendc_bin_${compute_unit} + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} + ADP_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/dynamic + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/binary/${compute_unit} + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/kernel + COMPUTE_UNIT ${compute_unit} + ) + add_dependencies(ascendc_bin_${compute_unit} ascendc_impl_gen) + endif() + + if (${ENABLE_CROSS_COMPILE} AND ${ENABLE_BINARY_PACKAGE}) + add_cross_compile_target( + TARGET bin_${compute_unit} + OUT_DIR ${CMAKE_CURRENT_SOURCE_DIR}/../kernel + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/ + ) + endif() +endforeach() + +# generate npu_supported_ops.json +add_npu_support_target(TARGET npu_supported_ops + OPS_INFO_DIR ${ASCEND_AUTOGEN_PATH} + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core + INSTALL_DIR packages/vendors/${vendor_name}/framework/${ASCEND_FRAMEWORK_TYPE} +) + +if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) + add_subdirectory(testcases) +endif() + +# install kernel file +if (${ENABLE_SOURCE_PACKAGE}) + file(GLOB KERNEL_FILES + ${CMAKE_CURRENT_SOURCE_DIR}/*.cpp + ${CMAKE_CURRENT_SOURCE_DIR}/*.h + ${CMAKE_CURRENT_SOURCE_DIR}/*.py + ) + install(FILES ${KERNEL_FILES} + DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl/dynamic + ) +endif() diff --git a/atb_operator/AtbTrilSample/TrilOp/op_kernel/tril.cpp b/atb_operator/AtbTrilSample/TrilOp/op_kernel/tril.cpp new file mode 100755 index 000000000..71711b6fa --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/op_kernel/tril.cpp @@ -0,0 +1,278 @@ +/* Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the Apache License Version 2.0. + * You may not use this file except in compliance with the License. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. + */ + +#include "kernel_operator.h" +using namespace AscendC; +constexpr int32_t BUFFER_NUM = 2; // tensor num for each queue +constexpr int32_t minNum = 1; + +constexpr int keyOne = 1; +constexpr int keyTwo = 2; +constexpr int keyThree = 3; +constexpr int keyFour = 4; + +constexpr int computeBatchSize = 256; + +struct IntegrateParam{ + uint32_t totalLengthAligned; + int32_t matrixNum; + int32_t matrixSize; + int32_t rowLength; + int32_t columnLength; + int32_t diagVal; + int32_t loopCnt; + uint32_t fullTileLength; + uint32_t lastTileLength; + int32_t fullCnt; + int32_t lastCnt; + uint32_t alignNum; + uint32_t typeSize; +}; + +class KernelTril { +public: + __aicore__ inline KernelTril() {} + //only pass the length this one is assigned to + __aicore__ inline void Init(GM_ADDR x, GM_ADDR y, IntegrateParam& paramList, uint32_t key) + { + this->matrixNum = paramList.matrixNum; + this->matrixSize = paramList.matrixSize; + this->rowLength = paramList.rowLength; + this->columnLength = paramList.columnLength; + this->diagVal = paramList.diagVal; + this->fullCnt = paramList.fullCnt; + this->lastCnt = paramList.lastCnt; + if(paramList.columnLength==0){ + paramList.columnLength = minNum; + } + this->fullRowInc = paramList.fullTileLength / paramList.columnLength; + this->initLength = 1; + // The result would not be the expected + if(paramList.typeSize==0){ + paramList.typeSize = sizeof(float); + } + + this->typeSize = paramList.typeSize; + + this->key=key; + + uint64_t gmBuffer=paramList.totalLengthAligned; + + xGm.SetGlobalBuffer((__gm__ DTYPE_X*)x, gmBuffer); + yGm.SetGlobalBuffer((__gm__ DTYPE_X*)y, gmBuffer); + + this->loopCnt = paramList.loopCnt; + this->fullTileLength = paramList.fullTileLength; + this->lastTileLength = paramList.lastTileLength; + + uint32_t singleBuffer = paramList.fullTileLength; + if(singleBuffer < paramList.lastTileLength){ + singleBuffer = paramList.lastTileLength; + } + if(key==keyThree || key==keyFour){ + pipe.InitBuffer(inQueueX, BUFFER_NUM, singleBuffer * this->typeSize); + pipe.InitBuffer(outQueueY, BUFFER_NUM, singleBuffer * this->typeSize); + }else{ + pipe.InitBuffer(queBind, BUFFER_NUM, singleBuffer * this->typeSize); + } + } + + __aicore__ inline void Process() + { + if(this->key==keyOne){ + NaivePath(); + }else if(this->key==keyTwo){ + SheerDup(); + }else if(this->key==keyThree){ + SheerZero(); + }else if(key==keyFour){ + FastPath(); + } + } + +private: + __aicore__ inline void SheerDup() + { + uint32_t GmOffset=0; + for (int i = 0; i < this->loopCnt-1; i++, GmOffset+=this->fullTileLength) { + auto bindLocal = queBind.AllocTensor(); + DataCopy(bindLocal, xGm[GmOffset], this->fullTileLength); + queBind.EnQue(bindLocal); + bindLocal = queBind.DeQue(); + DataCopy(yGm[GmOffset], bindLocal, this->fullTileLength); + queBind.FreeTensor(bindLocal); + } + auto bindLocal = queBind.AllocTensor(); + DataCopy(bindLocal, xGm[GmOffset], this->lastTileLength); + queBind.EnQue(bindLocal); + bindLocal = queBind.DeQue(); + DataCopy(yGm[GmOffset], bindLocal, this->lastTileLength); + queBind.FreeTensor(bindLocal); + } + + __aicore__ inline void SheerZero(){ + uint32_t GmOffset=0; + for (int i = 0; i < this->loopCnt-1; i++, GmOffset+=this->fullTileLength) { + CopyIn(GmOffset,this->fullTileLength); + AllZero(this->fullTileLength); + CopyOut(GmOffset,this->fullTileLength); + } + CopyIn(GmOffset,this->lastTileLength); + AllZero(this->lastTileLength); + CopyOut(GmOffset,this->lastTileLength); + } + + __aicore__ inline void NaivePath(){ + int32_t cnt=0; + for(int32_t i=0;imatrixNum;i++){ + for(int32_t j=0;jrowLength;j++){ + int32_t k=0; + while(kcolumnLength && k-j<=this->diagVal){ + DTYPE_X curr=xGm.GetValue(cnt); + yGm.SetValue(cnt,curr); + k++; + cnt++; + } + while(kcolumnLength){ + yGm.SetValue(cnt,(DTYPE_X)0); + k++; + cnt++; + } + } + } + } + + __aicore__ inline void FastPath(){ + uint32_t GmOffset=0; + int32_t init_row = 0; + for(int num=0;nummatrixNum;num++){ + uint32_t calLength=this->initLength; + if(this->diagVal<=0){ + init_row = -diagVal; + } + for (int32_t i = 0; i < this->loopCnt-1; i++) { + CopyIn(GmOffset,this->fullTileLength); + Compute(this->fullCnt, calLength, init_row); + CopyOut(GmOffset,this->fullTileLength); + if(init_row>0){ + init_row-=this->fullRowInc; + if(init_row<0){ + calLength-=init_row; + } + }else{ + calLength+=this->fullRowInc; + } + GmOffset+=this->fullTileLength; + } + CopyIn(GmOffset,this->lastTileLength); + Compute(this->lastCnt, calLength, init_row); + CopyOut(GmOffset,this->lastTileLength); + GmOffset+=this->lastTileLength; + } + } + + __aicore__ inline void CopyIn(uint32_t GmOffset, uint32_t tileLength){ + auto xLocal = inQueueX.AllocTensor(); + DataCopy(xLocal, xGm[GmOffset], tileLength); + inQueueX.EnQue(xLocal); + } + + __aicore__ inline void CopyOut(uint32_t GmOffset, uint32_t tileLength){ + auto yLocal=outQueueY.DeQue(); + DataCopy(yGm[GmOffset], yLocal, tileLength); + outQueueY.FreeTensor(yLocal); + } + + __aicore__ inline void Compute(int32_t cnt, uint32_t initLength, int32_t adjust){ + auto xLocal = inQueueX.DeQue(); + auto yLocal = outQueueY.AllocTensor(); + uint32_t localOffset=0; + uint32_t currLength=initLength; + DTYPE_X scalarZero=0; + for(int32_t i=0;icolumnLength); + if(i>=adjust){ + Adds(yLocal[localOffset],xLocal[localOffset],scalarZero,currLength); + currLength++; + } + localOffset+=this->columnLength; + } + outQueueY.EnQue(yLocal); + inQueueX.FreeTensor(xLocal); + } + + __aicore__ inline void AllZero(uint32_t tileLength){ + auto xLocal = inQueueX.DeQue(); + auto yLocal = outQueueY.AllocTensor(); + Sub(yLocal,xLocal,xLocal,tileLength); + outQueueY.EnQue(yLocal); + inQueueX.FreeTensor(xLocal); + } + +private: + TPipe pipe; + //queue for simple duplication + TQueBind queBind; // Use TQueBind to replace QueI,QueO if needed + + TQue inQueueX; + TQue outQueueY; + + GlobalTensor xGm; + GlobalTensor yGm; + + int32_t matrixNum; + int32_t matrixSize; + int32_t rowLength; + int32_t columnLength; + int32_t diagVal; + int32_t fullCnt; + int32_t lastCnt; + + int32_t loopCnt; + uint32_t fullTileLength; + uint32_t lastTileLength; + uint32_t fullRowInc; + uint32_t initLength; + + uint32_t typeSize; + uint32_t alignNum; + uint32_t key; +}; + +extern "C" __global__ __aicore__ void tril(GM_ADDR x, GM_ADDR y, GM_ADDR workspace, GM_ADDR tiling) { + GET_TILING_DATA(tiling_data, tiling); + KernelTril op; + IntegrateParam paramList = { + .totalLengthAligned=tiling_data.totalLengthAligned, + .matrixNum=tiling_data.matrixNum, + .matrixSize=tiling_data.matrixSize, + .rowLength=tiling_data.rowLength, + .columnLength=tiling_data.columnLength, + .diagVal=tiling_data.diagVal, + .loopCnt=tiling_data.loopCnt, + .fullTileLength=tiling_data.fullTileLength, + .lastTileLength=tiling_data.lastTileLength, + .fullCnt=tiling_data.fullCnt, + .lastCnt=tiling_data.lastCnt, + .alignNum=tiling_data.alignNum, + .typeSize=tiling_data.typeSize + }; + if(TILING_KEY_IS(1)){ + op.Init(x, y, paramList, 1); + }else if(TILING_KEY_IS(2)){ + op.Init(x, y, paramList, 2); + }else if(TILING_KEY_IS(3)){ + op.Init(x, y, paramList, 3); + }else if(TILING_KEY_IS(4)){ + op.Init(x, y, paramList, 4); + } + op.Process(); +} diff --git a/atb_operator/AtbTrilSample/TrilOp/scripts/install.sh b/atb_operator/AtbTrilSample/TrilOp/scripts/install.sh new file mode 100755 index 000000000..8468c5a25 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/scripts/install.sh @@ -0,0 +1,318 @@ +#!/bin/bash +vendor_name=customize +targetdir=/usr/local/Ascend/opp +target_custom=0 + +sourcedir=$PWD/packages +vendordir=vendors/$vendor_name + +QUIET="y" + +while true +do + case $1 in + --quiet) + QUIET="y" + shift + ;; + --install-path=*) + INSTALL_PATH=$(echo $1 | cut -d"=" -f2-) + INSTALL_PATH=${INSTALL_PATH%*/} + shift + ;; + --*) + shift + ;; + *) + break + ;; + esac +done + +log() { + cur_date=`date +"%Y-%m-%d %H:%M:%S"` + echo "[runtime] [$cur_date] "$1 +} + +if [ -n "${INSTALL_PATH}" ]; then + if [[ ! "${INSTALL_PATH}" = /* ]]; then + log "[ERROR] use absolute path for --install-path argument" + exit 1 + fi + if [ ! -d ${INSTALL_PATH} ]; then + mkdir ${INSTALL_PATH} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${INSTALL_PATH} failed" + exit 1 + fi + fi + targetdir=${INSTALL_PATH} +elif [ -n "${ASCEND_CUSTOM_OPP_PATH}" ]; then + if [ ! -d ${ASCEND_CUSTOM_OPP_PATH} ]; then + mkdir -p ${ASCEND_CUSTOM_OPP_PATH} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${ASCEND_CUSTOM_OPP_PATH} failed" + fi + fi + targetdir=${ASCEND_CUSTOM_OPP_PATH} +else + if [ "x${ASCEND_OPP_PATH}" == "x" ]; then + log "[ERROR] env ASCEND_OPP_PATH no exist" + exit 1 + fi + targetdir="${ASCEND_OPP_PATH}" +fi + +if [ ! -d $targetdir ];then + log "[ERROR] $targetdir no exist" + exit 1 +fi + +upgrade() +{ + if [ ! -d ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 files" + return 0 + fi + + if [ ! -d ${targetdir}/$vendordir/$1 ];then + log "[INFO] create ${targetdir}/$vendordir/$1." + mkdir -p ${targetdir}/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/$1 failed" + return 1 + fi + else + has_same_file=-1 + for file_a in ${sourcedir}/$vendordir/$1/*; do + file_b=${file_a##*/}; + if [ "ls ${targetdir}/$vendordir/$1" = "" ]; then + log "[INFO] ${targetdir}/$vendordir/$1 is empty !!" + return 1 + fi + grep -q $file_b <<<`ls ${targetdir}/$vendordir/$1`; + if [[ $? -eq 0 ]]; then + echo -n "${file_b} " + has_same_file=0 + fi + done + if [ 0 -eq $has_same_file ]; then + if test $QUIET = "n"; then + echo "[INFO]: has old version in ${targetdir}/$vendordir/$1, \ + you want to Overlay Installation , please enter:[o]; \ + or replace directory installation , please enter: [r]; \ + or not install , please enter:[n]." + + while true + do + read orn + if [ "$orn" = n ]; then + return 0 + elif [ "$orn" = m ]; then + break; + elif [ "$0rn" = r ]; then + [ -n "${targetdir}/$vendordir/$1/" ] && rm -rf "${targetdir}/$vendordir/$1"/* + break; + else + echo "[ERROR] input error, please input again!" + fi + done + fi + fi + log "[INFO] replace or merge old ops $1 files .g....." + fi + + log "copy new ops $1 files ......" + if [ -d ${targetdir}/$vendordir/$1/ ]; then + chmod -R +w "$targetdir/$vendordir/$1/" >/dev/null 2>&1 + fi + cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 files failed" + return 1 + fi + + return 0 +} +upgrade_proto() +{ + if [ ! -f ${sourcedir}/$vendordir/custom.proto ]; then + log "[INFO] no need to upgrade custom.proto files" + return 0 + fi + if [ ! -d ${targetdir}/$vendordir/framework/caffe ];then + log "[INFO] create ${targetdir}/$vendordir/framework/caffe." + mkdir -p ${targetdir}/$vendordir/framework/caffe + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/framework/caffe failed" + return 1 + fi + else + if [ -f ${targetdir}/$vendordir/framework/caffe/custom.proto ]; then + # 有老版本,判断是否要覆盖式安装 + if test $QUIET = "n"; then + echo "[INFO] ${targetdir}/$vendordir/framework/caffe has old version"\ + "custom.proto file. Do you want to replace? [y/n] " + + while true + do + read yn + if [ "$yn" = n ]; then + return 0 + elif [ "$yn" = y ]; then + break; + else + echo "[ERROR] input error, please input again!" + fi + done + fi + fi + log "[INFO] replace old caffe.proto files ......" + fi + chmod -R +w "$targetdir/$vendordir/framework/caffe/" >/dev/null 2>&1 + cp -rf ${sourcedir}/$vendordir/custom.proto ${targetdir}/$vendordir/framework/caffe/ + if [ $? -ne 0 ];then + log "[ERROR] copy new custom.proto failed" + return 1 + fi + log "[INFO] copy custom.proto success" + + return 0 +} + +upgrade_file() +{ + if [ ! -e ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 file" + return 0 + fi + + log "copy new $1 files ......" + cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 file failed" + return 1 + fi + + return 0 +} + +delete_optiling_file() +{ + if [ ! -d ${targetdir}/vendors ];then + log "[INFO] $1 not exist, no need to uninstall" + return 0 + fi + sys_info=$(uname -m) + if [ ! -d ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/lib/linux/${sys_info} ];then + rm -rf ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/liboptiling.so + fi + return 0 +} + +log "[INFO] copy uninstall sh success" + +if [ ! -d ${targetdir}/vendors ];then + log "[INFO] create ${targetdir}/vendors." + mkdir -p ${targetdir}/vendors + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/vendors failed" + return 1 + fi +fi +chmod u+w ${targetdir}/vendors + +echo "[ops_custom]upgrade framework" +upgrade framework +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op proto" +upgrade op_proto +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade version.info" +upgrade_file version.info +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op impl" +delete_optiling_file op_impl +upgrade op_impl +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op api" +upgrade op_api +if [ $? -ne 0 ];then + exit 1 +fi + +upgrade_proto +if [ $? -ne 0 ];then + exit 1 +fi + +# set the set_env.bash +if [ -n "${INSTALL_PATH}" ] && [ -d ${INSTALL_PATH} ]; then + _ASCEND_CUSTOM_OPP_PATH=${targetdir}/${vendordir} + bin_path="${_ASCEND_CUSTOM_OPP_PATH}/bin" + set_env_variable="#!/bin/bash\nexport ASCEND_CUSTOM_OPP_PATH=${_ASCEND_CUSTOM_OPP_PATH}:\${ASCEND_CUSTOM_OPP_PATH}" + if [ ! -d ${bin_path} ]; then + mkdir -p ${bin_path} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${bin_path} failed" + exit 1 + fi + fi + echo -e ${set_env_variable} > ${bin_path}/set_env.bash + if [ $? -ne 0 ]; then + log "[ERROR] write ASCEND_CUSTOM_OPP_PATH to set_env.bash failed" + exit 1 + else + log "[INFO] using requirements: when custom module install finished or before you run the custom module, \ + execute the command [ source ${bin_path}/set_env.bash ] to set the environment path" + fi +else + config_file=${targetdir}/vendors/config.ini + if [ ! -f ${config_file} ]; then + touch ${config_file} + chmod 640 ${config_file} + echo "load_priority=$vendor_name" > ${config_file} + if [ $? -ne 0 ];then + echo "echo load_priority failed" + exit 1 + fi + else + found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" + found_vendor=$(echo $found_vendors | sed "s/$vendor_name//g" | tr ',' ' ') + vendor=$(echo $found_vendor | tr -s ' ' ',') + if [ "$vendor" != "" ]; then + sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" + fi + fi +fi + +chmod u-w ${targetdir}/vendors + +if [ -d ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/ ]; then + chmod -R 440 ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/* >/dev/null 2>&1 +fi +if [ -f ${targetdir}/ascend_install.info ]; then + chmod -R 440 ${targetdir}/ascend_install.info +fi +if [ -f ${targetdir}/scene.info ]; then + chmod -R 440 ${targetdir}/scene.info +fi +if [ -f ${targetdir}/version.info ]; then + chmod -R 440 ${targetdir}/version.info +fi + +echo "SUCCESS" +exit 0 + diff --git a/atb_operator/AtbTrilSample/TrilOp/scripts/upgrade.sh b/atb_operator/AtbTrilSample/TrilOp/scripts/upgrade.sh new file mode 100755 index 000000000..e09173485 --- /dev/null +++ b/atb_operator/AtbTrilSample/TrilOp/scripts/upgrade.sh @@ -0,0 +1,151 @@ +#!/bin/bash +vendor_name=customize +targetdir=/usr/local/Ascend/opp +target_custom=0 + +sourcedir=$PWD/packages +vendordir=vendors/$vendor_name + +log() { + cur_date=`date +"%Y-%m-%d %H:%M:%S"` + echo "[runtime] [$cur_date] "$1 +} + +if [[ "x${ASCEND_OPP_PATH}" == "x" ]];then + log "[ERROR] env ASCEND_OPP_PATH no exist" + exit 1 +fi + +targetdir=${ASCEND_OPP_PATH} + +if [ ! -d $targetdir ];then + log "[ERROR] $targetdir no exist" + exit 1 +fi + +upgrade() +{ + if [ ! -d ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 files" + return 0 + fi + + if [ ! -d ${targetdir}/$vendordir/$1 ];then + log "[INFO] create ${targetdir}/$vendordir/$1." + mkdir -p ${targetdir}/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/$1 failed" + return 1 + fi + else + vendor_installed_dir=$(ls "$targetdir/vendors" 2> /dev/null) + for i in $vendor_installed_dir;do + vendor_installed_file=$(ls "$vendor_installed_dir/$vendor_name/$i" 2> /dev/null) + if [ "$i" = "$vendor_name" ] && [ "$vendor_installed_file" != "" ]; then + echo "[INFO]: $vendor_name custom opp package has been installed on the path $vendor_installed_dir, \ + you want to Overlay Installation , please enter:[o]; \ + or replace directory installation , please enter: [r]; \ + or not install , please enter:[n]." + fi + while true + do + read mrn + if [ "$mrn" = m ]; then + break + elif [ "$mrn" = r ]; then + [ -n "$vendor_installed_file"] && rm -rf "$vendor_installed_file" + break + elif [ "$mrn" = n ]; then + return 0 + else + echo "[WARNING]: Input error, please input m or r or n to choose!" + fi + done + done + log "[INFO] replace old ops $1 files ......" + fi + + log "copy new ops $1 files ......" + cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 files failed" + return 1 + fi + + return 0 +} + +upgrade_file() +{ + if [ ! -e ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 file" + return 0 + fi + + log "copy new $1 files ......" + cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 file failed" + return 1 + fi + + return 0 +} + +log "[INFO] copy uninstall sh success" + +echo "[ops_custom]upgrade framework" +upgrade framework +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op proto" +upgrade op_proto +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op impl" +upgrade op_impl +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op api" +upgrade op_api +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade version.info" +upgrade_file version.info +if [ $? -ne 0 ];then + exit 1 +fi + +config_file=${targetdir}/vendors/config.ini +found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" +found_vendor=$(echo $found_vendors | sed "s/$vendor_name//g" | tr ',' ' ') +vendor=$(echo $found_vendor | tr -s ' ' ',') +if [ "$vendor" != "" ]; then + sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" +fi + +changemode() +{ + if [ -d ${targetdir} ];then + chmod -R 550 ${targetdir}>/dev/null 2>&1 + fi + + return 0 +} +echo "[ops_custom]changemode..." +#changemode +if [ $? -ne 0 ];then + exit 1 +fi + +echo "SUCCESS" +exit 0 + diff --git a/atb_operator/README.md b/atb_operator/README.md new file mode 100644 index 000000000..6155a664d --- /dev/null +++ b/atb_operator/README.md @@ -0,0 +1,6 @@ + +| 目录名称 | 功能描述 | 运行环境 | 支持CANN版本 | +|------------------------------------------------------------|------------------------------------------| -- |---------------------| +| [AtbAddSample](./AtbAddSample) | Atb调用无属性算子样例 |Atlas A2训练系列产品| 8.0.0.alpha003及以上 | +| [AtbReduceSum](./AtbReduceSum) | Atb调用有属性算子样例 | Atlas A2训练系列产品 | 8.0.0.alpha003及以上 | + -- Gitee From 9b558fb8eeeef51953f9eaba0b9ae7846d8ccab0 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E5=88=98=E5=96=9C=E5=BC=BA?= Date: Mon, 31 Mar 2025 06:17:04 +0000 Subject: [PATCH 02/26] update --- .../AtbAddSample/AddAscendC/CMakeLists.txt | 69 ++ .../AtbAddSample/AddAscendC/CMakePresets.json | 63 ++ .../AtbAddSample/AddAscendC/README.md | 240 +++++ .../AtbAddSample/AddAscendC/build.sh | 37 + .../AddAscendC/cmake/config.cmake | 25 + .../AtbAddSample/AddAscendC/cmake/func.cmake | 228 +++++ .../AtbAddSample/AddAscendC/cmake/intf.cmake | 26 + .../AddAscendC/cmake/makeself.cmake | 17 + .../AddAscendC/cmake/util/__init__.py | 8 + .../cmake/util/ascendc_bin_param_build.py | 129 +++ .../cmake/util/ascendc_impl_build.py | 446 ++++++++++ .../cmake/util/ascendc_ops_config.py | 114 +++ .../cmake/util/ascendc_replay_build.py | 65 ++ .../cmake/util/batch_replay_impl.temp | 117 +++ .../cmake/util/code_channel_infer.py | 115 +++ .../AddAscendC/cmake/util/const_var.py | 33 + .../cmake/util/gen_impl_and_mrege_json.sh | 57 ++ .../AddAscendC/cmake/util/gen_ops_filter.sh | 62 ++ .../AddAscendC/cmake/util/gen_version_info.sh | 6 + .../AddAscendC/cmake/util/insert_op_info.py | 36 + .../cmake/util/insert_simplified_keys.py | 248 ++++++ .../AddAscendC/cmake/util/kernel_entry.py | 115 +++ .../AddAscendC/cmake/util/kernel_impl.temp | 10 + .../AddAscendC/cmake/util/makeself/COPYING | 339 ++++++++ .../AddAscendC/cmake/util/makeself/README.md | 246 ++++++ .../AddAscendC/cmake/util/makeself/VERSION | 1 + .../cmake/util/makeself/make-release.sh | 9 + .../cmake/util/makeself/makeself-header.sh | 660 ++++++++++++++ .../AddAscendC/cmake/util/makeself/makeself.1 | 110 +++ .../cmake/util/makeself/makeself.lsm | 16 + .../cmake/util/makeself/makeself.sh | 822 ++++++++++++++++++ .../cmake/util/makeself/run-tests.sh | 8 + .../cmake/util/merge_aicpu_info_json.sh | 31 + .../AddAscendC/cmake/util/opdesc_parser.py | 260 ++++++ .../cmake/util/parse_ini_to_json.py | 338 +++++++ .../AddAscendC/cmake/util/preset_parse.py | 23 + .../AddAscendC/cmake/util/replay_codegen.py | 105 +++ .../AddAscendC/cmake/util/replay_impl.temp | 120 +++ .../cmake/util/tiling_data_def_build.py | 84 ++ .../AddAscendC/framework/CMakeLists.txt | 11 + .../framework/tf_plugin/CMakeLists.txt | 14 + .../tf_plugin/tensorflow_add_custom_plugin.cc | 23 + .../AddAscendC/op_host/CMakeLists.txt | 82 ++ .../AddAscendC/op_host/add_custom.cpp | 117 +++ .../AddAscendC/op_host/add_custom_tiling.h | 22 + .../AddAscendC/op_kernel/CMakeLists.txt | 61 ++ .../AddAscendC/op_kernel/add_custom.cpp | 134 +++ .../AddAscendC/scripts/install.sh | 318 +++++++ .../AddAscendC/scripts/upgrade.sh | 151 ++++ .../AddOperationATBPlugin/CMakeLists.txt | 21 + .../aclnn_add_operation.cpp | 142 +++ .../aclnn_add_operation.h | 56 ++ .../AddOperationATBPlugin/build.sh | 33 + .../AddOperationTest/CMakeLists.txt | 40 + .../AddOperationTest/script/gen_data.py | 20 + .../AddOperationTest/script/run.sh | 52 ++ .../AddOperationTest/script/verify_result.py | 28 + .../AddOperationTest/src/main.cpp | 217 +++++ .../AtbAddSample/AddOperationTest/src/main.h | 55 ++ .../AtbAddSample/AtbAddSample/readme.md | 81 ++ 60 files changed, 7116 insertions(+) create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakePresets.json create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/README.md create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/build.sh create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/config.cmake create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/func.cmake create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/intf.cmake create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/makeself.cmake create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/__init__.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/const_var.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/README.md create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/preset_parse.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom.cpp create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/install.sh create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/upgrade.sh create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h create mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/build.sh create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/gen_data.py create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/run.sh create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/verify_result.py create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.cpp create mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.h create mode 100644 atb_operator/AtbAddSample/AtbAddSample/readme.md diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakeLists.txt new file mode 100644 index 000000000..584132d80 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakeLists.txt @@ -0,0 +1,69 @@ +cmake_minimum_required(VERSION 3.16.0) +project(opp) +if(ENABLE_CROSS_COMPILE) + if(${CMAKE_SYSTEM_PROCESSOR} STREQUAL x86_64) + set(CROSS_COMPILE_PLATFORM aarch64) + else() + set(CROSS_COMPILE_PLATFORM x86_64) + endif() + set(PLATFORM ${CMAKE_SYSTEM_PROCESSOR}) + set(CMAKE_COMPILE_COMPILER_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/linux/${CROSS_COMPILE_PLATFORM}/) + set(CMAKE_COMPILE_RUNTIME_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/${CROSS_COMPILE_PLATFORM}/) + set(CMAKE_SYSTEM_PROCESSOR ${CROSS_COMPILE_PLATFORM}) + set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) + set(CMAKE_CXX_COMPILER ${CMAKE_CROSS_PLATFORM_COMPILER}) +else() + set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) +endif() + +include(cmake/config.cmake) +include(cmake/func.cmake) +include(cmake/intf.cmake) + +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/framework) + add_subdirectory(framework) +endif() +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_host) + add_subdirectory(op_host) +endif() +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_kernel) + add_subdirectory(op_kernel) +endif() +if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) + add_subdirectory(testcases) +endif() + +# modify vendor_name in install.sh and upgrade.sh +add_custom_command(OUTPUT ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh + COMMAND mkdir -p ${CMAKE_BINARY_DIR}/scripts + COMMAND cp -r ${CMAKE_SOURCE_DIR}/scripts/* ${CMAKE_BINARY_DIR}/scripts/ + COMMAND sed -i "s/vendor_name=customize/vendor_name=${vendor_name}/g" ${CMAKE_BINARY_DIR}/scripts/* +) +add_custom_target(modify_vendor ALL DEPENDS ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh) +install(DIRECTORY ${CMAKE_BINARY_DIR}/scripts/ DESTINATION . FILE_PERMISSIONS OWNER_EXECUTE OWNER_READ GROUP_READ) + +install(FILES ${CMAKE_SOURCE_DIR}/custom.proto DESTINATION packages OPTIONAL) + +get_system_info(SYSTEM_INFO) + +# gen version.info +add_custom_target(gen_version_info ALL + COMMAND bash ${CMAKE_CURRENT_SOURCE_DIR}/cmake/util/gen_version_info.sh ${ASCEND_CANN_PACKAGE_PATH} ${CMAKE_CURRENT_BINARY_DIR} +) + +install(FILES ${CMAKE_CURRENT_BINARY_DIR}/version.info + DESTINATION packages/vendors/${vendor_name}/) + +# CPack config +set(CPACK_PACKAGE_NAME ${CMAKE_PROJECT_NAME}) +set(CPACK_PACKAGE_VERSION ${CMAKE_PROJECT_VERSION}) +set(CPACK_PACKAGE_DESCRIPTION "CPack opp project") +set(CPACK_PACKAGE_DESCRIPTION_SUMMARY "CPack opp project") +set(CPACK_PACKAGE_DIRECTORY ${CMAKE_INSTALL_PREFIX}) +set(CPACK_PACKAGE_FILE_NAME "custom_opp_${SYSTEM_INFO}.run") +set(CPACK_GENERATOR External) +set(CPACK_CMAKE_GENERATOR "Unix Makefiles") +set(CPACK_EXTERNAL_ENABLE_STAGING TRUE) +set(CPACK_EXTERNAL_PACKAGE_SCRIPT ${CMAKE_SOURCE_DIR}/cmake/makeself.cmake) +set(CPACK_EXTERNAL_BUILT_PACKAGES ${CPACK_PACKAGE_DIRECTORY}/_CPack_Packages/Linux/External/${CPACK_PACKAGE_FILE_NAME}/${CPACK_PACKAGE_FILE_NAME}) +include(CPack) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakePresets.json b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakePresets.json new file mode 100644 index 000000000..ddec9431b --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakePresets.json @@ -0,0 +1,63 @@ +{ + "version": 1, + "cmakeMinimumRequired": { + "major": 3, + "minor": 19, + "patch": 0 + }, + "configurePresets": [ + { + "name": "default", + "displayName": "Default Config", + "description": "Default build using Unix Makefiles generator", + "generator": "Unix Makefiles", + "binaryDir": "${sourceDir}/build_out", + "cacheVariables": { + "CMAKE_BUILD_TYPE": { + "type": "STRING", + "value": "Release" + }, + "ENABLE_SOURCE_PACKAGE": { + "type": "BOOL", + "value": "True" + }, + "ENABLE_BINARY_PACKAGE": { + "type": "BOOL", + "value": "True" + }, + "ASCEND_COMPUTE_UNIT": { + "type": "STRING", + "value": "ascend310b;ascend910b" + }, + "ENABLE_TEST": { + "type": "BOOL", + "value": "True" + }, + "vendor_name": { + "type": "STRING", + "value": "customize" + }, + "ASCEND_CANN_PACKAGE_PATH": { + "type": "PATH", + "value": "/usr/local/Ascend/ascend-toolkit/latest" + }, + "ASCEND_PYTHON_EXECUTABLE": { + "type": "STRING", + "value": "python3" + }, + "CMAKE_INSTALL_PREFIX": { + "type": "PATH", + "value": "${sourceDir}/build_out" + }, + "ENABLE_CROSS_COMPILE": { + "type": "BOOL", + "value": "False" + }, + "CMAKE_CROSS_PLATFORM_COMPILER": { + "type": "PATH", + "value": "/usr/bin/aarch64-linux-gnu-g++" + } + } + } + ] +} diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/README.md b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/README.md new file mode 100644 index 000000000..da59a83e7 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/README.md @@ -0,0 +1,240 @@ +<<<<<<< HEAD +## 概述 +本样例基于AddCustom算子工程,介绍了单算子工程及单算子调用。 + +## 算子描述 +======= +# 概述 +本样例基于AddCustom算子工程,介绍了单算子工程及aclnn接口说明。 + +## 算子工程介绍 +### 算子描述 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd +Add算子实现了两个数据相加,返回相加结果的功能,该算子实现了非32B对齐shape下的算子功能,完成了算子的泛化实现。 +对应的数学表达式为: +``` +z = x + y +``` +<<<<<<< HEAD +## 算子规格描述 +======= +### 算子规格描述 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd + + + + + + + + + + + +
算子类型(OpType)Add
算子输入nameshapedata typeformat
x-float32,float16,int32,int8ND
y-float32,float16,int32,int8ND
算子输出z-float32,float16,int32,int8ND
核函数名add_custom
+ + +<<<<<<< HEAD +## 算子工程介绍 +======= +### 算子工程文件介绍 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd +其中,算子工程目录AddCustom包含算子实现的模板文件、编译脚本等,如下所示: +``` +├── AddCustom //Add自定义算子工程 +│ ├── cmake +│ ├── framework // 算子插件实现文件目录,单算子模型文件的生成不依赖算子适配插件,无需关注 +│ ├── op_host // host侧实现文件 +│ ├── op_kernel // kernel侧实现文件 +│ ├── scripts // 自定义算子工程打包相关脚本所在目录 +│ ├── build.sh // 编译入口脚本 +│ ├── CMakeLists.txt // 算子工程的CMakeLists.txt +│ └── CMakePresets.json // 编译配置项 +``` +<<<<<<< HEAD +CANN软件包中提供了工程创建工具msopgen,AddCustom算子工程可通过AddCustom.json自动创建,具体请参考[Ascend C算子开发](https://hiascend.com/document/redirect/CannCommunityOpdevAscendC)>算子开发>算子开发工程>基于自定义算子工程的算子开发>创建算子工程 章节。 +## 编译运行样例算子 +======= + +### 编译运行样例算子 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd +针对自定义算子工程,编译运行包含如下步骤: +- 编译自定义算子工程生成算子安装包; +- 安装自定义算子到算子库中; +- 调用执行自定义算子; + +详细操作如下所示。 +<<<<<<< HEAD +### 1. 获取源码包 +编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 +### 2. 编译算子工程 +======= +#### 1. 获取源码包 +编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 +#### 2. 编译算子工程 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd + 编译自定义算子工程,构建生成自定义算子包。 + + - 执行如下命令,切换到算子工程AddCustom目录。 + + ```bash + cd ${git_clone_path}/samples/atb_operator/AtbAddSample/AddOp + ``` + + - 修改CMakePresets.json中ASCEND_CANN_PACKAGE_PATH为CANN软件包安装后的实际路径。 + + + ```json + { + …… + "configurePresets": [ + { + …… + "ASCEND_CANN_PACKAGE_PATH": { + "type": "PATH", + "value": "/usr/local/Ascend/ascend-toolkit/latest" //请替换为CANN软件包安装后的实际路径。eg:/home/HwHiAiUser/Ascend/ascend-toolkit/latest + }, + …… + } + ] + } + ``` + - 在算子工程AddOp目录下执行如下命令,进行算子工程编译。 + + ```bash + ./build.sh + ``` +编译成功后,会在当前目录下创建build_out目录,并在build_out目录下生成自定义算子安装包custom_opp_\_\.run,例如“custom_opp_ubuntu_x86_64.run”。 + +备注:如果要使用dump调试功能,需要移除op_host内和CMakeLists.txt内的Atlas 200/500 A2 推理产品的配置。 + +<<<<<<< HEAD +### 3. 部署算子包 +======= +#### 3. 部署算子包 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd + +执行如下命令,在自定义算子安装包所在路径下,安装自定义算子包。 + ```bash + cd build_out + ./custom_opp__.run + ``` +命令执行成功后,自定义算子包中的相关文件将部署至当前环境的OPP算子库的vendors/customize目录中。 + +<<<<<<< HEAD +### 4. 配置环境变量 +======= +#### 4. 配置环境变量 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd + + 请根据当前环境上CANN开发套件包的[安装方式](https://hiascend.com/document/redirect/CannCommunityInstSoftware),选择对应配置环境变量的命令。 + - 默认路径,root用户安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=/usr/local/Ascend/ascend-toolkit/latest + ``` + - 默认路径,非root用户安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=$HOME/Ascend/ascend-toolkit/latest + ``` + - 指定路径install_path,安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=${install_path}/ascend-toolkit/latest + ``` + +## aclnn接口说明 + +自定义算子编译部署后,会自动生成单算子API,可以直接在应用程序中调用。算子API的形式一般定义为“两段式接口”,形如: + ```cpp + aclnnStatus aclnnAddCustomGetWorkspaceSize(const aclTensor *x, const aclTensor *y, const alcTensor *out, uint64_t workspaceSize, aclOpExecutor **executor); + aclnnStatus aclnnAddCustom(void *workspace, int64_t workspaceSize, aclOpExecutor **executor, aclrtStream stream); + ``` +其中aclnnAddCustomGetWorkspaceSize为第一段接口,主要用于计算本次API调用计算过程中需要多少的workspace内存。获取到本次API计算需要的workspace大小之后,按照workspaceSize大小申请Device侧内存,然后调用第二段接口aclnnAddCustom执行计算。具体参考[AscendCL单算子调用](https://hiascend.com/document/redirect/CannCommunityAscendCInVorkSingleOp)>单算子API执行 章节。 + +<<<<<<< HEAD +======= +### 功能描述 +* 算子功能:完成加法计算。 +* 计算公式: +``` +z = x + y +``` + +#### 参数说明 +##### aclnnAddCustomGetWorkspaceSize: + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
接口参数描述
aclnnAddCustomGetWorkspaceSizex输入x的Tensor,支持flaot/half/int8/int32类型,ND排布格式
y输入y的Tensor,支持flaot/half/int8/int32类型,ND排布格式
out输出z的Tensor,支持flaot/half/int8/int32类型,ND排布格式
workspaceSize返回需要在Device侧申请的workspace大小
executor返回op执行器,包含了算子计算流程
+ + +| 返回值 | 描述 | +| ---- | ---- | +| ACL_SUCCESS | 成功 | +| ACL_ERROR | 失败 | + +##### aclnnAddCustom + + + + + + + + + + + + + + + + + + + + + + + +
接口参数描述
aclnnAddCustomworkspace输入Device侧申请的workspace内存
workspaceSizeworkspace大小
executor输入op执行器,包含了算子计算流程
stream输入执行流
+ +| 返回值 | 描述 | +| ---- | ---- | +| ACL_SUCCESS | 成功 | +| ACL_ERROR | 失败 | + +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd + +## 更新说明 +| 时间 | 更新事项 | +|----|------| +<<<<<<< HEAD +| 2025/03/27 | 修改不同用户环境配置 | +======= +| 2025/03/27 | 新增README | +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/build.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/build.sh new file mode 100755 index 000000000..4be96d7d8 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/build.sh @@ -0,0 +1,37 @@ +#!/bin/bash +script_path=$(realpath $(dirname $0)) + + +mkdir -p build_out +rm -rf build_out/* +cd build_out + +cmake_version=$(cmake --version | grep "cmake version" | awk '{print $3}') +if [ "$cmake_version" \< "3.19.0" ] ; then + opts=$(python3 $script_path/cmake/util/preset_parse.py $script_path/CMakePresets.json) + echo $opts + cmake .. $opts +else + cmake .. --preset=default +fi +target=package +if [ "$1"x != ""x ]; then target=$1; fi + +cmake --build . --target $target -j16 +if [ $? -ne 0 ]; then exit 1; fi + +if [ $target = "package" ]; then + if test -d ./op_kernel/binary ; then + ./cust*.run + if [ $? -ne 0 ]; then exit 1; fi + cmake --build . --target binary -j16 + if [ $? -ne 0 ]; then exit 1; fi + cmake --build . --target $target -j16 + fi +fi + +# for debug +# cd build_out +# make +# cpack +# verbose append -v diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/config.cmake b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/config.cmake new file mode 100755 index 000000000..886119daa --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/config.cmake @@ -0,0 +1,25 @@ + +set(CMAKE_CXX_FLAGS_DEBUG "") +set(CMAKE_CXX_FLAGS_RELEASE "") + +if (NOT DEFINED vendor_name) + set(vendor_name customize CACHE STRING "") +endif() +if (NOT DEFINED ASCEND_CANN_PACKAGE_PATH) + set(ASCEND_CANN_PACKAGE_PATH /usr/local/Ascend/latest CACHE PATH "") +endif() +if (NOT DEFINED ASCEND_PYTHON_EXECUTABLE) + set(ASCEND_PYTHON_EXECUTABLE python3 CACHE STRING "") +endif() +if (NOT DEFINED ASCEND_COMPUTE_UNIT) + message(FATAL_ERROR "ASCEND_COMPUTE_UNIT not set in CMakePreset.json ! +") +endif() +set(ASCEND_TENSOR_COMPILER_PATH ${ASCEND_CANN_PACKAGE_PATH}/compiler) +set(ASCEND_CCEC_COMPILER_PATH ${ASCEND_TENSOR_COMPILER_PATH}/ccec_compiler/bin) +set(ASCEND_AUTOGEN_PATH ${CMAKE_BINARY_DIR}/autogen) +set(ASCEND_FRAMEWORK_TYPE tensorflow) +file(MAKE_DIRECTORY ${ASCEND_AUTOGEN_PATH}) +set(CUSTOM_COMPILE_OPTIONS "custom_compile_options.ini") +execute_process(COMMAND rm -rf ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} + COMMAND touch ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS}) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/func.cmake b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/func.cmake new file mode 100755 index 000000000..ad187e7d6 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/func.cmake @@ -0,0 +1,228 @@ + +function(get_system_info SYSTEM_INFO) + if (UNIX) + execute_process(COMMAND grep -i ^id= /etc/os-release OUTPUT_VARIABLE TEMP) + string(REGEX REPLACE "\n|id=|ID=|\"" "" SYSTEM_NAME ${TEMP}) + set(${SYSTEM_INFO} ${SYSTEM_NAME}_${CMAKE_SYSTEM_PROCESSOR} PARENT_SCOPE) + elseif (WIN32) + message(STATUS "System is Windows. Only for pre-build.") + else () + message(FATAL_ERROR "${CMAKE_SYSTEM_NAME} not support.") + endif () +endfunction() + +function(opbuild) + message(STATUS "Opbuild generating sources") + cmake_parse_arguments(OPBUILD "" "OUT_DIR;PROJECT_NAME;ACCESS_PREFIX" "OPS_SRC" ${ARGN}) + execute_process(COMMAND ${CMAKE_COMPILE} -g -fPIC -shared -std=c++11 ${OPBUILD_OPS_SRC} -D_GLIBCXX_USE_CXX11_ABI=0 + -I ${ASCEND_CANN_PACKAGE_PATH}/include -L ${ASCEND_CANN_PACKAGE_PATH}/lib64 -lexe_graph -lregister -ltiling_api + -o ${OPBUILD_OUT_DIR}/libascend_all_ops.so + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("build ops lib info: ${EXEC_INFO}") + message("build ops lib error: ${EXEC_ERROR}") + message(FATAL_ERROR "opbuild run failed!") + endif() + set(proj_env "") + set(prefix_env "") + if (NOT "${OPBUILD_PROJECT_NAME}x" STREQUAL "x") + set(proj_env "OPS_PROJECT_NAME=${OPBUILD_PROJECT_NAME}") + endif() + if (NOT "${OPBUILD_ACCESS_PREFIX}x" STREQUAL "x") + set(prefix_env "OPS_DIRECT_ACCESS_PREFIX=${OPBUILD_ACCESS_PREFIX}") + endif() + execute_process(COMMAND ${proj_env} ${prefix_env} ${ASCEND_CANN_PACKAGE_PATH}/toolkit/tools/opbuild/op_build + ${OPBUILD_OUT_DIR}/libascend_all_ops.so ${OPBUILD_OUT_DIR} + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("opbuild ops info: ${EXEC_INFO}") + message("opbuild ops error: ${EXEC_ERROR}") + endif() + message(STATUS "Opbuild generating sources - done") +endfunction() + +function(add_ops_info_target) + cmake_parse_arguments(OPINFO "" "TARGET;OPS_INFO;OUTPUT;INSTALL_DIR" "" ${ARGN}) + get_filename_component(opinfo_file_path "${OPINFO_OUTPUT}" DIRECTORY) + add_custom_command(OUTPUT ${OPINFO_OUTPUT} + COMMAND mkdir -p ${opinfo_file_path} + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/parse_ini_to_json.py + ${OPINFO_OPS_INFO} ${OPINFO_OUTPUT} + ) + add_custom_target(${OPINFO_TARGET} ALL + DEPENDS ${OPINFO_OUTPUT} + ) + install(FILES ${OPINFO_OUTPUT} + DESTINATION ${OPINFO_INSTALL_DIR} + ) +endfunction() + +function(add_ops_compile_options OP_TYPE) + cmake_parse_arguments(OP_COMPILE "" "OP_TYPE" "COMPUTE_UNIT;OPTIONS" ${ARGN}) + file(APPEND ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} + "${OP_TYPE},${OP_COMPILE_COMPUTE_UNIT},${OP_COMPILE_OPTIONS}\n") +endfunction() + +function(add_ops_impl_target) + cmake_parse_arguments(OPIMPL "" "TARGET;OPS_INFO;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) + add_custom_command(OUTPUT ${OPIMPL_OUT_DIR}/.impl_timestamp + COMMAND mkdir -m 700 -p ${OPIMPL_OUT_DIR}/dynamic + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py + ${OPIMPL_OPS_INFO} + \"${OPIMPL_OPS_BATCH}\" \"${OPIMPL_OPS_ITERATE}\" + ${OPIMPL_IMPL_DIR} + ${OPIMPL_OUT_DIR}/dynamic + ${ASCEND_AUTOGEN_PATH} + + COMMAND rm -rf ${OPIMPL_OUT_DIR}/.impl_timestamp + COMMAND touch ${OPIMPL_OUT_DIR}/.impl_timestamp + DEPENDS ${OPIMPL_OPS_INFO} + ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py + ) + add_custom_target(${OPIMPL_TARGET} ALL + DEPENDS ${OPIMPL_OUT_DIR}/.impl_timestamp) + if (${ENABLE_SOURCE_PACKAGE}) + install(DIRECTORY ${OPIMPL_OUT_DIR}/dynamic + DESTINATION ${OPIMPL_INSTALL_DIR} + ) + endif() +endfunction() + +function(add_ops_replay_targets) + cmake_parse_arguments(OPREPLAY "" "OPS_INFO;COMPUTE_UNIT;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) + # ccec compile options + set(ccec_base_opts -c -O2 --cce-aicore-only -mllvm -cce-aicore-function-stack-size=16000 + -mllvm -cce-aicore-record-overflow=false -std=c++17) + set(ccec_extopts_ascend310p --cce-aicore-arch=dav-m200 -mllvm -cce-aicore-fp-ceiling=2) + set(ccec_extopts_ascend910 --cce-aicore-arch=dav-c100) + set(ccec_extopts_ascend910b --cce-aicore-arch=dav-c220-cube) + file(MAKE_DIRECTORY ${OPREPLAY_OUT_DIR}) + execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_replay_build.py + ${OPREPLAY_OPS_INFO} + "${OPREPLAY_OPS_BATCH}" "${OPREPLAY_OPS_ITERATE}" + ${OPREPLAY_IMPL_DIR} + ${OPREPLAY_OUT_DIR} + ${OPREPLAY_COMPUTE_UNIT} + ) + file(GLOB replay_kernel_entries ${OPREPLAY_OUT_DIR}/*.cce) + if (NOT "${replay_kernel_entries}x" STREQUAL "x") + foreach(replay_kernel_file ${replay_kernel_entries}) + get_filename_component(replay_kernel_file_name "${replay_kernel_file}" NAME) + string(REPLACE "_entry.cce" "" op_kerne_name ${replay_kernel_file_name}) + file(GLOB replay_lib_src ${OPREPLAY_OUT_DIR}/${op_kerne_name}*.cpp) + set(OP_TILING_DATA_H_PATH ${OPREPLAY_OUT_DIR}/${op_kerne_name}_tiling_data.h) + add_library(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} SHARED ${replay_lib_src}) + if(EXISTS ${OP_TILING_DATA_H_PATH}) + target_compile_options(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE + -include ${OP_TILING_DATA_H_PATH} + ) + endif() + target_compile_definitions(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE + ${op_kerne_name}=${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} + ) + target_compile_options(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE + -D__ASCENDC_REPLAY__ + ) + target_link_libraries(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE intf_pub + tikreplaylib::${OPREPLAY_COMPUTE_UNIT} + register + ) + add_custom_command(OUTPUT ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o + COMMAND ccec ${ccec_base_opts} ${ccec_extopts_${OPREPLAY_COMPUTE_UNIT}} ${replay_kernel_file} + -o ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o + DEPENDS ${replay_kernel_file} + ) + add_custom_target(replay_kernel_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} ALL + DEPENDS ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o + ) + install(TARGETS replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_replay + ) + install(FILES ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o + DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_replay + ) + endforeach() + endif() +endfunction() + +function(add_npu_support_target) + cmake_parse_arguments(NPUSUP "" "TARGET;OPS_INFO_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) + get_filename_component(npu_sup_file_path "${NPUSUP_OUT_DIR}" DIRECTORY) + add_custom_command(OUTPUT ${NPUSUP_OUT_DIR}/npu_supported_ops.json + COMMAND mkdir -p ${NPUSUP_OUT_DIR} + COMMAND ${CMAKE_SOURCE_DIR}/cmake/util/gen_ops_filter.sh + ${NPUSUP_OPS_INFO_DIR} + ${NPUSUP_OUT_DIR} + ) + add_custom_target(npu_supported_ops ALL + DEPENDS ${NPUSUP_OUT_DIR}/npu_supported_ops.json + ) + install(FILES ${NPUSUP_OUT_DIR}/npu_supported_ops.json + DESTINATION ${NPUSUP_INSTALL_DIR} + ) +endfunction() + +function(add_bin_compile_target) + cmake_parse_arguments(BINCMP "" "TARGET;OPS_INFO;COMPUTE_UNIT;IMPL_DIR;ADP_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/src) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/gen) + execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_bin_param_build.py + ${BINCMP_OPS_INFO} ${BINCMP_OUT_DIR}/gen ${BINCMP_COMPUTE_UNIT} + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("ops binary compile scripts gen info: ${EXEC_INFO}") + message("ops binary compile scripts gen error: ${EXEC_ERROR}") + message(FATAL_ERROR "ops binary compile scripts gen failed!") + endif() + if (NOT TARGET binary) + add_custom_target(binary) + endif() + add_custom_target(${BINCMP_TARGET} + COMMAND cp -r ${BINCMP_IMPL_DIR}/*.* ${BINCMP_OUT_DIR}/src + ) + add_custom_target(${BINCMP_TARGET}_gen_ops_config + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/insert_simplified_keys.py -p ${BINCMP_OUT_DIR}/bin + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_ops_config.py -p ${BINCMP_OUT_DIR}/bin + -s ${BINCMP_COMPUTE_UNIT} + ) + add_dependencies(binary ${BINCMP_TARGET}_gen_ops_config) + file(GLOB bin_scripts ${BINCMP_OUT_DIR}/gen/*.sh) + foreach(bin_script ${bin_scripts}) + get_filename_component(bin_file ${bin_script} NAME_WE) + string(REPLACE "-" ";" bin_sep ${bin_file}) + list(GET bin_sep 0 op_type) + list(GET bin_sep 1 op_file) + list(GET bin_sep 2 op_index) + if (NOT TARGET ${BINCMP_TARGET}_${op_file}_copy) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file}) + add_custom_target(${BINCMP_TARGET}_${op_file}_copy + COMMAND cp ${BINCMP_ADP_DIR}/${op_file}.py ${BINCMP_OUT_DIR}/src/${op_type}.py + ) + install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} + DESTINATION ${BINCMP_INSTALL_DIR}/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) + install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json + DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT}/ OPTIONAL + ) + endif() + add_custom_target(${BINCMP_TARGET}_${op_file}_${op_index} + COMMAND export HI_PYTHON=${ASCEND_PYTHON_EXECUTABLE} && bash ${bin_script} ${BINCMP_OUT_DIR}/src/${op_type}.py ${BINCMP_OUT_DIR}/bin/${op_file} + WORKING_DIRECTORY ${BINCMP_OUT_DIR} + ) + add_dependencies(${BINCMP_TARGET}_${op_file}_${op_index} ${BINCMP_TARGET} ${BINCMP_TARGET}_${op_file}_copy) + add_dependencies(${BINCMP_TARGET}_gen_ops_config ${BINCMP_TARGET}_${op_file}_${op_index}) + endforeach() + install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json + DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) +endfunction() diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/intf.cmake b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/intf.cmake new file mode 100755 index 000000000..2f362c396 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/intf.cmake @@ -0,0 +1,26 @@ + +add_library(intf_pub INTERFACE) +target_compile_options(intf_pub INTERFACE + -fPIC + -fvisibility=hidden + -fvisibility-inlines-hidden + $<$:-O2> + $<$:-O0 -g> + $<$:-std=c++11> + $<$,$>:-ftrapv -fstack-check> + $<$:-pthread -Wfloat-equal -Wshadow -Wformat=2 -Wno-deprecated -Wextra> + $,-fstack-protector-strong,-fstack-protector-all> +) +target_compile_definitions(intf_pub INTERFACE + _GLIBCXX_USE_CXX11_ABI=0 + $<$:_FORTIFY_SOURCE=2> +) +target_include_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/include) +target_link_options(intf_pub INTERFACE + $<$,EXECUTABLE>:-pie> + $<$:-s> + -Wl,-z,relro + -Wl,-z,now + -Wl,-z,noexecstack +) +target_link_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/lib64) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/makeself.cmake b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/makeself.cmake new file mode 100755 index 000000000..48c565bfb --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/makeself.cmake @@ -0,0 +1,17 @@ +execute_process(COMMAND chmod +x ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh) +execute_process(COMMAND ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh + --header ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself-header.sh + --help-header ./help.info + --gzip --complevel 4 --nomd5 --sha256 + ./ ${CPACK_PACKAGE_FILE_NAME} "version:1.0" ./install.sh + WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} + RESULT_VARIABLE EXEC_RESULT + ERROR_VARIABLE EXEC_ERROR +) +if (NOT "${EXEC_RESULT}x" STREQUAL "0x") + message(FATAL_ERROR "CPack Command error: ${EXEC_RESULT}\n${EXEC_ERROR}") +endif() +execute_process(COMMAND cp ${CPACK_EXTERNAL_BUILT_PACKAGES} ${CPACK_PACKAGE_DIRECTORY}/ + COMMAND echo "Copy ${CPACK_EXTERNAL_BUILT_PACKAGES} to ${CPACK_PACKAGE_DIRECTORY}/" + WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} +) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/__init__.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/__init__.py new file mode 100755 index 000000000..c4ddc893a --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/__init__.py @@ -0,0 +1,8 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- + +import sys +import os + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) +sys.path.append(PYF_PATH) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py new file mode 100755 index 000000000..decf34544 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py @@ -0,0 +1,129 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import json +import hashlib +import const_var +import opdesc_parser + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + + +class BinParamBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + super().__init__(op_type) + self.soc = '' + self.out_path = '' + + def set_soc_version(self: any, soc: str): + self.soc = soc + + def set_out_path(self: any, out_path: str): + self.out_path = out_path + + def gen_input_json(self: any): + key_map = {} + count = len(self.input_dtype[0].split(',')) + for i in range(0, count): + inputs = [] + outputs = [] + attrs = [] + op_node = {} + for idx in range(0, len(self.input_name)): + idtypes = self.input_dtype[idx].split(',') + ifmts = self.input_fmt[idx].split(',') + itype = self.input_type[idx] + para = {} + para['name'] = self.input_name[idx] + para['index'] = idx + para['dtype'] = idtypes[i] + para['format'] = ifmts[i] + para['paramType'] = itype + para['shape'] = [-2] + if itype == 'dynamic': + inputs.append([para]) + else: + inputs.append(para) + for idx in range(0, len(self.output_name)): + odtypes = self.output_dtype[idx].split(',') + ofmts = self.output_fmt[idx].split(',') + otype = self.output_type[idx] + para = {} + para['name'] = self.output_name[idx] + para['index'] = idx + para['dtype'] = odtypes[i] + para['format'] = ofmts[i] + para['paramType'] = otype + para['shape'] = [-2] + if otype == 'dynamic': + outputs.append([para]) + else: + outputs.append(para) + for attr in self.attr_list: + att = {} + att['name'] = attr + atype = self.attr_val.get(attr).get('type').lower() + atype = atype.replace('list', 'list_') + att['dtype'] = atype + att['value'] = const_var.ATTR_DEF_VAL.get(atype) + attrs.append(att) + op_node['bin_filename'] = '' + op_node['inputs'] = inputs + op_node['outputs'] = outputs + if len(attrs) > 0: + op_node['attrs'] = attrs + param = {} + param['op_type'] = self.op_type + param['op_list'] = [op_node] + objstr = json.dumps(param, indent=' ') + md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() + while key_map.get(md5sum) is not None: + objstr += '1' + md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() + key_map[md5sum] = md5sum + bin_file = self.op_type + '_' + md5sum + op_node['bin_filename'] = bin_file + param_file = os.path.join(self.out_path, bin_file + '_param.json') + param_file = os.path.realpath(param_file) + with os.fdopen(os.open(param_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + json.dump(param, fd, indent=' ') + self._write_buld_cmd(param_file, bin_file, i) + + + def _write_buld_cmd(self: any, param_file: str, bin_file: str, index: int): + hard_soc = const_var.SOC_MAP_EXT.get(self.soc) + if not hard_soc: + hard_soc = soc.capitalize() + name_com = [self.op_type, self.op_file, str(index)] + compile_file = os.path.join(self.out_path, '-'.join(name_com) + '.sh') + compile_file = os.path.realpath(compile_file) + with os.fdopen(os.open(compile_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + fd.write('#!/bin/bash\n') + fd.write('echo "[{}] Generating {} ..."\n'.format(hard_soc, bin_file)) + cmd = const_var.BIN_CMD.format(fun=self.op_intf, soc=hard_soc, param=param_file, impl='""') + fd.write(cmd) + chk = const_var.CHK_CMD.format(res_file=bin_file + '.json') + fd.write(chk) + chk = const_var.CHK_CMD.format(res_file=bin_file + '.o') + fd.write(chk) + fd.write('echo "[{}] Generating {} Done"\n'.format(hard_soc, bin_file)) + + +def gen_bin_param_file(cfgfile: str, out_dir: str, soc: str): + op_descs = opdesc_parser.get_op_desc(cfgfile, [], [], BinParamBuilder, None) + for op_desc in op_descs: + op_desc.set_soc_version(soc) + op_desc.set_out_path(out_dir) + op_desc.gen_input_json() + + +if __name__ == '__main__': + if len(sys.argv) <= 3: + raise RuntimeError('arguments must greater than 3') + gen_bin_param_file(sys.argv[1], sys.argv[2], sys.argv[3]) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py new file mode 100755 index 000000000..7fe177da1 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py @@ -0,0 +1,446 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import re +import stat +import opdesc_parser +import const_var + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + +IMPL_HEAD = ''' +import os, sys +import ctypes +import json +import shutil +from tbe.common.platform import get_soc_spec +from tbe.common.utils import para_check +from tbe.tikcpp import compile_op, replay_op, check_op_cap, generalize_op_params, get_code_channel, OpInfo +from tbe.common.buildcfg import get_default_build_config +from impl.util.platform_adapter import tbe_register +from tbe.common.buildcfg import get_current_build_config +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + +DTYPE_MAP = {"float32": ["DT_FLOAT", "float"], + "float16": ["DT_FLOAT16", "half"], + "int8": ["DT_INT8", "int8_t"], + "int16": ["DT_INT16", "int16_t"], + "int32": ["DT_INT32", "int32_t"], + "int64": ["DT_INT64", "int64_t"], + "uint1": ["DT_UINT1", "uint8_t"], + "uint8": ["DT_UINT8", "uint8_t"], + "uint16": ["DT_UINT16", "uint16_t"], + "uint32": ["DT_UINT32", "uint32_t"], + "uint64": ["DT_UINT64", "uint64_t"], + "bool": ["DT_BOOL", "bool"], + "double": ["DT_DOUBLE", "double"], + "dual": ["DT_DUAL", "unknown"], + "dual_sub_int8": ["DT_DUAL_SUB_INT8", "unknown"], + "dual_sub_uint8": ["DT_DUAL_SUB_UINT8", "unknown"], + "string": ["DT_STRING", "unknown"], + "complex64": ["DT_COMPLEX64", "unknown"], + "complex128": ["DT_COMPLEX128", "unknown"], + "qint8": ["DT_QINT8", "unknown"], + "qint16": ["DT_QINT16", "unknown"], + "qint32": ["DT_QINT32", "unknown"], + "quint8": ["DT_QUINT8", "unknown"], + "quint16": ["DT_QUINT16", "unknown"], + "resource": ["DT_RESOURCE", "unknown"], + "string_ref": ["DT_STRING_REF", "unknown"], + "int4": ["DT_INT4", "int8_t"], + "bfloat16": ["DT_BF16", "bfloat16_t"]} + +def get_dtype_fmt_options(__inputs__, __outputs__): + options = [] + for x in __inputs__ + __outputs__: + x_n = x.get("param_name").upper() + x_fmt = x.get("format") + x_dtype = x.get("dtype") + options.append("-DDTYPE_{n}={t}".format(n=x_n, t=DTYPE_MAP.get(x_dtype)[1])) + options.append("-DORIG_DTYPE_{n}={ot}".format(n=x_n, ot=DTYPE_MAP.get(x_dtype)[0])) + options.append("-DFORMAT_{n}=FORMAT_{f}".format(n=x_n, f=x_fmt)) + return options + +def load_dso(so_path): + try: + ctypes.CDLL(so_path) + except OSError as error : + print(error) + raise RuntimeError("cannot open %s" %(so_path)) + else: + print("load so succ ", so_path) + +def get_shortsoc_compile_option(compile_option_list: list, shortsoc:str): + compile_options = [] + if shortsoc in compile_option_list: + compile_options = compile_option_list[shortsoc] + elif '__ALLSOC__' in compile_option_list: + compile_options = compile_option_list['__ALLSOC__'] + return compile_options + +''' + +IMPL_API = ''' +@tbe_register.register_operator("{}") +@para_check.check_op_params({}) +def {}({}, kernel_name="{}", impl_mode=""): + if get_current_build_config("enable_op_prebuild"): + return + __inputs__, __outputs__, __attrs__ = _build_args({}) + options = get_dtype_fmt_options(__inputs__, __outputs__) + options += ["-x", "cce"] + ccec = os.environ.get('CCEC_REAL_PATH') + if ccec is None: + ccec = shutil.which("ccec") + if ccec != None: + ccec_path = os.path.dirname(ccec) + tikcpp_path = os.path.realpath(os.path.join(ccec_path, "..", "..", "tikcpp")) + else: + tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") + options.append("-I" + tikcpp_path) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) + options.append("-I" + os.path.join(PYF_PATH, "..", "ascendc", "common")) + if impl_mode == "high_performance": + options.append("-DHIGH_PERFORMANCE=1") + elif impl_mode == "high_precision": + options.append("-DHIGH_PRECISION=1") + if get_default_build_config("enable_deterministic_mode") == 1: + options.append("-DDETEMINISTIC_MODE=1") + + custom_compile_options = {}, + custom_all_compile_options = {}, + soc_version = get_soc_spec("SOC_VERSION") + soc_short = get_soc_spec("SHORT_SOC_VERSION").lower() + custom_compile_options_soc = get_shortsoc_compile_option(custom_compile_options[0], soc_short) + custom_all_compile_options_soc = get_shortsoc_compile_option(custom_all_compile_options[0], soc_short) + options += custom_all_compile_options_soc + options += custom_compile_options_soc + + origin_func_name = "{}" + ascendc_src_dir = "{}" + ascendc_src_file = "{}" + src = os.path.join(PYF_PATH, "..", "ascendc", ascendc_src_dir, ascendc_src_file) + if not os.path.exists(src): + src = os.path.join(PYF_PATH, ascendc_src_file) +''' + +REPLAY_OP_API = ''' + print("start replay Acend C Operator {}, kernel name is {}") + tikreplay_codegen_path = tikcpp_path + "/tikreplaylib/lib" + tikreplay_stub_path = tikcpp_path + "/tikreplaylib/lib/" + soc_version + print("start load libtikreplaylib_codegen.so and libtikreplaylib_stub.so") + codegen_so_path = tikreplay_codegen_path + "/libtikreplaylib_codegen.so" + replaystub_so_path = tikreplay_stub_path + "/libtikreplaylib_stub.so" + if PYF_PATH.endswith("dynamic"): + op_replay_path = os.path.join(PYF_PATH, "..", "..", "op_replay") + else: + op_replay_path = os.path.join(PYF_PATH, "..", "op_replay") + replayapi_so_path = os.path.join(op_replay_path, "libreplay_{}_" + soc_short + ".so") + load_dso(codegen_so_path) + load_dso(replaystub_so_path) + load_dso(replayapi_so_path) + op_type = "{}" + entry_obj = os.path.join(op_replay_path, "{}_entry_" + soc_short + ".o") + code_channel = get_code_channel(src, kernel_name, op_type, options) + op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ + attrs = __attrs__, impl_mode = impl_mode) + res, msg = replay_op(op_info, entry_obj, code_channel, src, options) + if not res: + print("call replay op failed for %s and get into call compile op" %(msg)) + compile_op(src, origin_func_name, op_info, options, code_channel, '{}') +''' + +COMPILE_OP_API = ''' + print("start compile Ascend C operator {}. kernel name is {}") + op_type = "{}" + code_channel = get_code_channel(src, kernel_name, op_type, options) + op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ + attrs = __attrs__, impl_mode = impl_mode, origin_inputs=[{}], origin_outputs = [{}]) + compile_op(src, origin_func_name, op_info, options, code_channel, '{}') +''' + +SUP_API = ''' +def {}({}, impl_mode=""): + __inputs__, __outputs__, __attrs__ = _build_args({}) + ret_str = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) + ret_dict = json.loads(ret_str) + err_code = ret_dict.get("ret_code") + sup = "Unknown" + reason = "Unknown reason" + if err_code is not None: + if err_code is 0: + sup = "True" + reason = "" + elif err_code is 1: + sup = "False" + reason = ret_dict.get("reason") + else: + sup = "Unknown" + reason = ret_dict.get("reason") + return sup, reason +''' +CAP_API = ''' +def {}({}, impl_mode=""): + __inputs__, __outputs__, __attrs__ = _build_args({}) + result = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) + return result.decode("utf-8") +''' +GLZ_API = ''' +@tbe_register.register_param_generalization("{}") +def {}_generalization({}, generalize_config=None): + __inputs__, __outputs__, __attrs__ = _build_args({}) + ret_str = generalize_op_params("{}", __inputs__, __outputs__, __attrs__, generalize_config) + return [json.loads(ret_str)] +''' + +ATTR_DEFAULT = {'bool': 'False', 'int': '0', 'float': '0.0', 'listInt': '[]', + 'listFloat': '[]', 'listBool': '[]', 'listListInt': '[[]]', 'str': ''} + + +def optype_snake(origin_str): + temp_str = origin_str[0].lower() + origin_str[1:] + new_str = re.sub(r'([A-Z])', r'_\1', temp_str).lower() + return new_str + + +class AdpBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + self.argsname = [] + self.argsdefv = [] + self.op_compile_option:str = '{}' + super().__init__(op_type) + + + def write_adapt(self: any, impl_path, path: str, op_compile_option_all:list = None): + self._build_paradefault() + if impl_path != "": + src_file = os.path.join(impl_path, self.op_file + '.cpp') + if not os.path.exists(src_file): + return + out_path = os.path.abspath(path) + if self.dynamic_shape and not out_path.endswith('dynamic'): + out_path = os.path.join(path, 'dynamic') + os.makedirs(out_path, mode=0o700, exist_ok=True) + adpfile = os.path.join(out_path, self.op_file + '.py') + self._gen_op_compile_option(op_compile_option_all) + with os.fdopen(os.open(adpfile, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + self._write_head(fd) + self._write_argparse(fd) + self._write_impl(fd) + if self.op_chk_support: + self._write_cap('check_supported', fd) + self._write_cap('get_op_support_info', fd) + if self.op_fmt_sel: + self._write_cap('op_select_format', fd) + self._write_cap('get_op_specific_info', fd) + if self.op_range_limit == 'limited' or self.op_range_limit == 'dynamic': + self._write_glz(fd) + + + def _gen_op_compile_option(self:any, op_compile_option_all:list =None): + if op_compile_option_all is not None: + if self.op_type in op_compile_option_all: + self.op_compile_option = op_compile_option_all[self.op_type] + elif "__all__" in op_compile_option_all: + self.op_compile_option = op_compile_option_all["__all__"] + + + def _ip_argpack(self: any, default: bool = True) -> list: + args = [] + for i in range(len(self.input_name)): + arg = self.input_name[i] + if default and self.argsdefv[i] is not None: + arg += '=' + self.argsdefv[i] + args.append(arg) + return args + + def _op_argpack(self: any, default: bool = True) -> list: + args = [] + argidx = len(self.input_name) + for i in range(len(self.output_name)): + arg = self.output_name[i] + if default and self.argsdefv[i + argidx] is not None: + arg += '=' + self.argsdefv[i + argidx] + args.append(arg) + return args + + def _attr_argpack(self: any, default: bool = True) -> list: + args = [] + argidx = len(self.input_name) + len(self.output_name) + for i in range(len(self.attr_list)): + att = self.attr_list[i] + arg = att + if default and self.argsdefv[i + argidx] is not None: + if self.attr_val.get(att).get('type') == 'str': + arg += '="' + self.argsdefv[i + argidx] + '"' + elif self.attr_val.get(att).get('type') == 'bool': + arg += '=' + self.argsdefv[i + argidx].capitalize() + else: + arg += '=' + self.argsdefv[i + argidx] + args.append(arg) + return args + + def _build_paralist(self: any, default: bool = True) -> str: + args = [] + args.extend(self._ip_argpack(default)) + args.extend(self._op_argpack(default)) + args.extend(self._attr_argpack(default)) + return ', '.join(args) + + def _io_parachk(self: any, types: list, type_name: str) -> list: + chk = [] + for iot in types: + if iot == 'optional': + ptype = 'OPTION' + else: + ptype = iot.upper() + chk.append('para_check.{}_{}'.format(ptype, type_name)) + return chk + + def _attr_parachk(self: any) -> list: + chk = [] + for att in self.attr_list: + if self.attr_val.get(att).get('paramType') == 'optional': + pt = 'OPTION' + else: + pt = self.attr_val.get(att).get('paramType').upper() + att_type = self.attr_val.get(att).get('type').upper() + att_type = att_type.replace('LIST', 'LIST_') + chk.append('para_check.{}_ATTR_{}'.format(pt, att_type)) + return chk + + def _build_parachk(self: any) -> str: + chk = [] + chk.extend(self._io_parachk(self.input_type, 'INPUT')) + chk.extend(self._io_parachk(self.output_type, 'OUTPUT')) + chk.extend(self._attr_parachk()) + chk.append('para_check.KERNEL_NAME') + return ', '.join(chk) + + def _build_paradefault(self: any): + optional = False + argtypes = [] + argtypes.extend(self.input_type) + argtypes.extend(self.output_type) + for atype in argtypes: + if atype == 'optional': + optional = True + if optional: + self.argsdefv.append('None') + else: + self.argsdefv.append(None) + for attr in self.attr_list: + atype = self.attr_val.get(attr).get('paramType') + if atype == 'optional': + optional = True + attrval = self.attr_val.get(attr).get('defaultValue') + if attrval is not None: + optional = True + if type == "bool": + attrval = attrval.capitalize() + elif type == "str": + attrval = "\"" + attrval + "\"" + self.argsdefv.append(attrval) + continue + if optional: + self.argsdefv.append(ATTR_DEFAULT.get(self.attr_val.get(attr).get('type'))) + else: + self.argsdefv.append(None) + + def _write_head(self: any, fd: object): + fd.write(IMPL_HEAD) + + def _write_argparse(self: any, fd: object): + args = self._build_paralist(False) + fd.write('def _build_args({}):\n'.format(args)) + fd.write(' __inputs__ = []\n') + fd.write(' for arg in [{}]:\n'.format(', '.join(self.input_name))) + fd.write(' if arg != None:\n') + fd.write(' if isinstance(arg, (list, tuple)):\n') + fd.write(' if len(arg) == 0:\n') + fd.write(' continue\n') + fd.write(' __inputs__.append(arg[0])\n') + fd.write(' else:\n') + fd.write(' __inputs__.append(arg)\n') + fd.write(' __outputs__ = []\n') + fd.write(' for arg in [{}]:\n'.format(', '.join(self.output_name))) + fd.write(' if arg != None:\n') + fd.write(' if isinstance(arg, (list, tuple)):\n') + fd.write(' if len(arg) == 0:\n') + fd.write(' continue\n') + fd.write(' __outputs__.append(arg[0])\n') + fd.write(' else:\n') + fd.write(' __outputs__.append(arg)\n') + fd.write(' __attrs__ = []\n') + for attr in self.attr_list: + fd.write(' if {} != None:\n'.format(attr)) + fd.write(' attr = {}\n') + fd.write(' attr["name"] = "{}"\n'.format(attr)) + fd.write(' attr["dtype"] = "{}"\n'.format(self.attr_val.get(attr).get('type'))) + fd.write(' attr["value"] = {}\n'.format(attr)) + fd.write(' __attrs__.append(attr)\n') + fd.write(' return __inputs__, __outputs__, __attrs__\n') + + def _write_impl(self: any, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + pchk = self._build_parachk() + if len(self.kern_name) > 0: + kern_name = self.kern_name + else: + kern_name = self.op_intf + src = self.op_file + '.cpp' + fd.write(IMPL_API.format(self.op_type, pchk, self.op_intf, argsdef, kern_name, argsval,\ + self.custom_compile_options, self.custom_all_compile_options, self.op_intf,\ + optype_snake(self.op_type), src)) + if self.op_replay_flag: + fd.write(REPLAY_OP_API.format(self.op_type, kern_name, self.op_file, self.op_type, self.op_file,\ + self.op_compile_option)) + else: + fd.write(COMPILE_OP_API.format(self.op_type, kern_name, self.op_type, ', '.join(self.input_name),\ + ', '.join(self.output_name), self.op_compile_option)) + + def _write_cap(self: any, cap_name: str, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + if cap_name == 'check_supported': + fd.write(SUP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) + else: + fd.write(CAP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) + + def _write_glz(self: any, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + fd.write(GLZ_API.format(self.op_type, self.op_intf, argsdef, argsval, self.op_type)) + + +def write_scripts(cfgfile: str, cfgs: dict, dirs: dict, ops: list = None, op_compile_option:list = None): + batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') + iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') + file_map = {} + op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, AdpBuilder,\ + ops, dirs.get(const_var.AUTO_GEN_DIR)) + for op_desc in op_descs: + op_desc.write_adapt(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), op_compile_option) + file_map[op_desc.op_type] = op_desc.op_file + return file_map + +if __name__ == '__main__': + if len(sys.argv) <= 6: + raise RuntimeError('arguments must greater equal than 6') + rep_cfg = {} + rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] + rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] + cfg_dir = {} + cfg_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] + cfg_dir[const_var.CFG_OUT_DIR] = sys.argv[5] + cfg_dir[const_var.AUTO_GEN_DIR] = sys.argv[6] + write_scripts(cfgfile=sys.argv[1], cfgs=rep_cfg, dirs=cfg_dir) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py new file mode 100755 index 000000000..7a97180be --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py @@ -0,0 +1,114 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import glob +import json +import argparse +import const_var + + +def load_json(json_file: str): + with open(json_file, encoding='utf-8') as file: + json_content = json.load(file) + return json_content + + +def get_specified_suffix_file(root_dir, suffix): + specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) + all_suffix_files = glob.glob(specified_suffix, recursive=True) + return all_suffix_files + + +def add_simplified_config(op_type, key, core_type, objfile, config): + simple_cfg = config.get('binary_info_config.json') + op_cfg = simple_cfg.get(op_type) + if not op_cfg: + op_cfg = {} + op_cfg['dynamicRankSupport'] = True + op_cfg['simplifiedKeyMode'] = 0 + op_cfg['binaryList'] = [] + simple_cfg[op_type] = op_cfg + bin_list = op_cfg.get('binaryList') + bin_list.append({'coreType': core_type, 'simplifiedKey': key, 'binPath': objfile}) + + +def add_op_config(op_file, bin_info, config): + op_cfg = config.get(op_file) + if not op_cfg: + op_cfg = {} + op_cfg['binList'] = [] + config[op_file] = op_cfg + op_cfg.get('binList').append(bin_info) + + +def gen_ops_config(json_file, soc, config): + core_type_map = {"MIX": 0, "AiCore": 1, "VectorCore": 2} + contents = load_json(json_file) + if ('binFileName' not in contents) or ('supportInfo' not in contents): + return + json_base_name = os.path.basename(json_file) + op_dir = os.path.basename(os.path.dirname(json_file)) + support_info = contents.get('supportInfo') + bin_name = contents.get('binFileName') + bin_suffix = contents.get('binFileSuffix') + core_type = core_type_map.get(contents.get("coreType")) + bin_file_name = bin_name + bin_suffix + op_type = bin_name.split('_')[0] + op_file = op_dir + '.json' + bin_info = {} + keys = support_info.get('simplifiedKey') + if keys: + bin_info['simplifiedKey'] = keys + for key in keys: + add_simplified_config(op_type, key, core_type, os.path.join(soc, op_dir, bin_file_name), config) + bin_info['staticKey'] = support_info.get('staticKey') + bin_info['int64Mode'] = support_info.get('int64Mode') + bin_info['inputs'] = support_info.get('inputs') + bin_info['outputs'] = support_info.get('outputs') + if support_info.get('attrs'): + bin_info['attrs'] = support_info.get('attrs') + bin_info['binInfo'] = {'jsonFilePath': os.path.join(soc, op_dir, json_base_name)} + add_op_config(op_file, bin_info, config) + + +def gen_all_config(root_dir, soc): + suffix = 'json' + config = {} + config['binary_info_config.json'] = {} + all_json_files = get_specified_suffix_file(root_dir, suffix) + for _json in all_json_files: + gen_ops_config(_json, soc, config) + for cfg_key in config.keys(): + cfg_file = os.path.join(root_dir, cfg_key) + with os.fdopen(os.open(cfg_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + json.dump(config.get(cfg_key), fd, indent=' ') + + +def args_prase(): + parser = argparse.ArgumentParser() + parser.add_argument('-p', + '--path', + nargs='?', + required=True, + help='Parse the path of the json file.') + parser.add_argument('-s', + '--soc', + nargs='?', + required=True, + help='Parse the soc_version of ops.') + return parser.parse_args() + + +def main(): + args = args_prase() + gen_all_config(args.path, args.soc) + + +if __name__ == '__main__': + main() diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py new file mode 100755 index 000000000..1cac7d911 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py @@ -0,0 +1,65 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import opdesc_parser +import replay_codegen +import const_var +from replay_codegen import ReplayCodeGenParams + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + + +class ReplayBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + super().__init__(op_type) + + def gen_replay_source(self: any, impl_path: str, out_path: str, ops_product: str): + if not self.op_replay_flag: + print('{} replay not enabled'.format(self.op_type)) + return + argn = len(self.input_name) + len(self.output_name) + 1 + if self.op_replay_batch: + print('{} replay in batch mode'.format(self.op_type)) + else: + print('{} replay in normal mode'.format(self.op_type)) + if impl_path.endswith('op_kernel'): + implf = os.path.join(impl_path, self.op_file + '.cpp') + tiling_file = os.path.join(impl_path, "../op_host", self.op_file + '_tiling.h') + else: + if self.dynamic_shape: + dyn_path = 'dynamic' + else: + dyn_path = '' + implf = os.path.join(impl_path, dyn_path, self.op_file + '.cpp') + tiling_file = os.path.join(impl_path, "../../op_tiling", self.op_file + '_tiling.h') + rep_conf = replay_codegen.ReplayCodeGen(ReplayCodeGenParams(self.op_type, implf, tiling_file, self.op_file, \ + self.op_intf, argn, self.op_replay_batch, self.max_block_dim, self.max_shape_size)) + rep_conf.set_batch(self.op_replay_batch) + rep_conf.set_outdir(out_path) + rep_conf.gen_replay(ops_product) + + +def gen_replay(cfgfile: str, cfgs: dict, dirs: dict, ops_product: str, ops: list = None): + batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') + iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') + op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, ReplayBuilder, ops) + for op_desc in op_descs: + op_desc.gen_replay_source(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), ops_product) + + +if __name__ == '__main__': + if len(sys.argv) <= 6: + raise RuntimeError('arguments must greater than 6') + rep_cfg = {} + rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] + rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] + rep_dir = {} + rep_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] + rep_dir[const_var.CFG_OUT_DIR] = sys.argv[5] + gen_replay(sys.argv[1], rep_cfg, rep_dir, sys.argv[6]) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp new file mode 100755 index 000000000..0e8834664 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp @@ -0,0 +1,117 @@ +#include +#include +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#include "register/op_check.h" +#define __ASCENDC_REPLAY_CODE__ +#include + +using namespace std; +using namespace optiling; +using namespace AscendCReplay; + +extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); +extern "C" int elf_batch_append(char *elf, uint32_t elfSize, char *jit, int kernum, char *atext[], int alen[], + int atlen, const char* kernelname[]); + +#define KERNEL_N 1 +#define ARG_N (__ARG_NUM__) +#define MAX_L (1024 * 1024 * 100) +#define MAX_E (1024 * 1024) + +int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) +{ + // gen type 1 : direct call codes 0: load .o file + if (param.gentype < 0 || param.gentype > 1) { + printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); + return 0; + } else if (param.gentype == 1 && param.objptr == nullptr) { + printf("Error: call replay with direct call mode, but code obj addr is null\n"); + return 0; + } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { + printf("Error: call replay with object file mode, but object file path is null\n"); + return 0; + } + // core_type 0:MIX 1:CUBE 2:VEC + if (core_type < 0 || core_type > 2) { + printf("Error: call replay core type is %d !\n", core_type); + return 0; + } + g_coreType = __CORE_TYPE__; + g_taskRation = param.task_ration; + g_tilingKey = param.tiling_key; + + unsigned char *buf, *jit; + char *kernel[KERNEL_N]; + int len[KERNEL_N]; + block_idx = 0; + block_num = param.block_dim; + g_ubBase = block_num; + uint8_t *code = (uint8_t *)malloc(MAX_L); + uint8_t *pos = code; + struct timespec tp1, tp2; + + clock_gettime(CLOCK_MONOTONIC, &tp1); + if (block_num > 32) { + printf("Error: block_num > 32\n"); + return 0; + } + //__OP_FOPEN__ + for (int i = 0; i < KERNEL_N; i++) { + //__OP_SET_KERNEL__ + for (int j = 0; j < ARG_N; j++) + AddArg(j, ARG_STEP * (j + 1)); +#ifdef FP_CEILING + SetCtrlFloatEnable(); +#else + SetCtrlFloatDisable(); +#endif + CodeInit(pos, true); + __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); + CodeEnd(); + kernel[i] = (char *)pos; + len[i] = CodeLen(); + pos += len[i]; + } + //__OP_FCLOSE__ + clock_gettime(CLOCK_MONOTONIC, &tp2); + buf = (unsigned char *)malloc(MAX_E); + int fd = open(param.entry_file, O_RDONLY); + if (fd < 0) { + printf("[error]: cannot find entry.o : %s\n", param.entry_file); + return 0; + } + uint32_t bufSize = read(fd, buf, MAX_E); + if (bufSize <= 0) { + printf("[error]: entry.o : %s is too small ! \n", param.entry_file); + } + close(fd); + jit = (unsigned char *)malloc(MAX_L); + printf("total code generated %ld\n", pos - code); + int sz = elf_batch_append((char *)buf, bufSize, (char *)jit, KERNEL_N, kernel, len, pos - code, ¶m.kernel_name); + if (tp1.tv_sec != tp2.tv_sec) { + printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); + } else { + printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); + } + printf("new elf size %d\n", sz); + if (param.gentype == 0) { + fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); + (void)write(fd, jit, sz); + close(fd); + free(jit); + } else if (param.gentype == 1) { + *param.objptr = (char*)jit; + } + free(buf); + free(code); + return sz; +} + +REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py new file mode 100755 index 000000000..137c9f39f --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py @@ -0,0 +1,115 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" +import os +import stat +import ctypes +import collections +import shutil +import subprocess +import copy + +"""CODE_* is used to cube/vector api is called in operator code +CODE_MIX means both cube and vector api is called +CODE_CUBE means only cube api is called +CODE_VEC means only vector api is called +""" +CODE_MIX = 0 +CODE_CUBE = 1 +CODE_VEC = 2 + + +def _is_v220(op_product: str): + """return if current soc version is V220 + + Returns: + res: True means V220 + """ + if op_product in ["ascend910b"]: + return True + return False + + +InfoCodeChanelParams = collections.namedtuple('InfoCodeChanelParams',\ +['src_file', 'tiling_header', 'kernel_name', 'outdir', 'op_product', 'compile_options']) + + +def infer_code_channel(params: InfoCodeChanelParams): + """get code channel for v220, return CODE_MIX if soc version is not V220 + + Args: + src_file (str): AscendC operator code file + src_file (str): AscendC operator tiling header file + kernel_name (str): kernel function name + optype (str): operator type + compile_options (list): compile options for ccec cmd + + Raises: + Exception: if not exist L1/L0/UB if code, it's not a aicore code + + Returns: + res (int): CODE_MIX/CODE_CUBE/CODE_VEC + """ + if not _is_v220(params.op_product): + return CODE_MIX + return CODE_VEC + if params.compile_options is None: + compile_options = [] + else: + compile_options = params.compile_options + ccec = shutil.which("ccec") + if ccec is not None: + ccec_path = os.path.dirname(ccec) + tikcpp_path = os.path.realpath(os.path.join(ccec_path, "..", "..", "tikcpp")) + else: + tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") + compile_options.append("-I" + tikcpp_path) + compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) + compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) + compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) + compile_options += ["-include", params.tiling_header] + arch = "dav-c220-cube" + sub_core_type = "AIC" + optional_core = "AiCore" + compile_cmd = [shutil.which("ccec"), '-c', '-O3'] + compile_cmd += compile_options + temp_file_name_tag = "_" + str(os.getpid()) + "_temp.o" + dst_file = os.path.join(kernel_meta_dir, kernel_name + temp_file_name_tag) + compile_cmd += [params.src_file, "--cce-aicore-arch={}".format(arch), + "--cce-aicore-only", "-o", dst_file, + "-mllvm", "-cce-aicore-function-stack-size=16000", + "-mllvm", "-cce-aicore-record-overflow=true", + "-mllvm", "-cce-aicore-addr-transform"] + compile_cmd += ["-std=c++17"] + print('get_code_channel: ', ' '.join(compile_cmd)) + proc = subprocess.Popen( + compile_cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) + (out, _) = proc.communicate() + if proc.returncode != 0: + print('get_code_channel coretype compile error: ', out.decode()) + msg = "compile %s error :%s\n" % (params.src_file, out.decode()) + raise Exception(f"get_code_channel coretype error, msg is{msg}") + objdump_cmd = ['objdump', '-s', '-j', '.text', '{}'.format(dst_file)] + + proc = subprocess.Popen( + objdump_cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) + (out, _) = proc.communicate() + if proc.returncode != 0: + print('get_code_channel objdump error: ', out.decode()) + msg = "get_code_channel objdump %s error :%s\n" % (src_file, out.decode()) + raise Exception(f"get_code_channel objdump error, msg is{msg}") + os.remove(dst_file) + lines = out.decode('utf-8').split('\n') + for line in lines: + insts = line.strip().split() + if len(insts) < 5: + continue + for inst in insts[1:]: + if len(inst) != 8: + continue + if inst[6] == 'f': + return CODE_MIX + return CODE_VEC diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/const_var.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/const_var.py new file mode 100755 index 000000000..8b32c3b91 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/const_var.py @@ -0,0 +1,33 @@ + +#!/usr/bin/env python +# coding=utf-8 +""" +Function: +The replay funtion entry +Copyright Information: +Huawei Technologies Co., Ltd. All Rights Reserved © 2020 +""" + +import os +import stat + + +REPLAY_BATCH = 'batch' +REPLAY_ITERATE = 'iterate' +CFG_IMPL_DIR = 'impl_dir' +CFG_OUT_DIR = 'out_dir' +AUTO_GEN_DIR = 'auto_gen_dir' +WFLAGS = os.O_WRONLY | os.O_CREAT | os.O_TRUNC +WMODES = stat.S_IWUSR | stat.S_IRUSR +SOC_MAP_EXT = {'ascend310p': 'Ascend310P3', 'ascend310b': 'Ascend310B1', + 'ascend910': 'Ascend910A', 'ascend910b': 'Ascend910B1'} +BIN_CMD = 'opc $1 --main_func={fun} --input_param={param} --soc_version={soc} \ +--output=$2 --impl_mode={impl} --simplified_key_mode=0 --op_mode=dynamic\n' +CHK_CMD = ''' +if ! test -f $2/{res_file} ; then + echo "$2/{res_file} not generated!" + exit 1 +fi +''' +ATTR_DEF_VAL = {'str' : '', 'int': 0, 'float': 0.0, 'bool': False, 'list_bool': [], + 'list_int': [], 'list_float': [], 'list_list_int': [[]]} diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh new file mode 100755 index 000000000..55e12e5ed --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh @@ -0,0 +1,57 @@ +#!/usr/bin/bash + +project_path=$1 +build_path=$2 +vendor_name=customize +if [[ ! -d "$project_path" ]]; then + echo "[ERROR] No projcet path is provided" + exit 1 +fi + +if [[ ! -d "$build_path" ]]; then + echo "[ERROR] No build path is provided" + exit 1 +fi + +# copy ai_core operators implements +# tbe_impl_files_num=$(ls $project_path/tbe/impl/* 2> /dev/null | wc -l) +# if [[ "$tbe_impl_files_num" -gt 0 ]];then +# cp -r ${project_path}/tbe/impl/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/ai_core/tbe/customize_impl +# cp -r ${project_path}/tbe/impl/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/vector_core/tbe/customize_impl +# fi + +# copy aicpu kernel so operators +if [[ -d "${project_path}/cpukernel/aicpu_kernel_lib" ]]; then + cp -f ${project_path}/cpukernel/aicpu_kernel_lib/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/cpu/aicpu_kernel/impl + rm -rf ${project_path}/cpukernel/aicpu_kernel_lib +fi + +# merge aicpu.ini and aicore.ini to generate npu_supported_ops.json +# mkdir -p ${build_path}/framework/op_info_cfg +# mkdir -p ${build_path}/framework/op_info_cfg/aicpu_kernel +# mkdir -p ${build_path}/framework/op_info_cfg/ai_core + +# if [[ -d "${project_path}/tbe/op_info_cfg/ai_core" ]]; then +# bash ${project_path}/cmake/util/gen_ops_filter.sh ${project_path}/tbe/op_info_cfg/ai_core ${build_path}/framework/op_info_cfg/ai_core +# fi + +# if [[ -d "${project_path}/cpukernel/op_info_cfg/aicpu_kernel" ]]; then +# bash ${project_path}/cmake/util/gen_ops_filter.sh ${project_path}/cpukernel/op_info_cfg/aicpu_kernel ${build_path}/framework/op_info_cfg/aicpu_kernel +# fi + +# aicpu_filter_file=${build_path}/framework/op_info_cfg/aicpu_kernel/npu_supported_ops.json +# aicore_filter_file=${build_path}/framework/op_info_cfg/ai_core/npu_supported_ops.json +# if [[ -f "${aicpu_filter_file}" ]] && [[ ! -f "${aicore_filter_file}" ]]; then +# cp $aicpu_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow +# fi +# if [[ -f "${aicore_filter_file}" ]] && [[ ! -f "${aicpu_filter_file}" ]]; then +# cp $aicore_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow +# fi + +# if [[ -f "${aicore_filter_file}" ]] && [[ -f "${aicpu_filter_file}" ]]; then +# chmod u+w ${aicpu_filter_file} +# python3 ${project_path}/cmake/util/insert_op_info.py ${aicore_filter_file} ${aicpu_filter_file} +# chmod u-w ${aicpu_filter_file} +# cp $aicpu_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow +# fi + diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh new file mode 100755 index 000000000..d4c27d17f --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh @@ -0,0 +1,62 @@ +#!/bin/bash +# Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +# Description: Generate npu_supported_ops.json +# ============================================================================== + +if [[ -z "$1" ]]; then + echo -e "[ERROR] No source dir provided" + exit 1 +fi + +if [[ -z "$2" ]]; then + echo -e "[ERROR] No destination dir provided" + exit 1 +fi + +src=$1 +dest_file=$2/npu_supported_ops.json + +if [ -f "$dest_file" ];then + chmod u+w $dest_file +fi + +echo $* + +add_ops() { + name=$1 + isHeavy=$2 + file=$3 + grep -w "\"$name\"" ${file} >/dev/null + if [ $? == 0 ];then + return + fi + echo " \"${name}\": {" >> ${file} + echo " \"isGray\": false," >> ${file} + echo " \"isHeavy\": ${isHeavy}" >> ${file} + echo " }," >> ${file} +} + +echo "{" > ${dest_file} +ini_files=$(find ${src} -name "*.ini") +for file in ${ini_files} ; do + name=$(grep '^\[' ${file} | sed 's/\[//g' | sed 's/]//g' | sed 's/\r//g') + grep 'heavyOp.flag' ${file} >/dev/null + if [ $? == 0 ];then + isHeavy=$(grep 'heavyOp.flag' ${file} | awk -F= '{print $2}') + else + isHeavy="false" + fi + for op in ${name} ; do + add_ops ${op} "false" ${dest_file} + done +done +echo "}" >> ${dest_file} +file_count=$(cat ${dest_file} | wc -l) +line=$(($file_count-1)) +sed -i "${line}{s/,//g}" ${dest_file} + +chmod 640 "${dest_file}" +echo -e "[INFO] Succed generated ${dest_file}" + +exit 0 + diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh new file mode 100755 index 000000000..a06cfc78d --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh @@ -0,0 +1,6 @@ +ascend_install_dir=$1 +gen_file_dir=$2 + +# create version.info +compiler_version=$(grep "Version" -w ${ascend_install_dir}/compiler/version.info | awk -F = '{print $2}') +echo "custom_opp_compiler_version=${compiler_version}" > ${gen_file_dir}/version.info \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py new file mode 100755 index 000000000..28ba08757 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py @@ -0,0 +1,36 @@ +# -*- coding: utf-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" +import json +import os +import sys +import stat +import const_var + + +if __name__ == '__main__': + if len(sys.argv) != 3: + print(sys.argv) + print('argv error, inert_op_info.py your_op_file lib_op_file') + sys.exit(2) + + with open(sys.argv[1], 'r') as load_f: + insert_operator = json.load(load_f) + + all_operators = {} + if os.path.exists(sys.argv[2]): + if os.path.getsize(sys.argv[2]) != 0: + with open(sys.argv[2], 'r') as load_f: + all_operators = json.load(load_f) + + for k in insert_operator.keys(): + if k in all_operators.keys(): + print('replace op:[', k, '] success') + else: + print('insert op:[', k, '] success') + all_operators[k] = insert_operator[k] + + with os.fdopen(os.open(sys.argv[2], const_var.WFLAGS, const_var.WMODES), 'w') as json_file: + json_file.write(json.dumps(all_operators, indent=4)) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py new file mode 100755 index 000000000..ace727b90 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py @@ -0,0 +1,248 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import re +import glob +import json +import argparse +import const_var + + +DATA_TPYE_DICT = { + 'float32': 0, + 'float16': 1, + 'int8': 2, + 'int16': 6, + 'uint16': 7, + 'uint8': 4, + 'int32': 3, + 'int64': 9, + 'uint32': 8, + 'uint64': 10, + 'bool': 12, + 'double': 11, + 'string': 13, + 'dual': 14, + 'dual': 15, + 'complex64': 16, + 'complex128': 17, + 'qint8': 18, + 'qint16': 19, + 'qint32': 20, + 'quint8': 21, + 'quint16': 22, + 'resource': 23, + 'string': 24, + 'dual': 25, + 'variant': 26, + 'bf16': 27, + 'bfloat16': 27, + 'undefined': 28, + 'int4': 29, + 'uint1': 30, + 'int2': 31 +} + +FORMAT_DICT = { + 'NCHW': 0, + 'NHWC': 1, + 'ND': 2, + 'NC1HWC0': 3, + 'FRACTAL_Z': 4, + 'NC1C0HWPAD': 5, + 'NHWC1C0': 6, + 'FSR_NCHW': 7, + 'FRACTAL_DECONV': 8, + 'C1HWNC0': 9, + 'FRACTAL_DECONV_TRANSPOSE': 10, + 'FRACTAL_DECONV_SP_STRIDE_TRANS': 11, + 'NC1HWC0_C04': 12, + 'FRACTAL_Z_C04': 13, + 'CHWN': 14, + 'FRACTAL_DECONV_SP_STRIDE8_TRANS': 15, + 'HWCN': 16, + 'NC1KHKWHWC0': 17, + 'BN_WEIGHT': 18, + 'FILTER_HWCK': 19, + 'HASHTABLE_LOOKUP_LOOKUPS': 20, + 'HASHTABLE_LOOKUP_KEYS': 21, + 'HASHTABLE_LOOKUP_VALUE': 22, + 'HASHTABLE_LOOKUP_OUTPUT': 23, + 'HASHTABLE_LOOKUP_HITS': 24, + 'C1HWNCoC0': 25, + 'MD': 26, + 'NDHWC': 27, + 'FRACTAL_ZZ': 28, + 'FRACTAL_NZ': 29, + 'NCDHW': 30, + 'DHWCN': 31, + 'NDC1HWC0': 32, + 'FRACTAL_Z_3D': 33, + 'CN': 34, + 'NC': 35, + 'DHWNC': 36, + 'FRACTAL_Z_3D_TRANSPOSE': 37, + 'FRACTAL_ZN_LSTM': 38, + 'FRACTAL_Z_G': 39, + 'RESERVED': 40, + 'ALL': 41, + 'NULL': 42, + 'ND_RNN_BIAS': 43, + 'FRACTAL_ZN_RNN': 44, + 'NYUV': 45, + 'NYUV_A': 46 +} + + +def load_json(json_file: str): + with open(json_file, encoding='utf-8') as file: + json_content = json.load(file) + return json_content + + +def get_specified_suffix_file(root_dir, suffix): + specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) + all_suffix_files = glob.glob(specified_suffix, recursive=True) + return all_suffix_files + + +def get_deterministic_value(support_info): + deterministic_key = 'deterministic' + if deterministic_key not in support_info: + return 0 + deterministic_value = support_info.get(deterministic_key) + if deterministic_value == 'true': + return 1 + else: + return 0 + + +def get_precision_value(support_info): + precision_key = 'implMode' + precision_value = support_info.get(precision_key) + if precision_value == 'high_performance': + _value = 1 + elif precision_value == 'high_precision': + _value = 2 + else: + _value = 0 + return _value + + +def get_overflow_value(support_info): + return 0 + + +def get_parameters(info): + if info: + if 'dtype' in info: + data_type = info['dtype'] + data_type_value = DATA_TPYE_DICT.get(data_type) + else: + data_type_value = 0 + if 'format' in info: + _format = info['format'] + _format_value = FORMAT_DICT.get(_format) + else: + _format_value = 0 + else: + data_type_value = 0 + _format_value = 0 + return str(data_type_value), str(_format_value) + + +def get_dynamic_parameters(info): + # 动态输入时只需获取第一个参数 + return get_parameters(info[0]) + + +def get_all_parameters(support_info, _type): + result_list = list() + info_lists = support_info.get(_type) + if info_lists: + for _info in info_lists: + # 输入为列表时是动态输入 + if isinstance(_info, (list, tuple)): + data_type_value, _format_value = get_dynamic_parameters(_info) + else: + data_type_value, _format_value = get_parameters(_info) + result_list.append("{},{}".format(data_type_value, _format_value)) + return result_list + + +def get_all_input_parameters(support_info): + result = get_all_parameters(support_info, 'inputs') + return '/'.join(result) + + +def insert_content_into_file(input_file, content): + with open(input_file, 'r+') as file: + lines = file.readlines() + for index, line in enumerate(lines): + match_result = re.search(r'"staticKey":', line) + if match_result: + count = len(line) - len(line.lstrip()) + new_content = "{}{}".format(' ' * count, content) + # 插入到前一行,防止插入最后时还需要考虑是否添加逗号 + lines.insert(index, new_content) + break + file.seek(0) + file.write(''.join(lines)) + + +def insert_simplified_keys(json_file): + contents = load_json(json_file) + # 不存在'binFileName'或者'supportInfo'字段时,非需要替换的解析json文件 + if ('binFileName' not in contents) or ('supportInfo' not in contents): + return + support_info = contents.get('supportInfo') + bin_file_name = contents.get('binFileName') + bin_suffix = contents.get('binFileSuffix') + # 'simplifiedKey'字段已经存在时,直接返回,不重复生成 + if 'simplifiedKey' in support_info: + return + op_type = bin_file_name.split('_')[0] + deterministic = str(get_deterministic_value(support_info)) + precision = str(get_precision_value(support_info)) + overflow = str(get_overflow_value(support_info)) + input_parameters = get_all_input_parameters(support_info) + key = '{}/d={},p={},o={}/{}/'.format( + op_type, + deterministic, + precision, + overflow, + input_parameters) + result = '"simplifiedKey": "' + key + '",\n' + insert_content_into_file(json_file, result) + + +def insert_all_simplified_keys(root_dir): + suffix = 'json' + all_json_files = get_specified_suffix_file(root_dir, suffix) + for _json in all_json_files: + insert_simplified_keys(_json) + + +def args_prase(): + parser = argparse.ArgumentParser() + parser.add_argument('-p', + '--path', + nargs='?', + required=True, + help='Parse the path of the json file.') + return parser.parse_args() + + +def main(): + args = args_prase() + insert_all_simplified_keys(args.path) + + +if __name__ == '__main__': + main() diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py new file mode 100755 index 000000000..2b77c970d --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py @@ -0,0 +1,115 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + + +def gen_fun_def(title, kernel, argn, arg_type, arg_name): + entry = [] + entry.append(title) + entry.append(kernel) + entry.append('(') + args = [] + for i in range(0, argn): + args.append(arg_type + ' ' + arg_name + str(i)) + entry.append(', '.join(args)) + entry.append(')') + return ' '.join(entry) + + +def gen_batch_kernel_body(fname, argn, arg_name): + body = [] + body.append('{') + fun = [] + fun.append(fname) + fun.append('(') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + fun.append(', '.join(args)) + fun.append(');') + body.append(' '.join(fun)) + body.append('}') + return '\n'.join(body) + + +def gen_mc_kernel_body(kn, argn, arg_name, blknum): + body = [] + body.append('{') + body.append(' switch(block_idx) {') + for blk in range(0, blknum): + fun = [] + fun.append('{}_blk{:02d}'.format(kn, blk)) + fun.append('(') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + fun.append(', '.join(args)) + fun.append(')') + body.append(' case {}: {}; break;'.format(blk, ' '.join(fun))) + body.append(' default: break;') + body.append(' }') + body.append('}') + return '\n'.join(body) + + +def gen_proc_body(argn, arg_name): + body = [] + body.append('{') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + body.append('uint64_t __x = (uint64_t)' + ' + (uint64_t)'.join(args) + ';') + body.append('__asm__ ("NOP");') + body.append('__asm__ ("NOP");') + body.append('__asm__ ("NOP");') + body.append('}') + return '\n'.join(body) + + +def batch_code_gen(kn, argn, argt): + codes = [] + kernel_name = kn + proc_name = kernel_name + '_percore' + arg_num = int(argn) + data_type = argt + arg_type = '__gm__ ' + data_type + '* __restrict__' + arg_name = 'arg' + kernel_title = 'extern \"C\" __global__ __aicore__ void' + proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' + codes.append('#ifndef __aicore__') + codes.append('#define __aicore__ [aicore]') + codes.append('#endif') + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') + codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) + codes.append(gen_batch_kernel_body(proc_name, arg_num, arg_name)) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) + codes.append(gen_proc_body(arg_num, arg_name)) + return '\n'.join(codes) + '\n' + + +def mc_code_gen(kn, argn, argt, blknum): + codes = [] + kernel_name = kn + core_num = int(blknum) + arg_num = int(argn) + data_type = argt + arg_type = '__gm__ ' + data_type + '* __restrict__' + arg_name = 'arg' + kernel_title = 'extern \"C\" __global__ __aicore__ void' + proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' + codes.append('#ifndef __aicore__') + codes.append('#define __aicore__ [aicore]') + codes.append('#endif') + for i in range(0, core_num): + proc_name = '{}_blk{:02d}'.format(kernel_name, i) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') + codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) + codes.append(gen_mc_kernel_body(kernel_name, arg_num, arg_name, core_num)) + for i in range(0, core_num): + proc_name = '{}_blk{:02d}'.format(kernel_name, i) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) + codes.append(gen_proc_body(arg_num, arg_name)) + return '\n'.join(codes) + '\n' diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp new file mode 100755 index 000000000..5079a1043 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp @@ -0,0 +1,10 @@ +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#define __ASCENDC_REPLAY_CODE__ +#include "__CCE_FILE__" diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING new file mode 100644 index 000000000..d159169d1 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING @@ -0,0 +1,339 @@ + GNU GENERAL PUBLIC LICENSE + Version 2, June 1991 + + Copyright (C) 1989, 1991 Free Software Foundation, Inc., + 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + Everyone is permitted to copy and distribute verbatim copies + of this license document, but changing it is not allowed. + + Preamble + + The licenses for most software are designed to take away your +freedom to share and change it. By contrast, the GNU General Public +License is intended to guarantee your freedom to share and change free +software--to make sure the software is free for all its users. This +General Public License applies to most of the Free Software +Foundation's software and to any other program whose authors commit to +using it. (Some other Free Software Foundation software is covered by +the GNU Lesser General Public License instead.) You can apply it to +your programs, too. + + When we speak of free software, we are referring to freedom, not +price. Our General Public Licenses are designed to make sure that you +have the freedom to distribute copies of free software (and charge for +this service if you wish), that you receive source code or can get it +if you want it, that you can change the software or use pieces of it +in new free programs; and that you know you can do these things. + + To protect your rights, we need to make restrictions that forbid +anyone to deny you these rights or to ask you to surrender the rights. +These restrictions translate to certain responsibilities for you if you +distribute copies of the software, or if you modify it. + + For example, if you distribute copies of such a program, whether +gratis or for a fee, you must give the recipients all the rights that +you have. You must make sure that they, too, receive or can get the +source code. And you must show them these terms so they know their +rights. + + We protect your rights with two steps: (1) copyright the software, and +(2) offer you this license which gives you legal permission to copy, +distribute and/or modify the software. + + Also, for each author's protection and ours, we want to make certain +that everyone understands that there is no warranty for this free +software. If the software is modified by someone else and passed on, we +want its recipients to know that what they have is not the original, so +that any problems introduced by others will not reflect on the original +authors' reputations. + + Finally, any free program is threatened constantly by software +patents. We wish to avoid the danger that redistributors of a free +program will individually obtain patent licenses, in effect making the +program proprietary. To prevent this, we have made it clear that any +patent must be licensed for everyone's free use or not licensed at all. + + The precise terms and conditions for copying, distribution and +modification follow. + + GNU GENERAL PUBLIC LICENSE + TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION + + 0. This License applies to any program or other work which contains +a notice placed by the copyright holder saying it may be distributed +under the terms of this General Public License. The "Program", below, +refers to any such program or work, and a "work based on the Program" +means either the Program or any derivative work under copyright law: +that is to say, a work containing the Program or a portion of it, +either verbatim or with modifications and/or translated into another +language. (Hereinafter, translation is included without limitation in +the term "modification".) Each licensee is addressed as "you". + +Activities other than copying, distribution and modification are not +covered by this License; they are outside its scope. The act of +running the Program is not restricted, and the output from the Program +is covered only if its contents constitute a work based on the +Program (independent of having been made by running the Program). +Whether that is true depends on what the Program does. + + 1. You may copy and distribute verbatim copies of the Program's +source code as you receive it, in any medium, provided that you +conspicuously and appropriately publish on each copy an appropriate +copyright notice and disclaimer of warranty; keep intact all the +notices that refer to this License and to the absence of any warranty; +and give any other recipients of the Program a copy of this License +along with the Program. + +You may charge a fee for the physical act of transferring a copy, and +you may at your option offer warranty protection in exchange for a fee. + + 2. You may modify your copy or copies of the Program or any portion +of it, thus forming a work based on the Program, and copy and +distribute such modifications or work under the terms of Section 1 +above, provided that you also meet all of these conditions: + + a) You must cause the modified files to carry prominent notices + stating that you changed the files and the date of any change. + + b) You must cause any work that you distribute or publish, that in + whole or in part contains or is derived from the Program or any + part thereof, to be licensed as a whole at no charge to all third + parties under the terms of this License. + + c) If the modified program normally reads commands interactively + when run, you must cause it, when started running for such + interactive use in the most ordinary way, to print or display an + announcement including an appropriate copyright notice and a + notice that there is no warranty (or else, saying that you provide + a warranty) and that users may redistribute the program under + these conditions, and telling the user how to view a copy of this + License. (Exception: if the Program itself is interactive but + does not normally print such an announcement, your work based on + the Program is not required to print an announcement.) + +These requirements apply to the modified work as a whole. If +identifiable sections of that work are not derived from the Program, +and can be reasonably considered independent and separate works in +themselves, then this License, and its terms, do not apply to those +sections when you distribute them as separate works. But when you +distribute the same sections as part of a whole which is a work based +on the Program, the distribution of the whole must be on the terms of +this License, whose permissions for other licensees extend to the +entire whole, and thus to each and every part regardless of who wrote it. + +Thus, it is not the intent of this section to claim rights or contest +your rights to work written entirely by you; rather, the intent is to +exercise the right to control the distribution of derivative or +collective works based on the Program. + +In addition, mere aggregation of another work not based on the Program +with the Program (or with a work based on the Program) on a volume of +a storage or distribution medium does not bring the other work under +the scope of this License. + + 3. You may copy and distribute the Program (or a work based on it, +under Section 2) in object code or executable form under the terms of +Sections 1 and 2 above provided that you also do one of the following: + + a) Accompany it with the complete corresponding machine-readable + source code, which must be distributed under the terms of Sections + 1 and 2 above on a medium customarily used for software interchange; or, + + b) Accompany it with a written offer, valid for at least three + years, to give any third party, for a charge no more than your + cost of physically performing source distribution, a complete + machine-readable copy of the corresponding source code, to be + distributed under the terms of Sections 1 and 2 above on a medium + customarily used for software interchange; or, + + c) Accompany it with the information you received as to the offer + to distribute corresponding source code. (This alternative is + allowed only for noncommercial distribution and only if you + received the program in object code or executable form with such + an offer, in accord with Subsection b above.) + +The source code for a work means the preferred form of the work for +making modifications to it. For an executable work, complete source +code means all the source code for all modules it contains, plus any +associated interface definition files, plus the scripts used to +control compilation and installation of the executable. However, as a +special exception, the source code distributed need not include +anything that is normally distributed (in either source or binary +form) with the major components (compiler, kernel, and so on) of the +operating system on which the executable runs, unless that component +itself accompanies the executable. + +If distribution of executable or object code is made by offering +access to copy from a designated place, then offering equivalent +access to copy the source code from the same place counts as +distribution of the source code, even though third parties are not +compelled to copy the source along with the object code. + + 4. You may not copy, modify, sublicense, or distribute the Program +except as expressly provided under this License. Any attempt +otherwise to copy, modify, sublicense or distribute the Program is +void, and will automatically terminate your rights under this License. +However, parties who have received copies, or rights, from you under +this License will not have their licenses terminated so long as such +parties remain in full compliance. + + 5. You are not required to accept this License, since you have not +signed it. However, nothing else grants you permission to modify or +distribute the Program or its derivative works. These actions are +prohibited by law if you do not accept this License. Therefore, by +modifying or distributing the Program (or any work based on the +Program), you indicate your acceptance of this License to do so, and +all its terms and conditions for copying, distributing or modifying +the Program or works based on it. + + 6. Each time you redistribute the Program (or any work based on the +Program), the recipient automatically receives a license from the +original licensor to copy, distribute or modify the Program subject to +these terms and conditions. You may not impose any further +restrictions on the recipients' exercise of the rights granted herein. +You are not responsible for enforcing compliance by third parties to +this License. + + 7. If, as a consequence of a court judgment or allegation of patent +infringement or for any other reason (not limited to patent issues), +conditions are imposed on you (whether by court order, agreement or +otherwise) that contradict the conditions of this License, they do not +excuse you from the conditions of this License. If you cannot +distribute so as to satisfy simultaneously your obligations under this +License and any other pertinent obligations, then as a consequence you +may not distribute the Program at all. For example, if a patent +license would not permit royalty-free redistribution of the Program by +all those who receive copies directly or indirectly through you, then +the only way you could satisfy both it and this License would be to +refrain entirely from distribution of the Program. + +If any portion of this section is held invalid or unenforceable under +any particular circumstance, the balance of the section is intended to +apply and the section as a whole is intended to apply in other +circumstances. + +It is not the purpose of this section to induce you to infringe any +patents or other property right claims or to contest validity of any +such claims; this section has the sole purpose of protecting the +integrity of the free software distribution system, which is +implemented by public license practices. Many people have made +generous contributions to the wide range of software distributed +through that system in reliance on consistent application of that +system; it is up to the author/donor to decide if he or she is willing +to distribute software through any other system and a licensee cannot +impose that choice. + +This section is intended to make thoroughly clear what is believed to +be a consequence of the rest of this License. + + 8. If the distribution and/or use of the Program is restricted in +certain countries either by patents or by copyrighted interfaces, the +original copyright holder who places the Program under this License +may add an explicit geographical distribution limitation excluding +those countries, so that distribution is permitted only in or among +countries not thus excluded. In such case, this License incorporates +the limitation as if written in the body of this License. + + 9. The Free Software Foundation may publish revised and/or new versions +of the General Public License from time to time. Such new versions will +be similar in spirit to the present version, but may differ in detail to +address new problems or concerns. + +Each version is given a distinguishing version number. If the Program +specifies a version number of this License which applies to it and "any +later version", you have the option of following the terms and conditions +either of that version or of any later version published by the Free +Software Foundation. If the Program does not specify a version number of +this License, you may choose any version ever published by the Free Software +Foundation. + + 10. If you wish to incorporate parts of the Program into other free +programs whose distribution conditions are different, write to the author +to ask for permission. For software which is copyrighted by the Free +Software Foundation, write to the Free Software Foundation; we sometimes +make exceptions for this. Our decision will be guided by the two goals +of preserving the free status of all derivatives of our free software and +of promoting the sharing and reuse of software generally. + + NO WARRANTY + + 11. BECAUSE THE PROGRAM IS LICENSED FREE OF CHARGE, THERE IS NO WARRANTY +FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN +OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES +PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED +OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF +MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS +TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE +PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, +REPAIR OR CORRECTION. + + 12. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING +WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MAY MODIFY AND/OR +REDISTRIBUTE THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, +INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING +OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED +TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY +YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER +PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE +POSSIBILITY OF SUCH DAMAGES. + + END OF TERMS AND CONDITIONS + + How to Apply These Terms to Your New Programs + + If you develop a new program, and you want it to be of the greatest +possible use to the public, the best way to achieve this is to make it +free software which everyone can redistribute and change under these terms. + + To do so, attach the following notices to the program. It is safest +to attach them to the start of each source file to most effectively +convey the exclusion of warranty; and each file should have at least +the "copyright" line and a pointer to where the full notice is found. + + + Copyright (C) + + This program is free software; you can redistribute it and/or modify + it under the terms of the GNU General Public License as published by + the Free Software Foundation; either version 2 of the License, or + (at your option) any later version. + + This program is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + GNU General Public License for more details. + + You should have received a copy of the GNU General Public License along + with this program; if not, write to the Free Software Foundation, Inc., + 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. + +Also add information on how to contact you by electronic and paper mail. + +If the program is interactive, make it output a short notice like this +when it starts in an interactive mode: + + Gnomovision version 69, Copyright (C) year name of author + Gnomovision comes with ABSOLUTELY NO WARRANTY; for details type `show w'. + This is free software, and you are welcome to redistribute it + under certain conditions; type `show c' for details. + +The hypothetical commands `show w' and `show c' should show the appropriate +parts of the General Public License. Of course, the commands you use may +be called something other than `show w' and `show c'; they could even be +mouse-clicks or menu items--whatever suits your program. + +You should also get your employer (if you work as a programmer) or your +school, if any, to sign a "copyright disclaimer" for the program, if +necessary. Here is a sample; alter the names: + + Yoyodyne, Inc., hereby disclaims all copyright interest in the program + `Gnomovision' (which makes passes at compilers) written by James Hacker. + + , 1 April 1989 + Ty Coon, President of Vice + +This General Public License does not permit incorporating your program into +proprietary programs. If your program is a subroutine library, you may +consider it more useful to permit linking proprietary applications with the +library. If this is what you want to do, use the GNU Lesser General +Public License instead of this License. diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/README.md b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/README.md new file mode 100644 index 000000000..b41f01682 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/README.md @@ -0,0 +1,246 @@ +[![License: GPL v2](https://img.shields.io/badge/License-GPL%20v2-blue.svg)](https://www.gnu.org/licenses/old-licenses/gpl-2.0.en.html) +![Build Status](https://github.com/megastep/makeself/workflows/CI/badge.svg) + +# makeself - Make self-extractable archives on Unix + +[makeself.sh][1] is a small shell script that generates a self-extractable +compressed tar archive from a directory. The resulting file appears as a shell script +(many of those have a **.run** suffix), and can be launched as is. The archive +will then uncompress itself to a temporary directory and an optional arbitrary +command will be executed (for example an installation script). This is pretty +similar to archives generated with WinZip Self-Extractor in the Windows world. +Makeself archives also include checksums for integrity self-validation (CRC +and/or MD5/SHA256 checksums). + +The makeself.sh script itself is used only to create the archives from a +directory of files. The resultant archive is actually a compressed (using +gzip, bzip2, or compress) TAR archive, with a small shell script stub at the +beginning. This small stub performs all the steps of extracting the files, +running the embedded command, and removing the temporary files when done. +All the user has to do to install the software contained in such an +archive is to "run" the archive, i.e **sh nice-software.run**. I recommend +using the ".run" (which was introduced by some Makeself archives released by +Loki Software) or ".sh" suffix for such archives not to confuse the users, +so that they will know they are actually shell scripts (with quite a lot of binary data +attached to them though!). + +I am trying to keep the code of this script as portable as possible, i.e it is +not relying on any bash-specific features and only calls commands that are +installed on any functioning UNIX-compatible system. This script as well as +the archives it generates should run on any Unix flavor, with any compatible +Bourne shell, provided of course that the compression programs are available. + +As of version 2.1, Makeself has been rewritten and tested on the following +platforms : + + * Linux (all distributions) + * Sun Solaris (8 and above) + * HP-UX (tested on 11.0 and 11i on HPPA RISC) + * SCO OpenUnix and OpenServer + * IBM AIX 5.1L + * macOS (Darwin) + * SGI IRIX 6.5 + * FreeBSD + * UnicOS / Cray + * Cygwin (Windows) + +If you successfully run Makeself and/or archives created with it on another +system, then please [let me know][2]! + +Examples of publicly available archives made using makeself are : + + * Game patches and installers for [Id Software][3] games like Quake 3 for Linux or Return To Castle Wolfenstein ; + * All game patches released by [Loki Software][4] for the Linux version of popular games ; + * The [nVidia drivers][5] for Linux + * The installer for the Linux version of [Google Earth][6] + * The [VirtualBox][7] installers for Linux + * The [Makeself][1] distribution itself ;-) + * and countless others... + +**Important note for Apache users:** By default, most Web servers will think that Makeself archives are regular text files and thus they may show up as text in a Web browser. The correct way to prevent this is to add a MIME type for this file format, like so (in httpd.conf) : + +`AddType application/x-makeself .run` + +**Important note for certain GNU/Linux distributions:** Archives created with Makeself prior to v2.1.2 were using an old syntax for the _head_ and _tail_ Unix commands that is being progressively obsoleted in their GNU forms. Therefore you may have problems uncompressing some of these archives. A workaround for this is to set the environment variable $_POSIX2_VERSION to enable the old syntax, i.e. : + +`export _POSIX2_VERSION=199209` + +## Usage + +The syntax of makeself is the following: + +``` +makeself.sh [args] archive_dir file_name label startup_script [script_args] +``` + + * _args_ are optional options for Makeself. The available ones are : + + * **`--version`** : Prints the version number on stdout, then exits immediately + * **`--gzip`** : Use gzip for compression (the default on platforms on which gzip is commonly available, like Linux) + * **`--bzip2`** : Use bzip2 instead of gzip for better compression. The bzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. + * **`--pbzip2`** : Use pbzip2 instead of gzip for better and faster compression on machines having multiple CPUs. The pbzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. + * **`--xz`** : Use xz instead of gzip for better compression. The xz command must be available in the command path. It is recommended that the archive prefix be set to something like '.xz.run' for the archive, so that potential users know that they'll need xz to extract it. + * **`--lzo`** : Use lzop instead of gzip for better compression. The lzop command must be available in the command path. It is recommended that the archive prefix be set to something like `.lzo.run` for the archive, so that potential users know that they'll need lzop to extract it. + * **`--lz4`** : Use lz4 instead of gzip for better compression. The lz4 command must be available in the command path. It is recommended that the archive prefix be set to something like '.lz4.run' for the archive, so that potential users know that they'll need lz4 to extract it. + * **`--zstd`** : Use zstd instead of gzip for better compression. The zstd command must be available in the command path. It is recommended that the archive prefix be set to something like '.zstd.run' for the archive, so that potential users know that they'll need zstd to extract it. + * **`--pigz`** : Use pigz for compression. + * **`--base64`** : Encode the archive to ASCII in Base64 format instead of compressing (base64 command required). + * **`--gpg-encrypt`** : Encrypt the archive using `gpg -ac -z $COMPRESS_LEVEL`. This will prompt for a password to encrypt with. Assumes that potential users have `gpg` installed. + * **`--ssl-encrypt`** : Encrypt the archive using `openssl aes-256-cbc -a -salt`. This will prompt for a password to encrypt with. Assumes that the potential users have the OpenSSL tools installed. + * **`--compress`** : Use the UNIX `compress` command to compress the data. This should be the default on all platforms that don't have gzip available. + * **`--nocomp`** : Do not use any compression for the archive, which will then be an uncompressed TAR. + * **`--complevel`** : Specify the compression level for gzip, bzip2, pbzip2, zstd, xz, lzo or lz4. (defaults to 9) + * **`--threads`** : Specify the number of threads to be used by compressors that support parallelization. Omit to use compressor's default. Most useful (and required) for opting into xz's threading, usually with `--threads=0` for all available cores. pbzip2 and pigz are parallel by default, and setting this value allows limiting the number of threads they use. + * **`--notemp`** : The generated archive will not extract the files to a temporary directory, but in a new directory created in the current directory. This is better to distribute software packages that may extract and compile by themselves (i.e. launch the compilation through the embedded script). + * **`--current`** : Files will be extracted to the current directory, instead of in a subdirectory. This option implies `--notemp` above. + * **`--follow`** : Follow the symbolic links inside of the archive directory, i.e. store the files that are being pointed to instead of the links themselves. + * **`--append`** _(new in 2.1.x)_: Append data to an existing archive, instead of creating a new one. In this mode, the settings from the original archive are reused (compression type, label, embedded script), and thus don't need to be specified again on the command line. + * **`--header`** : Makeself uses a separate file to store the header stub, called `makeself-header.sh`. By default, it is assumed that it is stored in the same location as makeself.sh. This option can be used to specify its actual location if it is stored someplace else. + * **`--cleanup`** : Specify a script that is run when execution is interrupted or finishes successfully. The script is executed with the same environment and initial `script_args` as `startup_script`. + * **`--copy`** : Upon extraction, the archive will first extract itself to a temporary directory. The main application of this is to allow self-contained installers stored in a Makeself archive on a CD, when the installer program will later need to unmount the CD and allow a new one to be inserted. This prevents "Filesystem busy" errors for installers that span multiple CDs. + * **`--nox11`** : Disable the automatic spawning of a new terminal in X11. + * **`--nowait`** : When executed from a new X11 terminal, disable the user prompt at the end of the script execution. + * **`--nomd5`** and **`--nocrc`** : Disable the creation of a MD5 / CRC checksum for the archive. This speeds up the extraction process if integrity checking is not necessary. + * **`--sha256`** : Adds a SHA256 checksum for the archive. This is in addition to the MD5 / CRC checksums unless `--nomd5` is also used. + * **`--lsm` _file_** : Provide and LSM file to makeself, that will be embedded in the generated archive. LSM files are describing a software package in a way that is easily parseable. The LSM entry can then be later retrieved using the `--lsm` argument to the archive. An example of a LSM file is provided with Makeself. + * **`--tar-format opt`** : Specify the tar archive format (default is ustar); you may use any value accepted by your tar command (such as posix, v7, etc). + * **`--tar-extra opt`** : Append more options to the tar command line. + + For instance, in order to exclude the `.git` directory from the packaged archive directory using the GNU `tar`, one can use `makeself.sh --tar-extra "--exclude=.git" ...` + + * **`--keep-umask`** : Keep the umask set to shell default, rather than overriding when executing self-extracting archive. + * **`--packaging-date date`** : Use provided string as the packaging date instead of the current date. + * **`--license`** : Append a license file. + * **`--nooverwrite`** : Do not extract the archive if the specified target directory already exists. + * **`--help-header file`** : Add a header to the archive's `--help` output. + * `archive_dir` is the name of the directory that contains the files to be archived + * `file_name` is the name of the archive to be created + * `label` is an arbitrary text string describing the package. It will be displayed while extracting the files. + * `startup_script` is the command to be executed _from within_ the directory of extracted files. Thus, if you wish to execute a program contained in this directory, you must prefix your command with `./`. For example, `./program` will be fine. The `script_args` are additional arguments for this command. + +Here is an example, assuming the user has a package image stored in a **/home/joe/mysoft**, and he wants to generate a self-extracting package named +**mysoft.sh**, which will launch the "setup" script initially stored in /home/joe/mysoft : + +`makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup +` + +Here is also how I created the [makeself.run][9] archive which contains the Makeself distribution : + +`makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" ` + +Archives generated with Makeself can be passed the following arguments: + + * **`--keep`** : Prevent the files to be extracted in a temporary directory that will be removed after the embedded script's execution. The files will then be extracted in the current working directory and will stay here until you remove them. + * **`--verbose`** : Will prompt the user before executing the embedded command + * **`--target dir`** : Allows to extract the archive in an arbitrary place. + * **`--nox11`** : Do not spawn a X11 terminal. + * **`--confirm`** : Prompt the user for confirmation before running the embedded command. + * **`--info`** : Print out general information about the archive (does not extract). + * **`--lsm`** : Print out the LSM entry, if it is present. + * **`--list`** : List the files in the archive. + * **`--check`** : Check the archive for integrity using the embedded checksums. Does not extract the archive. + * **`--nochown`** : By default, a `chown -R` command is run on the target directory after extraction, so that all files belong to the current user. This is mostly needed if you are running as root, as tar will then try to recreate the initial user ownerships. You may disable this behavior with this flag. + * **`--tar`** : Run the tar command on the contents of the archive, using the following arguments as parameter for the command. + * **`--noexec`** : Do not run the embedded script after extraction. + * **`--noexec-cleanup`** : Do not run the embedded cleanup script. + * **`--nodiskspace`** : Do not check for available disk space before attempting to extract. + * **`--cleanup-args`** : Specify arguments to be passed to the cleanup script. Wrap value in quotes to specify multiple arguments. + +Any subsequent arguments to the archive will be passed as additional arguments to the embedded command. You must explicitly use the `--` special command-line construct before any such options to make sure that Makeself will not try to interpret them. + +## Startup Script + +The startup script must be a regular Shell script. + +Within the startup script, you can use the `$USER_PWD` variable to get the path of the folder from which the self-extracting script is executed. This is especially useful to access files that are located in the same folder as the script, as shown in the example below. + +`my-self-extracting-script.sh --fooBarFileParameter foo.bar` + +## Building and Testing + +Clone the git repo and execute `git submodule update --init --recursive` to obtain all submodules. + +* To make a release: `make` +* To run all tests: `make test` + +## Maven Usage + +Makeself is now supported by the following maven plugin [makeself-maven-plugin](https://github.com/hazendaz/makeself-maven-plugin). Please refer to project for usage and report any bugs in regards to maven plugin on that project. + +## License + +Makeself itself is covered by the [GNU General Public License][8] (GPL) version 2 and above. Archives generated by Makeself don't have to be placed under this license (although I encourage it ;-)), since the archive itself is merely data for Makeself. + +## Contributing + +I will gladly consider merging your pull requests on the [GitHub][10] repository. However, please keep the following in mind: + + * One of the main purposes of Makeself is portability. Do not submit patches that will break supported platforms. The more platform-agnostic, the better. + * Please explain clearly what the purpose of the patch is, and how you achieved it. + +## Download + +Get the latest official distribution [here][9] (version 2.4.2). + +The latest development version can be grabbed from [GitHub][10]. Feel free to submit any patches there through the fork and pull request process. + +## Version history + + * **v1.0:** Initial public release + * **v1.1:** The archive can be passed parameters that will be passed on to the embedded script, thanks to John C. Quillan + * **v1.2:** Cosmetic updates, support for bzip2 compression and non-temporary archives. Many ideas thanks to Francois Petitjean. + * **v1.3:** More patches from Bjarni R. Einarsson and Francois Petitjean: Support for no compression (`--nocomp`), script is no longer mandatory, automatic launch in an xterm, optional verbose output, and -target archive option to indicate where to extract the files. + * **v1.4:** Many patches from Francois Petitjean: improved UNIX compatibility, automatic integrity checking, support of LSM files to get info on the package at run time.. + * **v1.5.x:** A lot of bugfixes, and many other patches, including automatic verification through the usage of checksums. Version 1.5.5 was the stable release for a long time, even though the Web page didn't get updated ;-). Makeself was also officially made a part of the [Loki Setup installer][11], and its source is being maintained as part of this package. + * **v2.0:** Complete internal rewrite of Makeself. The command-line parsing was vastly improved, the overall maintenance of the package was greatly improved by separating the stub from makeself.sh. Also Makeself was ported and tested to a variety of Unix platforms. + * **v2.0.1:** First public release of the new 2.0 branch. Prior versions are officially obsoleted. This release introduced the `--copy` argument that was introduced in response to a need for the [UT2K3][12] Linux installer. + * **v2.1.0:** Big change : Makeself can now support multiple embedded tarballs, each stored separately with their own checksums. An existing archive can be updated with the `--append` flag. Checksums are also better managed, and the `--nochown` option for archives appeared. + * **v2.1.1:** Fixes related to the Unix compression (compress command). Some Linux distributions made the insane choice to make it unavailable, even though gzip is capable of uncompressing these files, plus some more bugfixes in the extraction and checksum code. + * **v2.1.2:** Some bug fixes. Use head -n to avoid problems with POSIX conformance. + * **v2.1.3:** Bug fixes with the command line when spawning terminals. Added `--tar`, `--noexec` for archives. Added `--nomd5` and `--nocrc` to avoid creating checksums in archives. The embedded script is now run through "eval". The `--info` output now includes the command used to create the archive. A man page was contributed by Bartosz Fenski. + * **v2.1.4:** Fixed `--info` output. Generate random directory name when extracting files to . to avoid problems. Better handling of errors with wrong permissions for the directory containing the files. Avoid some race conditions, Unset the $CDPATH variable to avoid problems if it is set. Better handling of dot files in the archive directory. + * **v2.1.5:** Made the md5sum detection consistent with the header code. Check for the presence of the archive directory. Added `--encrypt` for symmetric encryption through gpg (Eric Windisch). Added support for the digest command on Solaris 10 for MD5 checksums. Check for available disk space before extracting to the target directory (Andreas Schweitzer). Allow extraction to run asynchronously (patch by Peter Hatch). Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo). + * **v2.1.6:** Replaced one dot per file progress with a realtime progress percentage and a spinning cursor. Added `--noprogress` to prevent showing the progress during the decompression. Added `--target` dir to allow extracting directly to a target directory. (Guy Baconniere) + * **v2.2.0:** First major new release in years! Includes many bugfixes and user contributions. Please look at the [project page on Github][10] for all the details. + * **v2.3.0:** Support for archive encryption via GPG or OpenSSL. Added LZO and LZ4 compression support. Options to set the packaging date and stop the umask from being overriden. Optionally ignore check for available disk space when extracting. New option to check for root permissions before extracting. + * **v2.3.1:** Various compatibility updates. Added unit tests for Travis CI in the GitHub repo. New `--tar-extra`, `--untar-extra`, `--gpg-extra`, `--gpg-asymmetric-encrypt-sign` options. + * **v2.4.0:** Added optional support for SHA256 archive integrity checksums. + * **v2.4.2:** New --cleanup and --cleanup-args arguments for cleanup scripts. Added threading support for supported compressors. Now supports zstd compression. + * **v2.4.3:** Make explicit POSIX tar archives for increased compatibility. + * **v2.4.4:** Fixed various compatibility issues (no longer use POSIX tar archives), Github Actions to check on Solaris and FreeBSD. + * **v2.4.5:** Added `--tar-format` option to set the tar archive format (default is ustar) + +## Links + + * Check out the ["Loki Setup"][11] installer, used to install many Linux games and other applications, and of which I am the co-author. Since the demise of Loki, I am now the official maintainer of the project, and it is now being hosted here on GitHub. + * Bjarni R. Einarsson also wrote the **setup.sh** installer script, inspired by Makeself. [Check it out !][14] + +## Contact + +This script was written by [Stéphane Peter][15] (megastep at megastep.org). Any enhancements and suggestions are welcome. + +Contributions were included from John C. Quillan, Bjarni R. Einarsson, +Francois Petitjean, Ryan C. Gordon, and many contributors on GitHub. If you think I forgot +your name, don't hesitate to contact me. + +This project is now hosted on GitHub. Feel free to submit patches and bug reports on the [project page][10]. + +* * * + +[Stephane Peter][2] + + [1]: http://makeself.io/ + [2]: mailto:megastep@megastep.org + [3]: http://www.idsoftware.com/ + [4]: http://www.lokigames.com/products/myth2/updates.php3 + [5]: http://www.nvidia.com/ + [6]: http://earth.google.com/ + [7]: http://www.virtualbox.org/ + [8]: http://www.gnu.org/copyleft/gpl.html + [9]: https://github.com/megastep/makeself/releases/download/release-2.4.5/makeself-2.4.5.run + [10]: https://github.com/megastep/makeself + [11]: https://github.com/megastep/loki_setup/ + [12]: http://www.unrealtournament2003.com/ + [13]: http://www.icculus.org/ + [14]: http://bre.klaki.net/programs/setup.sh/ + [15]: https://stephanepeter.com/ diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION new file mode 100644 index 000000000..59aa62c1f --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION @@ -0,0 +1 @@ +2.4.5 diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh new file mode 100644 index 000000000..b5692d490 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh @@ -0,0 +1,9 @@ +#!/bin/sh +# +# Create a distributable archive of the current version of Makeself + +VER=`cat VERSION` +mkdir -p /tmp/makeself-$VER release +cp -pPR makeself* test README.md COPYING VERSION .gitmodules /tmp/makeself-$VER/ +./makeself.sh --notemp /tmp/makeself-$VER release/makeself-$VER.run "Makeself v$VER" echo "Makeself has extracted itself" + diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh new file mode 100644 index 000000000..940903148 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh @@ -0,0 +1,660 @@ +cat << EOF > "$archname" +#!/bin/bash +# This script was generated using Makeself $MS_VERSION +# The license covering this archive and its contents, if any, is wholly independent of the Makeself license (GPL) +# 2022.3.19-Modified the MS_Help function and some options +# Huawei Technologies Co., Ltd. + +ORIG_UMASK=\`umask\` + +CRCsum="$CRCsum" +MD5="$MD5sum" +SHA="$SHAsum" +SIGNATURE="$Signature" +TMPROOT=\${TMPDIR:="\$HOME"} +if ! test -d "\$TMPROOT"; then + TMPROOT="\$PWD" +fi +export TMPDIR="\$TMPROOT" +USER_PWD="\$PWD" +if ! test -d "\$USER_PWD"; then + exit 1 +fi +export USER_PWD +ARCHIVE_DIR=\`dirname "\$0"\` +export ARCHIVE_DIR + +name_of_file="\$0 " +pwd_of_file="\$PWD" +label="$LABEL" +script="$SCRIPT" +scriptargs="$SCRIPTARGS" +cleanup_script="${CLEANUP_SCRIPT}" +licensetxt="$LICENSE" +helpheader='$HELPHEADER' +targetdir="$archdirname" +filesizes="$filesizes" +totalsize="$totalsize" +keep="$KEEP" +nooverwrite="$NOOVERWRITE" +quiet="n" +accept="n" +nodiskspace="n" +export_conf="$EXPORT_CONF" +decrypt_cmd="$DECRYPT_CMD" +skip="$SKIP" + +print_cmd_arg="" +if type printf > /dev/null; then + print_cmd="printf" +elif test -x /usr/ucb/echo; then + print_cmd="/usr/ucb/echo" +else + print_cmd="echo" +fi + +if test -d /usr/xpg4/bin; then + PATH=/usr/xpg4/bin:\$PATH + export PATH +fi + +if test -d /usr/sfw/bin; then + PATH=\$PATH:/usr/sfw/bin + export PATH +fi + +unset CDPATH + +MS_Printf() +{ + \$print_cmd \$print_cmd_arg "\$1" +} + +MS_PrintLicense() +{ + PAGER=\${PAGER:=more} + if test x"\$licensetxt" != x; then + PAGER_PATH=\`exec <&- 2>&-; which \$PAGER || command -v \$PAGER || type \$PAGER\` + if test -x "\$PAGER_PATH"; then + echo "\$licensetxt" | \$PAGER + else + echo "\$licensetxt" + fi + if test x"\$accept" != xy; then + while true + do + MS_Printf "Please type y to accept, n otherwise: " + read yn + if test x"\$yn" = xn; then + keep=n + eval \$finish; exit 1 + break; + elif test x"\$yn" = xy; then + break; + fi + done + fi + fi +} + +MS_diskspace() +{ + ( + df -kP "\$1" | tail -1 | awk '{ if (\$4 ~ /%/) {print \$3} else {print \$4} }' + ) +} + +MS_dd() +{ + blocks=\`expr \$3 / 1024\` + bytes=\`expr \$3 % 1024\` + # Test for ibs, obs and conv feature + if dd if=/dev/zero of=/dev/null count=1 ibs=512 obs=512 conv=sync 2> /dev/null; then + dd if="\$1" ibs=\$2 skip=1 obs=1024 conv=sync 2> /dev/null | \\ + { test \$blocks -gt 0 && dd ibs=1024 obs=1024 count=\$blocks ; \\ + test \$bytes -gt 0 && dd ibs=1 obs=1024 count=\$bytes ; } 2> /dev/null + else + dd if="\$1" bs=\$2 skip=1 2> /dev/null + fi +} + +MS_dd_Progress() +{ + if test x"\$noprogress" = xy; then + MS_dd "\$@" + return \$? + fi + file="\$1" + offset=\$2 + length=\$3 + pos=0 + bsize=4194304 + while test \$bsize -gt \$length; do + bsize=\`expr \$bsize / 4\` + done + blocks=\`expr \$length / \$bsize\` + bytes=\`expr \$length % \$bsize\` + ( + dd ibs=\$offset skip=1 2>/dev/null + pos=\`expr \$pos \+ \$bsize\` + MS_Printf " 0%% " 1>&2 + if test \$blocks -gt 0; then + while test \$pos -le \$length; do + dd bs=\$bsize count=1 2>/dev/null + pcent=\`expr \$length / 100\` + pcent=\`expr \$pos / \$pcent\` + if test \$pcent -lt 100; then + MS_Printf "\b\b\b\b\b\b\b" 1>&2 + if test \$pcent -lt 10; then + MS_Printf " \$pcent%% " 1>&2 + else + MS_Printf " \$pcent%% " 1>&2 + fi + fi + pos=\`expr \$pos \+ \$bsize\` + done + fi + if test \$bytes -gt 0; then + dd bs=\$bytes count=1 2>/dev/null + fi + MS_Printf "\b\b\b\b\b\b\b" 1>&2 + MS_Printf " 100%% " 1>&2 + ) < "\$file" +} + +MS_Help() +{ + cat << EOH >&2 +Usage: \$0 [options] +Options: + --help | -h Print this message + --info Print embedded info : title, default target directory, embedded script ... + --list Print the list of files in the archive + --check Checks integrity and version dependency of the archive + --quiet Quiet install mode, skip human-computer interactions + --nox11 Do not spawn an xterm + --noexec Do not run embedded script + --extract= Extract directly to a target directory (absolute or relative) + Usually used with --noexec to just extract files without running + --tar arg1 [arg2 ...] Access the contents of the archive through the tar command +\${helpheader} +EOH +} + +MS_Verify_Sig() +{ + GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` + MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` + test -x "\$GPG_PATH" || GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` + test -x "\$MKTEMP_PATH" || MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` + offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` + temp_sig=\`mktemp -t XXXXX\` + echo \$SIGNATURE | base64 --decode > "\$temp_sig" + gpg_output=\`MS_dd "\$1" \$offset \$totalsize | LC_ALL=C "\$GPG_PATH" --verify "\$temp_sig" - 2>&1\` + gpg_res=\$? + rm -f "\$temp_sig" + if test \$gpg_res -eq 0 && test \`echo \$gpg_output | grep -c Good\` -eq 1; then + if test \`echo \$gpg_output | grep -c \$sig_key\` -eq 1; then + test x"\$quiet" = xn && echo "GPG signature is good" >&2 + else + echo "GPG Signature key does not match" >&2 + exit 2 + fi + else + test x"\$quiet" = xn && echo "GPG signature failed to verify" >&2 + exit 2 + fi +} + +MS_Check() +{ + OLD_PATH="\$PATH" + PATH=\${GUESS_MD5_PATH:-"\$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} + MD5_ARG="" + MD5_PATH=\`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum\` + test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which md5 || command -v md5 || type md5\` + test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which digest || command -v digest || type digest\` + PATH="\$OLD_PATH" + + SHA_PATH=\`exec <&- 2>&-; which shasum || command -v shasum || type shasum\` + test -x "\$SHA_PATH" || SHA_PATH=\`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum\` + + if test x"\$quiet" = xn; then + MS_Printf "Verifying archive integrity..." + fi + offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` + fsize=\`cat "\$1" | wc -c | tr -d " "\` + if test \$totalsize -ne \`expr \$fsize - \$offset\`; then + echo " Unexpected archive size." >&2 + exit 2 + fi + verb=\$2 + i=1 + for s in \$filesizes + do + crc=\`echo \$CRCsum | cut -d" " -f\$i\` + if test -x "\$SHA_PATH"; then + if test x"\`basename \$SHA_PATH\`" = xshasum; then + SHA_ARG="-a 256" + fi + sha=\`echo \$SHA | cut -d" " -f\$i\` + if test x"\$sha" = x0000000000000000000000000000000000000000000000000000000000000000; then + test x"\$verb" = xy && echo " \$1 does not contain an embedded SHA256 checksum." >&2 + else + shasum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$SHA_PATH \$SHA_ARG" | cut -b-64\`; + if test x"\$shasum" != x"\$sha"; then + echo "Error in SHA256 checksums: \$shasum is different from \$sha" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " SHA256 checksums are OK." >&2 + fi + crc="0000000000"; + fi + fi + if test -x "\$MD5_PATH"; then + if test x"\`basename \$MD5_PATH\`" = xdigest; then + MD5_ARG="-a md5" + fi + md5=\`echo \$MD5 | cut -d" " -f\$i\` + if test x"\$md5" = x00000000000000000000000000000000; then + test x"\$verb" = xy && echo " \$1 does not contain an embedded MD5 checksum." >&2 + else + md5sum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$MD5_PATH \$MD5_ARG" | cut -b-32\`; + if test x"\$md5sum" != x"\$md5"; then + echo "Error in MD5 checksums: \$md5sum is different from \$md5" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " MD5 checksums are OK." >&2 + fi + crc="0000000000"; verb=n + fi + fi + if test x"\$crc" = x0000000000; then + test x"\$verb" = xy && echo " \$1 does not contain a CRC checksum." >&2 + else + sum1=\`MS_dd_Progress "\$1" \$offset \$s | CMD_ENV=xpg4 cksum | awk '{print \$1}'\` + if test x"\$sum1" != x"\$crc"; then + echo "Error in checksums: \$sum1 is different from \$crc" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " CRC checksums are OK." >&2 + fi + fi + i=\`expr \$i + 1\` + offset=\`expr \$offset + \$s\` + done + if test x"\$quiet" = xn; then + echo " All good." + fi +} + +MS_Decompress() +{ + if test x"\$decrypt_cmd" != x""; then + { eval "\$decrypt_cmd" || echo " ... Decryption failed." >&2; } | eval "$GUNZIP_CMD" + else + eval "$GUNZIP_CMD" + fi + + if test \$? -ne 0; then + echo " ... Decompression failed." >&2 + fi +} + +UnTAR() +{ + if test x"\$quiet" = xn; then + tar \$1vf - $UNTAR_EXTRA 2>&1 || { echo " ... Extraction failed." >&2; kill -15 \$$; } + else + tar \$1f - $UNTAR_EXTRA 2>&1 || { echo Extraction failed. >&2; kill -15 \$$; } + fi +} + +MS_exec_cleanup() { + if test x"\$cleanup" = xy && test x"\$cleanup_script" != x""; then + cleanup=n + cd "\$tmpdir" + eval "\"\$cleanup_script\" \$scriptargs \$cleanupargs" + fi +} + +MS_cleanup() +{ + echo 'Signal caught, cleaning up' >&2 + MS_exec_cleanup + cd "\$TMPROOT" + rm -rf "\$tmpdir" + eval \$finish; exit 15 +} + +Script_Args_Check() +{ + script_supported_args=\$(echo \${helpheader} | grep -o -E "\-\-[^ ]+" | awk -F"=" {'print \$1'}) + arg_to_test=\$(echo \$1|awk -F"=" {'print \$1'}) + + for arg in \${script_supported_args}; + do + if test x"\$arg_to_test" = x"\$arg" ;then + return + fi + done + + MS_Help + exit 1 +} + +finish=true +xterm_loop= +noprogress=$NOPROGRESS +nox11=$NOX11 +copy=$COPY +ownership=$OWNERSHIP +verbose=n +cleanup=y +cleanupargs= +sig_key= + +initargs="\$@" + +while [ -n "\$*" ] +do + case "\$1" in + -h | --help) + MS_Help + exit 0 + ;; + -q | --quiet) + quiet=y + noprogress=y + shift + ;; + --info) + echo Identification: "\$label" + echo Target directory: "\$targetdir" + echo Uncompressed size: $USIZE KB + echo Compression: $COMPRESS + if test x"$ENCRYPT" != x""; then + echo Encryption: $ENCRYPT + fi + echo Date of packaging: $DATE + echo Built with Makeself version $MS_VERSION + echo Build command was: "$MS_COMMAND" + if test x"\$script" != x; then + echo Script run after extraction: + echo " " \$script \$scriptargs + fi + if test x"$copy" = xcopy; then + echo "Archive will copy itself to a temporary location" + fi + if test x"$NEED_ROOT" = xy; then + echo "Root permissions required for extraction" + fi + if test x"$KEEP" = xy; then + echo "directory \$targetdir is permanent" + else + echo "\$targetdir will be removed after extraction" + fi + exit 0 + ;; + --list) + echo Target directory: \$targetdir + offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + for s in \$filesizes + do + MS_dd "\$0" \$offset \$s | MS_Decompress | UnTAR t + offset=\`expr \$offset + \$s\` + done + exit 0 + ;; + --tar) + offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + arg1="\$2" + shift 2 || { MS_Help; exit 1; } + for s in \$filesizes + do + MS_dd "\$0" \$offset \$s | MS_Decompress | tar "\$arg1" - "\$@" + offset=\`expr \$offset + \$s\` + done + exit 0 + ;; + --check) + MS_Check "\$0" y + scriptargs="\$scriptargs \$1" + shift + ;; + --noexec) + script="" + cleanup_script="" + shift + ;; + --extract=*) + keep=y + targetdir=\`echo \$1 | cut -d"=" -f2 \` + if ! shift; then MS_Help; exit 1; fi + ;; + --nox11) + nox11=y + shift + ;; + --xwin) + if test "$NOWAIT" = n; then + finish="echo Press Return to close this window...; read junk" + fi + xterm_loop=1 + shift + ;; + --phase2) + copy=phase2 + shift + ;; + --repack | --repack-path=*) + Script_Args_Check \$1 + scriptargs="\$scriptargs '\$1'" + shift + if [[ ! "\$1" =~ ^-.* ]]; then + scriptargs="\$scriptargs '\$1'" + shift + fi + ;; + *) + Script_Args_Check \$1 + scriptargs="\$scriptargs '\$1'" + shift + ;; + esac +done + +quiet_para="" +if test x"\$quiet" = xy; then + quiet_para="--quiet " +fi +scriptargs="--\$name_of_file""--\"\$pwd_of_file\""" \$quiet_para""\$scriptargs" + +if test x"\$quiet" = xy -a x"\$verbose" = xy; then + echo Cannot be verbose and quiet at the same time. >&2 + exit 1 +fi + +if test x"$NEED_ROOT" = xy -a \`id -u\` -ne 0; then + echo "Administrative privileges required for this archive (use su or sudo)" >&2 + exit 1 +fi + +if test x"\$copy" \!= xphase2; then + MS_PrintLicense +fi + +case "\$copy" in +copy) + tmpdir="\$TMPROOT"/makeself.\$RANDOM.\`date +"%y%m%d%H%M%S"\`.\$\$ + mkdir "\$tmpdir" || { + echo "Could not create temporary directory \$tmpdir" >&2 + exit 1 + } + SCRIPT_COPY="\$tmpdir/makeself" + echo "Copying to a temporary location..." >&2 + cp "\$0" "\$SCRIPT_COPY" + chmod +x "\$SCRIPT_COPY" + cd "\$TMPROOT" + exec "\$SCRIPT_COPY" --phase2 -- \$initargs + ;; +phase2) + finish="\$finish ; rm -rf \`dirname \$0\`" + ;; +esac + +if test x"\$nox11" = xn; then + if tty -s; then # Do we have a terminal? + : + else + if test x"\$DISPLAY" != x -a x"\$xterm_loop" = x; then # No, but do we have X? + if xset q > /dev/null 2>&1; then # Check for valid DISPLAY variable + GUESS_XTERMS="xterm gnome-terminal rxvt dtterm eterm Eterm xfce4-terminal lxterminal kvt konsole aterm terminology" + for a in \$GUESS_XTERMS; do + if type \$a >/dev/null 2>&1; then + XTERM=\$a + break + fi + done + chmod a+x \$0 || echo Please add execution rights on \$0 + if test \`echo "\$0" | cut -c1\` = "/"; then # Spawn a terminal! + exec \$XTERM -e "\$0 --xwin \$initargs" + else + exec \$XTERM -e "./\$0 --xwin \$initargs" + fi + fi + fi + fi +fi + +if test x"\$targetdir" = x.; then + tmpdir="." +else + if test x"\$keep" = xy; then + if test x"\$nooverwrite" = xy && test -d "\$targetdir"; then + echo "Target directory \$targetdir already exists, aborting." >&2 + exit 1 + fi + if test x"\$quiet" = xn; then + echo "Creating directory \$targetdir" >&2 + fi + tmpdir="\$targetdir" + dashp="-p" + else + tmpdir="\$TMPROOT/selfgz\$\$\$RANDOM" + dashp="" + fi + mkdir \$dashp "\$tmpdir" || { + echo 'Cannot create target directory' \$tmpdir >&2 + echo 'You should try option --extract=' >&2 + eval \$finish + exit 1 + } +fi + +location="\`pwd\`" +if test x"\$SETUP_NOCHECK" != x1; then + MS_Check "\$0" +fi +offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + +if test x"\$verbose" = xy; then + MS_Printf "About to extract $USIZE KB in \$tmpdir ... Proceed ? [Y/n] " + read yn + if test x"\$yn" = xn; then + eval \$finish; exit 1 + fi +fi + +if test x"\$quiet" = xn; then + # Decrypting with openssl will ask for password, + # the prompt needs to start on new line + if test x"$ENCRYPT" = x"openssl"; then + echo "Decrypting and uncompressing \$label..." + else + MS_Printf "Uncompressing \$label" + fi +fi +res=3 +if test x"\$keep" = xn; then + trap MS_cleanup 1 2 3 15 +fi + +if test x"\$nodiskspace" = xn; then + leftspace=\`MS_diskspace "\$tmpdir"\` + if test -n "\$leftspace"; then + if test "\$leftspace" -lt $USIZE; then + echo + echo "Not enough space left in "\`dirname \$tmpdir\`" (\$leftspace KB) to decompress \$0 ($USIZE KB)" >&2 + if test x"\$keep" = xn; then + echo "Consider setting TMPDIR to a directory with more free space." + fi + eval \$finish; exit 1 + fi + fi +fi + +for s in \$filesizes +do + if MS_dd_Progress "\$0" \$offset \$s | MS_Decompress | ( cd "\$tmpdir"; umask \$ORIG_UMASK ; UnTAR xp ) 1>/dev/null; then + if test x"\$ownership" = xy; then + (cd "\$tmpdir"; chown -R \`id -u\` .; chgrp -R \`id -g\` .) + fi + else + echo >&2 + echo "Unable to decompress \$0" >&2 + eval \$finish; exit 1 + fi + offset=\`expr \$offset + \$s\` +done +if test x"\$quiet" = xn; then + echo +fi + +cd "\$tmpdir" +res=0 +if test x"\$script" != x; then + if test x"\$export_conf" = x"y"; then + MS_BUNDLE="\$0" + MS_LABEL="\$label" + MS_SCRIPT="\$script" + MS_SCRIPTARGS="\$scriptargs" + MS_ARCHDIRNAME="\$archdirname" + MS_KEEP="\$KEEP" + MS_NOOVERWRITE="\$NOOVERWRITE" + MS_COMPRESS="\$COMPRESS" + MS_CLEANUP="\$cleanup" + export MS_BUNDLE MS_LABEL MS_SCRIPT MS_SCRIPTARGS + export MS_ARCHDIRNAME MS_KEEP MS_NOOVERWRITE MS_COMPRESS + fi + + if test x"\$verbose" = x"y"; then + yn="x" + while test x"\$yn" != x -a x"\$yn" != xy -a x"\$yn" != xY -a x"\$yn" != xn -a x"\$yn" != xN + do + MS_Printf "OK to execute: \$script \$scriptargs \$* ? [Y/n] " + read yn + if test x"\$yn" = x -o x"\$yn" = xy -o x"\$yn" = xY; then + eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$?; + elif test x"\$yn" = xn -o x"\$yn" = xN; then + echo "Unable to decompress \$script ,because of aborting! ";res=\$? + else + echo "Input value is unacceptable,please try again." + fi + done + else + eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$? + fi + if test "\$res" -ne 0; then + test x"\$verbose" = xy && echo "The program '\$script' returned an error code (\$res)" >&2 + fi +fi + +MS_exec_cleanup + +if test x"\$keep" = xn; then + cd "\$TMPROOT" + rm -rf "\$tmpdir" +fi +eval \$finish; exit \$res +EOF diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 new file mode 100644 index 000000000..81bf6e4ff --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 @@ -0,0 +1,110 @@ +.TH "MAKESELF" "1" "2.4.5" +.SH "NAME" +makeself \- An utility to generate self-extractable archives. +.SH "SYNTAX" +.B makeself [\fIoptions\fP] archive_dir file_name label +.B [\fIstartup_script\fP] [\fIargs\fP] +.SH "DESCRIPTION" +This program is a free (GPL) utility designed to create self-extractable +archives from a directory. +.SH "OPTIONS" +The following options are supported. +.TP 15 +.B -v, --version +Prints out the makeself version number and exits. +.TP +.B -h, --help +Print out help information. +.TP +.B --tar-quietly +Suppress verbose output from the tar command +.TP +.B --quiet +Do not print any messages other than errors +.TP +.B --gzip +Compress using gzip (default if detected). +.TP +.B --bzip2 +Compress using bzip2. +.TP +.B --pbzip2 +Compress using pbzip2. +.TP +.B --xz +Compress using xz. +.TP +.B --lzo +Compress using lzop. +.TP +.B --lz4 +Compress using lz4. +.TP +.B --compress +Compress using the UNIX 'compress' command. +.TP +.B --nocomp +Do not compress the data. +.TP +.B --complevel lvl +Specify the compression level for gzip,bzip2,pbzui2,xz,lzo or lz4 +.TP +.B --notemp +The archive will create archive_dir in the current directory and +uncompress in ./archive_dir. +.TP +.B --copy +Upon extraction, the archive will first copy itself to a temporary directory. +.TP +.B --append +Append more files to an existing makeself archive. The label and startup scripts will then be ignored. +.TP +.B --current +Files will be extracted to the current directory. Both --current and --target dir imply --notemp. +.TP +.B --target dir +Extract directly to a target directory. Directory path can be either absolute or relative. +.TP +.B --header file +Specify location of the header script. +.TP +.B --cleanup file +Specify a cleanup script that executes on interrupt and when finished successfully. +.TP +.B --follow +Follow the symlinks in the archive. +.TP +.B --noprogress +Do not show the progress during the decompression. +.TP +.B --nox11 +Disable automatic spawn of an xterm if running in X11. +.TP +.B --nowait +Do not wait for user input after executing embedded program from an xterm. +.TP +.B --nomd5 +Do not create a MD5 checksum for the archive. +.TP +.B --nocrc +Do not create a CRC32 checksum for the archive. +.TP +.B --lsm file +LSM file describing the package. +.B --packaging-date date +Use provided string as the packaging date instead of the current date. +.SH "EXAMPLES" +Here is an example, assuming the user has a package image stored in a /home/joe/mysoft, +and he wants to generate a self-extracting package named mysoft.sh, which will launch +the "setup" script initially stored in /home/joe/mysoft: +.TP +makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup +.TP +Here is also how I created the makeself.run archive which contains the Makeself distribution: +.TP +makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" +.SH "AUTHORS" +Makeself has been written by Stéphane Peter . +.BR +This man page was originally written by Bartosz Fenski for the +Debian GNU/Linux distribution (but it may be used by others). diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm new file mode 100644 index 000000000..3c4cea8c1 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm @@ -0,0 +1,16 @@ +Begin3 +Title: makeself.sh +Version: 2.4.5 +Description: makeself.sh is a shell script that generates a self-extractable + tar.gz archive from a directory. The resulting file appears as a shell + script, and can be launched as is. The archive will then uncompress + itself to a temporary directory and an arbitrary command will be + executed (for example an installation script). This is pretty similar + to archives generated with WinZip Self-Extractor in the Windows world. +Keywords: Installation archive tar winzip +Author: Stephane Peter (megastep@megastep.org) +Maintained-by: Stephane Peter (megastep@megastep.org) +Original-site: https://makeself.io/ +Platform: Unix +Copying-policy: GPL +End diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh new file mode 100755 index 000000000..c8ea56597 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh @@ -0,0 +1,822 @@ +#!/bin/sh +# +# Makeself version 2.4.x +# by Stephane Peter +# +# Utility to create self-extracting tar.gz archives. +# The resulting archive is a file holding the tar.gz archive with +# a small Shell script stub that uncompresses the archive to a temporary +# directory and then executes a given script from withing that directory. +# +# Makeself home page: https://makeself.io/ +# +# Version 2.0 is a rewrite of version 1.0 to make the code easier to read and maintain. +# +# Version history : +# - 1.0 : Initial public release +# - 1.1 : The archive can be passed parameters that will be passed on to +# the embedded script, thanks to John C. Quillan +# - 1.2 : Package distribution, bzip2 compression, more command line options, +# support for non-temporary archives. Ideas thanks to Francois Petitjean +# - 1.3 : More patches from Bjarni R. Einarsson and Francois Petitjean: +# Support for no compression (--nocomp), script is no longer mandatory, +# automatic launch in an xterm, optional verbose output, and -target +# archive option to indicate where to extract the files. +# - 1.4 : Improved UNIX compatibility (Francois Petitjean) +# Automatic integrity checking, support of LSM files (Francois Petitjean) +# - 1.5 : Many bugfixes. Optionally disable xterm spawning. +# - 1.5.1 : More bugfixes, added archive options -list and -check. +# - 1.5.2 : Cosmetic changes to inform the user of what's going on with big +# archives (Quake III demo) +# - 1.5.3 : Check for validity of the DISPLAY variable before launching an xterm. +# More verbosity in xterms and check for embedded command's return value. +# Bugfix for Debian 2.0 systems that have a different "print" command. +# - 1.5.4 : Many bugfixes. Print out a message if the extraction failed. +# - 1.5.5 : More bugfixes. Added support for SETUP_NOCHECK environment variable to +# bypass checksum verification of archives. +# - 1.6.0 : Compute MD5 checksums with the md5sum command (patch from Ryan Gordon) +# - 2.0 : Brand new rewrite, cleaner architecture, separated header and UNIX ports. +# - 2.0.1 : Added --copy +# - 2.1.0 : Allow multiple tarballs to be stored in one archive, and incremental updates. +# Added --nochown for archives +# Stopped doing redundant checksums when not necesary +# - 2.1.1 : Work around insane behavior from certain Linux distros with no 'uncompress' command +# Cleaned up the code to handle error codes from compress. Simplified the extraction code. +# - 2.1.2 : Some bug fixes. Use head -n to avoid problems. +# - 2.1.3 : Bug fixes with command line when spawning terminals. +# Added --tar for archives, allowing to give arbitrary arguments to tar on the contents of the archive. +# Added --noexec to prevent execution of embedded scripts. +# Added --nomd5 and --nocrc to avoid creating checksums in archives. +# Added command used to create the archive in --info output. +# Run the embedded script through eval. +# - 2.1.4 : Fixed --info output. +# Generate random directory name when extracting files to . to avoid problems. (Jason Trent) +# Better handling of errors with wrong permissions for the directory containing the files. (Jason Trent) +# Avoid some race conditions (Ludwig Nussel) +# Unset the $CDPATH variable to avoid problems if it is set. (Debian) +# Better handling of dot files in the archive directory. +# - 2.1.5 : Made the md5sum detection consistent with the header code. +# Check for the presence of the archive directory +# Added --encrypt for symmetric encryption through gpg (Eric Windisch) +# Added support for the digest command on Solaris 10 for MD5 checksums +# Check for available disk space before extracting to the target directory (Andreas Schweitzer) +# Allow extraction to run asynchronously (patch by Peter Hatch) +# Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo) +# - 2.1.6 : Replaced one dot per file progress with a realtime progress percentage and a spining cursor (Guy Baconniere) +# Added --noprogress to prevent showing the progress during the decompression (Guy Baconniere) +# Added --target dir to allow extracting directly to a target directory (Guy Baconniere) +# - 2.2.0 : Many bugfixes, updates and contributions from users. Check out the project page on Github for the details. +# - 2.3.0 : Option to specify packaging date to enable byte-for-byte reproducibility. (Marc Pawlowsky) +# - 2.4.0 : Optional support for SHA256 checksums in archives. +# - 2.4.2 : Add support for threads for several compressors. (M. Limber) +# Added zstd support. +# - 2.4.3 : Make explicit POSIX tar archives for increased compatibility. +# - 2.4.5 : Added --tar-format to override ustar tar archive format +# +# (C) 1998-2021 by Stephane Peter +# +# This software is released under the terms of the GNU GPL version 2 and above +# Please read the license at http://www.gnu.org/copyleft/gpl.html +# Self-extracting archives created with this script are explictly NOT released under the term of the GPL +# + +MS_VERSION=2.4.5 +MS_COMMAND="$0" +unset CDPATH + +for f in ${1+"$@"}; do + MS_COMMAND="$MS_COMMAND \\\\ + \\\"$f\\\"" +done + +# For Solaris systems +if test -d /usr/xpg4/bin; then + PATH=/usr/xpg4/bin:$PATH + export PATH +fi + +# Procedures + +MS_Usage() +{ + echo "Usage: $0 [args] archive_dir file_name label startup_script [script_args]" + echo "args can be one or more of the following :" + echo " --version | -v : Print out Makeself version number and exit" + echo " --help | -h : Print out this help message" + echo " --tar-quietly : Suppress verbose output from the tar command" + echo " --quiet | -q : Do not print any messages other than errors." + echo " --gzip : Compress using gzip (default if detected)" + echo " --pigz : Compress with pigz" + echo " --zstd : Compress with zstd" + echo " --bzip2 : Compress using bzip2 instead of gzip" + echo " --pbzip2 : Compress using pbzip2 instead of gzip" + echo " --xz : Compress using xz instead of gzip" + echo " --lzo : Compress using lzop instead of gzip" + echo " --lz4 : Compress using lz4 instead of gzip" + echo " --compress : Compress using the UNIX 'compress' command" + echo " --complevel lvl : Compression level for gzip pigz zstd xz lzo lz4 bzip2 and pbzip2 (default 9)" + echo " --threads thds : Number of threads to be used by compressors that support parallelization." + echo " Omit to use compressor's default. Most useful (and required) for opting" + echo " into xz's threading, usually with '--threads=0' for all available cores." + echo " pbzip2 and pigz are parallel by default, and setting this value allows" + echo " limiting the number of threads they use." + echo " --base64 : Instead of compressing, encode the data using base64" + echo " --gpg-encrypt : Instead of compressing, encrypt the data using GPG" + echo " --gpg-asymmetric-encrypt-sign" + echo " : Instead of compressing, asymmetrically encrypt and sign the data using GPG" + echo " --gpg-extra opt : Append more options to the gpg command line" + echo " --ssl-encrypt : Instead of compressing, encrypt the data using OpenSSL" + echo " --ssl-passwd pass : Use the given password to encrypt the data using OpenSSL" + echo " --ssl-pass-src src : Use the given src as the source of password to encrypt the data" + echo " using OpenSSL. See \"PASS PHRASE ARGUMENTS\" in man openssl." + echo " If this option is not supplied, the user will be asked to enter" + echo " encryption password on the current terminal." + echo " --ssl-no-md : Do not use \"-md\" option not supported by older OpenSSL." + echo " --nochown : Do not give the target folder to the current user (default)" + echo " --chown : Give the target folder to the current user recursively" + echo " --nocomp : Do not compress the data" + echo " --notemp : The archive will create archive_dir in the" + echo " current directory and uncompress in ./archive_dir" + echo " --needroot : Check that the root user is extracting the archive before proceeding" + echo " --copy : Upon extraction, the archive will first copy itself to" + echo " a temporary directory" + echo " --append : Append more files to an existing Makeself archive" + echo " The label and startup scripts will then be ignored" + echo " --target dir : Extract directly to a target directory" + echo " directory path can be either absolute or relative" + echo " --nooverwrite : Do not extract the archive if the specified target directory exists" + echo " --current : Files will be extracted to the current directory" + echo " Both --current and --target imply --notemp" + echo " --tar-format opt : Specify a tar archive format (default is ustar)" + echo " --tar-extra opt : Append more options to the tar command line" + echo " --untar-extra opt : Append more options to the during the extraction of the tar archive" + echo " --nomd5 : Don't calculate an MD5 for archive" + echo " --nocrc : Don't calculate a CRC for archive" + echo " --sha256 : Compute a SHA256 checksum for the archive" + echo " --header file : Specify location of the header script" + echo " --cleanup file : Specify a cleanup script that executes on interrupt and when finished successfully." + echo " --follow : Follow the symlinks in the archive" + echo " --noprogress : Do not show the progress during the decompression" + echo " --nox11 : Disable automatic spawn of a xterm" + echo " --nowait : Do not wait for user input after executing embedded" + echo " program from an xterm" + echo " --sign passphrase : Signature private key to sign the package with" + echo " --lsm file : LSM file describing the package" + echo " --license file : Append a license file" + echo " --help-header file : Add a header to the archive's --help output" + echo " --packaging-date date" + echo " : Use provided string as the packaging date" + echo " instead of the current date." + echo + echo " --keep-umask : Keep the umask set to shell default, rather than overriding when executing self-extracting archive." + echo " --export-conf : Export configuration variables to startup_script" + echo + echo "Do not forget to give a fully qualified startup script name" + echo "(i.e. with a ./ prefix if inside the archive)." + exit 1 +} + +# Default settings +if type gzip >/dev/null 2>&1; then + COMPRESS=gzip +elif type compress >/dev/null 2>&1; then + COMPRESS=compress +else + echo "ERROR: missing commands: gzip, compress" >&2 + MS_Usage +fi +ENCRYPT=n +PASSWD="" +PASSWD_SRC="" +OPENSSL_NO_MD=n +COMPRESS_LEVEL=9 +DEFAULT_THREADS=123456 # Sentinel value +THREADS=$DEFAULT_THREADS +KEEP=n +CURRENT=n +NOX11=n +NOWAIT=n +APPEND=n +TAR_QUIETLY=n +KEEP_UMASK=n +QUIET=n +NOPROGRESS=n +COPY=none +NEED_ROOT=n +TAR_ARGS=rvf +TAR_FORMAT=ustar +TAR_EXTRA="" +GPG_EXTRA="" +DU_ARGS=-ks +HEADER=`dirname "$0"`/makeself-header.sh +SIGNATURE="" +TARGETDIR="" +NOOVERWRITE=n +DATE=`LC_ALL=C date` +EXPORT_CONF=n +SHA256=n +OWNERSHIP=n +SIGN=n +GPG_PASSPHRASE="" + +# LSM file stuff +LSM_CMD="echo No LSM. >> \"\$archname\"" + +while true +do + case "$1" in + --version | -v) + echo Makeself version $MS_VERSION + exit 0 + ;; + --pbzip2) + COMPRESS=pbzip2 + shift + ;; + --bzip2) + COMPRESS=bzip2 + shift + ;; + --gzip) + COMPRESS=gzip + shift + ;; + --pigz) + COMPRESS=pigz + shift + ;; + --zstd) + COMPRESS=zstd + shift + ;; + --xz) + COMPRESS=xz + shift + ;; + --lzo) + COMPRESS=lzo + shift + ;; + --lz4) + COMPRESS=lz4 + shift + ;; + --compress) + COMPRESS=compress + shift + ;; + --base64) + COMPRESS=base64 + shift + ;; + --gpg-encrypt) + COMPRESS=gpg + shift + ;; + --gpg-asymmetric-encrypt-sign) + COMPRESS=gpg-asymmetric + shift + ;; + --gpg-extra) + GPG_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-encrypt) + ENCRYPT=openssl + shift + ;; + --ssl-passwd) + PASSWD=$2 + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-pass-src) + PASSWD_SRC=$2 + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-no-md) + OPENSSL_NO_MD=y + shift + ;; + --nocomp) + COMPRESS=none + shift + ;; + --complevel) + COMPRESS_LEVEL="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --threads) + THREADS="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --nochown) + OWNERSHIP=n + shift + ;; + --chown) + OWNERSHIP=y + shift + ;; + --notemp) + KEEP=y + shift + ;; + --copy) + COPY=copy + shift + ;; + --current) + CURRENT=y + KEEP=y + shift + ;; + --tar-format) + TAR_FORMAT="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --tar-extra) + TAR_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --untar-extra) + UNTAR_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --target) + TARGETDIR="$2" + KEEP=y + shift 2 || { MS_Usage; exit 1; } + ;; + --sign) + SIGN=y + GPG_PASSPHRASE="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --nooverwrite) + NOOVERWRITE=y + shift + ;; + --needroot) + NEED_ROOT=y + shift + ;; + --header) + HEADER="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --cleanup) + CLEANUP_SCRIPT="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --license) + # We need to escape all characters having a special meaning in double quotes + LICENSE=$(sed 's/\\/\\\\/g; s/"/\\\"/g; s/`/\\\`/g; s/\$/\\\$/g' "$2") + shift 2 || { MS_Usage; exit 1; } + ;; + --follow) + TAR_ARGS=rvhf + DU_ARGS=-ksL + shift + ;; + --noprogress) + NOPROGRESS=y + shift + ;; + --nox11) + NOX11=y + shift + ;; + --nowait) + NOWAIT=y + shift + ;; + --nomd5) + NOMD5=y + shift + ;; + --sha256) + SHA256=y + shift + ;; + --nocrc) + NOCRC=y + shift + ;; + --append) + APPEND=y + shift + ;; + --lsm) + LSM_CMD="cat \"$2\" >> \"\$archname\"" + shift 2 || { MS_Usage; exit 1; } + ;; + --packaging-date) + DATE="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --help-header) + HELPHEADER=`sed -e "s/'/'\\\\\''/g" $2` + shift 2 || { MS_Usage; exit 1; } + [ -n "$HELPHEADER" ] && HELPHEADER="$HELPHEADER +" + ;; + --tar-quietly) + TAR_QUIETLY=y + shift + ;; + --keep-umask) + KEEP_UMASK=y + shift + ;; + --export-conf) + EXPORT_CONF=y + shift + ;; + -q | --quiet) + QUIET=y + shift + ;; + -h | --help) + MS_Usage + ;; + -*) + echo Unrecognized flag : "$1" + MS_Usage + ;; + *) + break + ;; + esac +done + +if test $# -lt 1; then + MS_Usage +else + if test -d "$1"; then + archdir="$1" + else + echo "Directory $1 does not exist." >&2 + exit 1 + fi +fi +archname="$2" + +if test "$QUIET" = "y" || test "$TAR_QUIETLY" = "y"; then + if test "$TAR_ARGS" = "rvf"; then + TAR_ARGS="rf" + elif test "$TAR_ARGS" = "rvhf"; then + TAR_ARGS="rhf" + fi +fi + +if test "$APPEND" = y; then + if test $# -lt 2; then + MS_Usage + fi + + # Gather the info from the original archive + OLDENV=`sh "$archname" --dumpconf` + if test $? -ne 0; then + echo "Unable to update archive: $archname" >&2 + exit 1 + else + eval "$OLDENV" + OLDSKIP=`expr $SKIP + 1` + fi +else + if test "$KEEP" = n -a $# = 3; then + echo "ERROR: Making a temporary archive with no embedded command does not make sense!" >&2 + echo >&2 + MS_Usage + fi + # We don't want to create an absolute directory unless a target directory is defined + if test "$CURRENT" = y; then + archdirname="." + elif test x"$TARGETDIR" != x; then + archdirname="$TARGETDIR" + else + archdirname=`basename "$1"` + fi + + if test $# -lt 3; then + MS_Usage + fi + + LABEL="$3" + SCRIPT="$4" + test "x$SCRIPT" = x || shift 1 + shift 3 + SCRIPTARGS="$*" +fi + +if test "$KEEP" = n -a "$CURRENT" = y; then + echo "ERROR: It is A VERY DANGEROUS IDEA to try to combine --notemp and --current." >&2 + exit 1 +fi + +case $COMPRESS in +gzip) + GZIP_CMD="gzip -c$COMPRESS_LEVEL" + GUNZIP_CMD="gzip -cd" + ;; +pigz) + GZIP_CMD="pigz -$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD --processes $THREADS" + fi + GUNZIP_CMD="gzip -cd" + ;; +zstd) + GZIP_CMD="zstd -$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD --threads=$THREADS" + fi + GUNZIP_CMD="zstd -cd" + ;; +pbzip2) + GZIP_CMD="pbzip2 -c$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD -p$THREADS" + fi + GUNZIP_CMD="bzip2 -d" + ;; +bzip2) + GZIP_CMD="bzip2 -$COMPRESS_LEVEL" + GUNZIP_CMD="bzip2 -d" + ;; +xz) + GZIP_CMD="xz -c$COMPRESS_LEVEL" + # Must opt-in by specifying a value since not all versions of xz support threads + if test $THREADS -ne $DEFAULT_THREADS; then + GZIP_CMD="$GZIP_CMD --threads=$THREADS" + fi + GUNZIP_CMD="xz -d" + ;; +lzo) + GZIP_CMD="lzop -c$COMPRESS_LEVEL" + GUNZIP_CMD="lzop -d" + ;; +lz4) + GZIP_CMD="lz4 -c$COMPRESS_LEVEL" + GUNZIP_CMD="lz4 -d" + ;; +base64) + GZIP_CMD="base64" + GUNZIP_CMD="base64 --decode -i -" + ;; +gpg) + GZIP_CMD="gpg $GPG_EXTRA -ac -z$COMPRESS_LEVEL" + GUNZIP_CMD="gpg -d" + ENCRYPT="gpg" + ;; +gpg-asymmetric) + GZIP_CMD="gpg $GPG_EXTRA -z$COMPRESS_LEVEL -es" + GUNZIP_CMD="gpg --yes -d" + ENCRYPT="gpg" + ;; +compress) + GZIP_CMD="compress -fc" + GUNZIP_CMD="(type compress >/dev/null 2>&1 && compress -fcd || gzip -cd)" + ;; +none) + GZIP_CMD="cat" + GUNZIP_CMD="cat" + ;; +esac + +if test x"$ENCRYPT" = x"openssl"; then + if test x"$APPEND" = x"y"; then + echo "Appending to existing archive is not compatible with OpenSSL encryption." >&2 + fi + + ENCRYPT_CMD="openssl enc -aes-256-cbc -salt" + DECRYPT_CMD="openssl enc -aes-256-cbc -d" + + if test x"$OPENSSL_NO_MD" != x"y"; then + ENCRYPT_CMD="$ENCRYPT_CMD -md sha256" + DECRYPT_CMD="$DECRYPT_CMD -md sha256" + fi + + if test -n "$PASSWD_SRC"; then + ENCRYPT_CMD="$ENCRYPT_CMD -pass $PASSWD_SRC" + elif test -n "$PASSWD"; then + ENCRYPT_CMD="$ENCRYPT_CMD -pass pass:$PASSWD" + fi +fi + +tmpfile="${TMPDIR:-/tmp}/mkself$$" + +if test -f "$HEADER"; then + oldarchname="$archname" + archname="$tmpfile" + # Generate a fake header to count its lines + SKIP=0 + . "$HEADER" + SKIP=`cat "$tmpfile" |wc -l` + # Get rid of any spaces + SKIP=`expr $SKIP` + rm -f "$tmpfile" + if test "$QUIET" = "n"; then + echo "Header is $SKIP lines long" >&2 + fi + archname="$oldarchname" +else + echo "Unable to open header file: $HEADER" >&2 + exit 1 +fi + +if test "$QUIET" = "n"; then + echo +fi + +if test "$APPEND" = n; then + if test -f "$archname"; then + echo "WARNING: Overwriting existing file: $archname" >&2 + fi +fi + +USIZE=`du $DU_ARGS "$archdir" | awk '{print $1}'` + +if test "." = "$archdirname"; then + if test "$KEEP" = n; then + archdirname="makeself-$$-`date +%Y%m%d%H%M%S`" + fi +fi + +test -d "$archdir" || { echo "Error: $archdir does not exist."; rm -f "$tmpfile"; exit 1; } +if test "$QUIET" = "n"; then + echo "About to compress $USIZE KB of data..." + echo "Adding files to archive named \"$archname\"..." +fi + +# See if we have GNU tar +TAR=`exec <&- 2>&-; which gtar || command -v gtar || type gtar` +test -x "$TAR" || TAR=tar + +tmparch="${TMPDIR:-/tmp}/mkself$$.tar" +( + if test "$APPEND" = "y"; then + tail -n "+$OLDSKIP" "$archname" | eval "$GUNZIP_CMD" > "$tmparch" + fi + cd "$archdir" + # "Determining if a directory is empty" + # https://www.etalabs.net/sh_tricks.html + find . \ + \( \ + ! -type d \ + -o \ + \( -links 2 -exec sh -c ' + is_empty () ( + cd "$1" + set -- .[!.]* ; test -f "$1" && return 1 + set -- ..?* ; test -f "$1" && return 1 + set -- * ; test -f "$1" && return 1 + return 0 + ) + is_empty "$0"' {} \; \ + \) \ + \) -print \ + | LC_ALL=C sort \ + | sed 's/./\\&/g' \ + | xargs $TAR $TAR_EXTRA --format $TAR_FORMAT -$TAR_ARGS "$tmparch" +) || { + echo "ERROR: failed to create temporary archive: $tmparch" + rm -f "$tmparch" "$tmpfile" + exit 1 +} + +USIZE=`du $DU_ARGS "$tmparch" | awk '{print $1}'` + +eval "$GZIP_CMD" <"$tmparch" >"$tmpfile" || { + echo "ERROR: failed to create temporary file: $tmpfile" + rm -f "$tmparch" "$tmpfile" + exit 1 +} +rm -f "$tmparch" + +if test x"$ENCRYPT" = x"openssl"; then + echo "About to encrypt archive \"$archname\"..." + { eval "$ENCRYPT_CMD -in $tmpfile -out ${tmpfile}.enc" && mv -f ${tmpfile}.enc $tmpfile; } || \ + { echo Aborting: could not encrypt temporary file: "$tmpfile".; rm -f "$tmpfile"; exit 1; } +fi + +fsize=`cat "$tmpfile" | wc -c | tr -d " "` + +# Compute the checksums + +shasum=0000000000000000000000000000000000000000000000000000000000000000 +md5sum=00000000000000000000000000000000 +crcsum=0000000000 + +if test "$NOCRC" = y; then + if test "$QUIET" = "n"; then + echo "skipping crc at user request" + fi +else + crcsum=`CMD_ENV=xpg4 cksum < "$tmpfile" | sed -e 's/ /Z/' -e 's/ /Z/' | cut -dZ -f1` + if test "$QUIET" = "n"; then + echo "CRC: $crcsum" + fi +fi + +if test "$SHA256" = y; then + SHA_PATH=`exec <&- 2>&-; which shasum || command -v shasum || type shasum` + if test -x "$SHA_PATH"; then + shasum=`eval "$SHA_PATH -a 256" < "$tmpfile" | cut -b-64` + else + SHA_PATH=`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum` + shasum=`eval "$SHA_PATH" < "$tmpfile" | cut -b-64` + fi + if test "$QUIET" = "n"; then + if test -x "$SHA_PATH"; then + echo "SHA256: $shasum" + else + echo "SHA256: none, SHA command not found" + fi + fi +fi +if test "$NOMD5" = y; then + if test "$QUIET" = "n"; then + echo "Skipping md5sum at user request" + fi +else + # Try to locate a MD5 binary + OLD_PATH=$PATH + PATH=${GUESS_MD5_PATH:-"$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} + MD5_ARG="" + MD5_PATH=`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum` + test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which md5 || command -v md5 || type md5` + test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which digest || command -v digest || type digest` + PATH=$OLD_PATH + if test -x "$MD5_PATH"; then + if test `basename ${MD5_PATH}`x = digestx; then + MD5_ARG="-a md5" + fi + md5sum=`eval "$MD5_PATH $MD5_ARG" < "$tmpfile" | cut -b-32` + if test "$QUIET" = "n"; then + echo "MD5: $md5sum" + fi + else + if test "$QUIET" = "n"; then + echo "MD5: none, MD5 command not found" + fi + fi +fi +if test "$SIGN" = y; then + GPG_PATH=`exec <&- 2>&-; which gpg || command -v gpg || type gpg` + if test -x "$GPG_PATH"; then + SIGNATURE=`$GPG_PATH --pinentry-mode=loopback --batch --yes --passphrase "$GPG_PASSPHRASE" --output - --detach-sig $tmpfile | base64 | tr -d \\\\n` + if test "$QUIET" = "n"; then + echo "Signature: $SIGNATURE" + fi + else + echo "Missing gpg command" >&2 + fi +fi + +totalsize=0 +for size in $fsize; +do + totalsize=`expr $totalsize + $size` +done + +if test "$APPEND" = y; then + mv "$archname" "$archname".bak || exit + + # Prepare entry for new archive + filesizes="$fsize" + CRCsum="$crcsum" + MD5sum="$md5sum" + SHAsum="$shasum" + Signature="$SIGNATURE" + # Generate the header + . "$HEADER" + # Append the new data + cat "$tmpfile" >> "$archname" + + chmod +x "$archname" + rm -f "$archname".bak + if test "$QUIET" = "n"; then + echo "Self-extractable archive \"$archname\" successfully updated." + fi +else + filesizes="$fsize" + CRCsum="$crcsum" + MD5sum="$md5sum" + SHAsum="$shasum" + Signature="$SIGNATURE" + + # Generate the header + . "$HEADER" + + # Append the compressed tar data after the stub + if test "$QUIET" = "n"; then + echo + fi + cat "$tmpfile" >> "$archname" + chmod +x "$archname" + if test "$QUIET" = "n"; then + echo Self-extractable archive \"$archname\" successfully created. + fi +fi +rm -f "$tmpfile" diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh new file mode 100644 index 000000000..31ee16511 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh @@ -0,0 +1,8 @@ +#!/bin/sh +# Run every available test - Bash needed +cd test +for test in *test; +do + echo "Running test $test ..." + bash $test || { echo "*** ERROR: Test '$test' failed!"; exit 1; } +done diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh new file mode 100755 index 000000000..a977bd51d --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh @@ -0,0 +1,31 @@ +#!/bin/bash + +project_path=$1 +build_path=$2 +vendor_name=customize +echo $@ +if [[ ! -d "$project_path" ]]; then + echo "[ERROR] No projcet path is provided" + exit 1 +fi + +if [[ ! -d "$build_path" ]]; then + echo "[ERROR] No build path is provided" + exit 1 +fi + +if [[ ! -d "$ASCEND_OPP_PATH" ]]; then + echo "[ERROR] No opp install path is provided" + exit 1 +fi +custom_exist_info_json=$ASCEND_OPP_PATH/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json +custom_new_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json +temp_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/temp_cust_aicpu_kernel.json + +if [[ -f "$custom_exist_info_json" ]] && [[ -f "$custom_new_info_json" ]]; then + cp -f $custom_exist_info_json $temp_info_json + chmod +w $temp_info_json + python3 ${project_path}/cmake/util/insert_op_info.py ${custom_new_info_json} ${temp_info_json} + cp -f $temp_info_json $custom_new_info_json + rm -f $temp_info_json +fi diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py new file mode 100755 index 000000000..c38e79edc --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py @@ -0,0 +1,260 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os + + +OP_ALL = '__ALLOP__' +SOC_ALL = '__ALLSOC__' +SOC_TO_SHORT_SOC_MAP = { + "ascend910a": "ascend910", + "ascend910proa": "ascend910", + "ascend910b": "ascend910", + "ascend910prob": "ascend910", + "ascend910premiuma": "ascend910", + "ascend910b1": "ascend910b", + "ascend910b2": "ascend910b", + "ascend910b3": "ascend910b", + "ascend910b4": "ascend910b", + "ascend310p1": "ascend310p", + "ascend310p3": "ascend310p", + "ascend310p3vir01": "ascend310p", + "ascend310p3vir02": "ascend310p", + "ascend310p3vir04": "ascend310p", + "ascend310p3vir08": "ascend310p", + "ascend310b1": "ascend310b", + "bs9sx1aa": "bs9sx1a" +} + + +class OpDesc: + def __init__(self: any, op_type: str): + self.op_type = op_type + self.attr_list = [] + self.attr_val = {} + self.input_name = [] + self.input_type = [] + self.input_dtype = [] + self.input_fmt = [] + self.output_name = [] + self.output_type = [] + self.output_dtype = [] + self.output_fmt = [] + self.op_fmt_sel = False + self.op_chk_support = False + self.op_intf = '' + self.kern_name = '' + self.op_file = '' + self.op_replay_flag = False + self.op_replay_batch = False + self.input_idx = -1 + self.output_idx = -1 + self.max_block_dim = 32 + self.max_shape_size = 268435456 + self.dynamic_shape = False + self.op_range_limit = '' + self.custom_compile_options = {} + self.custom_all_compile_options = {} + + @staticmethod + def _parse_digit(conf: str) -> int: + return int(conf.split('=')[1]) + + @staticmethod + def _parse_flag(conf: str) -> bool: + if 'true' == conf.split('=')[1]: + return True + return False + + @staticmethod + def _parse_str(conf: str) -> str: + return conf.split('=')[1] + + @staticmethod + def _parse_list(conf: str) -> list: + return conf.split('=')[1].split(',') + + def parse_input(self: any, conf: str): + if conf.startswith('input{}.name'.format(int(self.input_idx) + 1)): + self.input_idx += 1 + self.input_name.append(self._parse_str(conf)) + elif conf.startswith('input{}.paramType'.format(int(self.input_idx))): + self.input_type.append(self._parse_str(conf)) + elif conf.startswith('input{}.dtype'.format(int(self.input_idx))): + self.input_dtype.append(self._parse_str(conf)) + elif conf.startswith('input{}.format'.format(int(self.input_idx))): + self.input_fmt.append(self._parse_str(conf)) + else: + return + + def parse_output(self: any, conf: str): + if conf.startswith('output{}.name'.format(int(self.output_idx) + 1)): + self.output_idx += 1 + self.output_name.append(self._parse_str(conf)) + elif conf.startswith('output{}.paramType'.format(int(self.output_idx))): + self.output_type.append(self._parse_str(conf)) + elif conf.startswith('output{}.dtype'.format(int(self.output_idx))): + self.output_dtype.append(self._parse_str(conf)) + elif conf.startswith('output{}.format'.format(int(self.output_idx))): + self.output_fmt.append(self._parse_str(conf)) + else: + return + + def parse_op_format(self: any, conf: str): + self.op_fmt_sel = self._parse_flag(conf) + + def parse_check_support(self: any, conf: str): + self.op_chk_support = self._parse_flag(conf) + + def parse_range_limit(self: any, conf: str): + self.op_range_limit = self._parse_str(conf) + + def parse_kern_name(self: any, conf: str): + self.kern_name = self._parse_str(conf) + + def parse_op_intf(self: any, conf: str): + self.op_intf = self._parse_str(conf) + + def parse_op_file(self: any, conf: str): + self.op_file = self._parse_str(conf) + + def parse_dynamic_shape(self: any, conf: str): + self.dynamic_shape = self._parse_flag(conf) + + def parse_attr_list(self: any, conf: str): + self.attr_list = self._parse_list(conf) + + def parse_attr_val(self: any, conf: str): + for attr in self.attr_list: + if self.attr_val.get(attr) is None: + self.attr_val[attr] = {} + if conf.startswith('attr_{}.type'.format(attr)): + self.attr_val.get(attr)['type'] = self._parse_str(conf) + elif conf.startswith('attr_{}.paramType'.format(attr)): + self.attr_val.get(attr)['paramType'] = self._parse_str(conf) + elif conf.startswith('attr_{}.defaultValue'.format(attr)): + self.attr_val.get(attr)['defaultValue'] = self._parse_str(conf) + + def parse_replay_val(self: any, batch_list: list, iterator_list: list): + if self.op_type in batch_list: + self.op_replay_flag = True + self.op_replay_batch = True + elif self.op_type in iterator_list: + self.op_replay_flag = True + self.op_replay_batch = False + + +def _is_op_type_in_opdesc(op_descs: list, op_type: str): + for op in op_descs: + if op_type == op.op_type: + return True + return False + + +def _set_all_options_to_opdescs(op_descs, soc_ver_compile_options): + for op in op_descs: + op.custom_all_compile_options = soc_ver_compile_options + + +def _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options): + for op in op_descs: + if op.op_type != op_type: + continue + op.custom_compile_options = soc_ver_compile_options + + +def _trans_soc_ver_to_short(soc_ver: str): + low_soc_ver = soc_ver.lower() + if low_soc_ver not in SOC_TO_SHORT_SOC_MAP: + print(f'WARNING: caution: {soc_ver} will trans into ascend910, if not your intention,' + f'use ascend910b1~4 instead') + return SOC_TO_SHORT_SOC_MAP[low_soc_ver] + + +def _get_op_custom_options(op_descs: list, auto_gen_dir: str): + if auto_gen_dir is None: + return {} + file = os.path.join(auto_gen_dir, "custom_compile_options.ini") + if not os.path.exists(file): + print(f'WARNING: cannot find {auto_gen_dir}/custom_compile_options.ini') + return {} + with open (file, 'r') as fd: + lines = fd.readlines() + for line in lines: + param_list = str.split(line.rstrip('\n'), ',') + if len(param_list) != 3: + raise Exception(f'ERROR: custom compile option {param_list} len is not 3') + op_type = param_list[0] + if op_type.upper() == 'ALL': + op_type = OP_ALL + if op_type != OP_ALL and _is_op_type_in_opdesc(op_descs, op_type) == False: + print(f'WARNING: op: {op_type} are not exists in this project') + continue + soc_ver_compile_options = {} + soc_ver = param_list[1] + options_str = param_list[2] + options = str.split(options_str, ';') + if soc_ver == '': + soc_ver_compile_options[SOC_ALL] = options + else: + soc_ver_list = str.split(soc_ver, ';') + for ver in soc_ver_list: + short_ver = _trans_soc_ver_to_short(ver) + soc_ver_compile_options[short_ver] = options + if op_type == OP_ALL: + _set_all_options_to_opdescs(op_descs, soc_ver_compile_options) + else: + _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options) + + +def get_op_desc(file: str, batch_list: list, iterator_list: list, builder: any, + op_type: list, auto_gen_dir: str = None) -> list: + op_descs = [] + op_match = False + with open (file, 'r') as fd: + lines = fd.readlines() + for line in lines: + line = line.strip() + if line.startswith('['): + name = line[1:-1] + if op_type is None or name in op_type: + op_match = True + op_desc = builder(name) + op_desc.parse_replay_val(batch_list, iterator_list) + op_descs.append(op_desc) + else: + op_match = False + if op_type is not None and len(op_descs) == len(op_type): + return op_descs + continue + if not op_match: + continue + if line.startswith('input'): + op_desc.parse_input(line) + elif line.startswith('output'): + op_desc.parse_output(line) + elif line.startswith('dynamicFormat.flag'): + op_desc.parse_op_format(line) + elif line.startswith('needCheckSupport.flag'): + op_desc.parse_check_support(line) + elif line.startswith('rangeLimit.value'): + op_desc.parse_range_limit(line) + elif line.startswith('opInterface.value'): + op_desc.parse_op_intf(line) + elif line.startswith('kernel.name'): + op_desc.parse_kern_name(line) + elif line.startswith('opFile.value'): + op_desc.parse_op_file(line) + elif line.startswith('dynamicShapeSupport.flag'): + op_desc.parse_dynamic_shape(line) + elif line.startswith('attr.list'): + op_desc.parse_attr_list(line) + elif line.startswith('attr_'): + op_desc.parse_attr_val(line) + _get_op_custom_options(op_descs, auto_gen_dir) + return op_descs diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py new file mode 100755 index 000000000..721465fee --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py @@ -0,0 +1,338 @@ +# Copyright 2020-2021 Huawei Technologies Co., Ltd +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +""" +parser ini to json +""" + +import json +import os +import stat +import sys + + +ATTR_TYPE_LIST = ["int", "float", "bool", "str", "listInt", "listFloat", "listBool", "listStr", "listListInt", + "type", "listType", "tensor", "listTensor"] +ATTR_PARAMTYPE_LIST = ["optional", "required"] +BOOL_FLAG_KEY = ["dynamicFormat", "dynamicShapeSupport", "dynamicRankSupport", "precision_reduce", "heavyOp", + "needCheckSupport"] +BOOL_LIST = ["true", "false"] +DTYPE_LIST = ["float16", "float", "float32", "int8", "int16", "int32", "uint8", "uint16", "uint32", "bool", + "int64", "uint64", "qint8", "qint16", "qint32", "quint8", "quint16", "double", "complex64", + "complex128", "string", "resource", "dual", "dual_sub_int8", "dual_sub_uint8", "string_ref", + "int4", "bfloat16", "uint1"] +FORMAT_LIST = ["NCHW", "NHWC", "ND", "NC1HWC0", "FRACTAL_Z", "NC1C0HWPAD", "NHWC1C0", "FSR_NCHW", "FRACTAL_DECONV", + "C1HWNC0", "FRACTAL_DECONV_TRANSPOSE", "FRACTAL_DECONV_SP_STRIDE_TRANS", "NC1HWC0_C04", + "FRACTAL_Z_C04", "CHWN", "FRACTAL_DECONV_SP_STRIDE8_TRANS", "HWCN", "NC1KHKWHWC0", "BN_WEIGHT", + "FILTER_HWCK", "HASHTABLE_LOOKUP_LOOKUPS", "HASHTABLE_LOOKUP_KEYS", "HASHTABLE_LOOKUP_VALUE", + "HASHTABLE_LOOKUP_OUTPUT", "HASHTABLE_LOOKUP_HITS", "C1HWNCoC0", "MD", "NDHWC", "FRACTAL_ZZ", + "FRACTAL_NZ", "NCDHW", "DHWCN", "NDC1HWC0", "FRACTAL_Z_3D", "CN", "NC", "DHWNC", + "FRACTAL_Z_3D_TRANSPOSE", "FRACTAL_ZN_LSTM", "FRACTAL_ZN_RNN", "FRACTAL_Z_G", "NULL"] + + +def parse_ini_files(ini_files): + """ + parse ini files to json + Parameters: + ---------------- + ini_files:input file list + return:ops_info + ---------------- + """ + tbe_ops_info = {} + for ini_file in ini_files: + check_file_size(ini_file) + parse_ini_to_obj(ini_file, tbe_ops_info) + return tbe_ops_info + + +def check_file_size(input_file): + try: + file_size = os.path.getsize(input_file) + except OSError as os_error: + print('[ERROR] Failed to open "%s". %s' % (input_file, str(os_error))) + raise OSError from os_error + if file_size > 10*1024*1024: + print('[WARN] The size of %s exceeds 10MB, it may take more time to run, please wait.' % input_file) + + +def parse_ini_to_obj(ini_file, tbe_ops_info): + """ + parse ini file to json obj + Parameters: + ---------------- + ini_file:ini file path + tbe_ops_info:ops_info + ---------------- + """ + with open(ini_file) as ini_file: + lines = ini_file.readlines() + op_dict = {} + op_name = "" + find_op_type = False + for line in lines: + line = line.rstrip() + if line == "": + continue + if line.startswith("["): + if line.endswith("]"): + op_name = line[1:-1] + op_dict = {} + tbe_ops_info[op_name] = op_dict + find_op_type = True + elif "=" in line: + key1 = line[:line.index("=")] + key2 = line[line.index("=")+1:] + key1_0, key1_1 = key1.split(".") + if key1_0 not in op_dict: + op_dict[key1_0] = {} + if key1_1 in op_dict.get(key1_0): + raise RuntimeError("Op:" + op_name + " " + key1_0 + " " + + key1_1 + " is repeated!") + dic_key = op_dict.get(key1_0) + dic_key[key1_1] = key2 + else: + continue + if not find_op_type: + raise RuntimeError("Not find OpType in .ini file.") + + +def check_output_exist(op_dict, is_valid): + """ + Function Description: + Check output is exist + Parameter: op_dict + Parameter: is_valid + """ + if "output0" in op_dict: + output0_dict = op_dict.get("output0") + if output0_dict.get("name", None) is None: + is_valid = False + print("output0.name is required in .ini file!") + else: + is_valid = False + print("output0 is required in .ini file!") + return is_valid + + +def check_attr_dict(attr_dict, is_valid, attr): + """ + Function Description: + Check attr_dict + Parameter: attr_dict + Parameter: is_valid + Parameter: attr + """ + attr_type = attr_dict.get("type") + value = attr_dict.get("value") + param_type = attr_dict.get("paramType") + if attr_type is None or value is None: + is_valid = False + print("If attr.list is exist, {0}.type and {0}.value is required".format(attr)) + if param_type and param_type not in ATTR_PARAMTYPE_LIST: + is_valid = False + print("{0}.paramType only support {1}.".format(attr, ATTR_PARAMTYPE_LIST)) + if attr_type and attr_type not in ATTR_TYPE_LIST: + is_valid = False + print("{0}.type only support {1}.".format(attr, ATTR_TYPE_LIST)) + return is_valid + + +def check_attr(op_dict, is_valid): + """ + Function Description: + Check attr + Parameter: op_dict + Parameter: is_valid + """ + if "attr" in op_dict: + attr_dict = op_dict.get("attr") + attr_list_str = attr_dict.get("list", None) + if attr_list_str is None: + is_valid = False + print("attr.list is required in .ini file!") + else: + attr_list = attr_list_str.split(",") + for attr_name in attr_list: + attr = "attr_" + attr_name.strip() + attr_dict = op_dict.get(attr) + if attr_dict: + is_valid = check_attr_dict(attr_dict, is_valid, attr) + else: + is_valid = False + print("%s is required in .ini file, when attr.list is %s!" % (attr, attr_list_str)) + return is_valid + + +def check_bool_flag(op_dict, is_valid): + """ + Function Description: + check_bool_flag + Parameter: op_dict + Parameter: is_valid + """ + for key in BOOL_FLAG_KEY: + if key in op_dict: + op_bool_key = op_dict.get(key) + if op_bool_key.get("flag").strip() not in BOOL_LIST: + is_valid = False + print("{0}.flag only support {1}.".format(key, BOOL_LIST)) + return is_valid + + +def check_type_format(op_info, is_valid, op_info_key): + """ + Function Description: + Check type and format + Parameter: op_info + Parameter: is_valid + Parameter: op_info_key + """ + op_info_dtype_str = op_info.get("dtype") + op_info_dtype_num = 0 + op_info_format_num = 0 + if op_info_dtype_str: + op_info_dtype = op_info_dtype_str.split(",") + op_info_dtype_num = len(op_info_dtype) + for dtype in op_info_dtype: + if dtype.strip() not in DTYPE_LIST: + is_valid = False + print("{0}.dtype not support {1}.".format(op_info_key, dtype)) + op_info_format_str = op_info.get("format") + if op_info_format_str: + op_info_format = op_info_format_str.split(",") + op_info_format_num = len(op_info_format) + for op_format in op_info_format: + if op_format.strip() not in FORMAT_LIST: + is_valid = False + print("{0}.format not support {1}.".format(op_info_key, op_format)) + if op_info_dtype_num > 0 and op_info_format_num > 0: + if op_info_dtype_num != op_info_format_num: + is_valid = False + print("The number of {0}.dtype not match the number of {0}.format.".format(op_info_key)) + return is_valid + + +def check_op_info(tbe_ops): + """ + Function Description: + Check info. + Parameter: tbe_ops + Return Value: is_valid + """ + print("\n\n==============check valid for ops info start==============") + required_op_input_info_keys = ["paramType", "name"] + required_op_output_info_keys = ["paramType", "name"] + param_type_valid_value = ["dynamic", "optional", "required"] + is_valid = True + for op_key in tbe_ops: + op_dict = tbe_ops[op_key] + for op_info_key in op_dict: + if op_info_key.startswith("input"): + op_input_info = op_dict[op_info_key] + missing_keys = [] + for required_op_input_info_key in required_op_input_info_keys: + if required_op_input_info_key not in op_input_info: + missing_keys.append(required_op_input_info_key) + if len(missing_keys) > 0: + print("op: " + op_key + " " + op_info_key + " missing: " + + ",".join(missing_keys)) + is_valid = False + else: + if not op_input_info["paramType"] in param_type_valid_value: + print("op: " + op_key + " " + op_info_key + \ + " paramType not valid, valid key:[dynamic, " + "optional, required]") + is_valid = False + is_valid = check_type_format(op_input_info, is_valid, op_info_key) + if op_info_key.startswith("output"): + op_input_info = op_dict[op_info_key] + missing_keys = [] + for required_op_input_info_key in required_op_output_info_keys: + if required_op_input_info_key not in op_input_info: + missing_keys.append(required_op_input_info_key) + if len(missing_keys) > 0: + print("op: " + op_key + " " + op_info_key + " missing: " + + ",".join(missing_keys)) + is_valid = False + else: + if not op_input_info["paramType"] in param_type_valid_value: + print("op: " + op_key + " " + op_info_key + + " paramType not valid, valid key:[dynamic, " + "optional, required]") + is_valid = False + is_valid = check_type_format(op_input_info, is_valid, op_info_key) + is_valid = check_attr(op_dict, is_valid) + is_valid = check_bool_flag(op_dict, is_valid) + print("==============check valid for ops info end================\n\n") + return is_valid + + +def write_json_file(tbe_ops_info, json_file_path): + """ + Save info to json file + Parameters: + ---------------- + tbe_ops_info: ops_info + json_file_path: json file path + ---------------- + """ + json_file_real_path = os.path.realpath(json_file_path) + wr_flag = os.O_WRONLY | os.O_CREAT + wr_mode = stat.S_IWUSR | stat.S_IRUSR + with os.fdopen(os.open(json_file_real_path, wr_flag, wr_mode), 'w') as file_path: + # Only the owner and group have rights + os.chmod(json_file_real_path, stat.S_IWGRP + stat.S_IWUSR + stat.S_IRGRP + + stat.S_IRUSR) + json.dump(tbe_ops_info, file_path, sort_keys=True, indent=4, + separators=(',', ':')) + print("Compile op info cfg successfully.") + + +def parse_ini_to_json(ini_file_paths, outfile_path): + """ + parse ini files to json file + Parameters: + ---------------- + ini_file_paths: list of ini file path + outfile_path: output file path + ---------------- + """ + tbe_ops_info = parse_ini_files(ini_file_paths) + if not check_op_info(tbe_ops_info): + print("Compile op info cfg failed.") + return False + write_json_file(tbe_ops_info, outfile_path) + return True + + +if __name__ == '__main__': + args = sys.argv + + OUTPUT_FILE_PATH = "tbe_ops_info.json" + ini_file_path_list = [] + + for arg in args: + if arg.endswith("ini"): + ini_file_path_list.append(arg) + OUTPUT_FILE_PATH = arg.replace(".ini", ".json") + if arg.endswith("json"): + OUTPUT_FILE_PATH = arg + + if len(ini_file_path_list) == 0: + ini_file_path_list.append("tbe_ops_info.ini") + + if not parse_ini_to_json(ini_file_path_list, OUTPUT_FILE_PATH): + sys.exit(1) + sys.exit(0) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/preset_parse.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/preset_parse.py new file mode 100755 index 000000000..8f1124b1d --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/preset_parse.py @@ -0,0 +1,23 @@ +import json +import sys +import os + + +def get_config_opts(file): + src_dir = os.path.abspath(os.path.dirname(file)) + opts = '' + with open(file, 'r') as fd: + config = json.load(fd) + for conf in config: + if conf == 'configurePresets': + for node in config[conf]: + macros = node.get('cacheVariables') + if macros is not None: + for key in macros: + opts += '-D{}={} '.format(key, macros[key]['value']) + opts = opts.replace('${sourceDir}', src_dir) + print(opts) + + +if __name__ == "__main__": + get_config_opts(sys.argv[1]) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py new file mode 100755 index 000000000..1baa364ef --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py @@ -0,0 +1,105 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import os +import stat +import collections +import kernel_entry as keb +from tiling_data_def_build import gen_tiling +import code_channel_infer +import const_var + +PYF_PATH = os.path.dirname(__file__) + +ReplayCodeGenParams = collections.namedtuple('ReplayCodeGenParams',\ +['op_type', 'impl', 'tiling_file', 'kernel', 'entry', 'argn', 'op_replay_batch', 'max_block_dim', 'max_shape_size']) + + +class ReplayCodeGen: + def __init__(self, replayCodeGenParams): + self.op_type = replayCodeGenParams.op_type + self.impl = replayCodeGenParams.impl + self.tiling_file = replayCodeGenParams.tiling_file + self.tiling_data_file = '' + self.kernel = replayCodeGenParams.kernel + self.entry = replayCodeGenParams.entry + self.argn = replayCodeGenParams.argn + self.batch = False + self.outdir = '' + self.data_type = 'uint8_t' + self.blknum = 32 + self.op_replay_batch = replayCodeGenParams.op_replay_batch + self.max_block_dim = replayCodeGenParams.max_block_dim + self.max_shape_size = replayCodeGenParams.max_shape_size + + def set_batch(self, is_batch): + self.batch = is_batch + + def set_outdir(self, outdir): + self.outdir = outdir + + def gen_replay(self, ops_product: str): + kerentry = os.path.join(self.outdir, self.kernel + '_entry.cce') + kerimpl = os.path.join(self.outdir, self.kernel + '_impl.cpp') + replayimpl = os.path.join(self.outdir, self.kernel + '_replay.cpp') + if self.batch: + reptmp = os.path.join(PYF_PATH, 'batch_replay_impl.temp') + else: + reptmp = os.path.join(PYF_PATH, 'replay_impl.temp') + kertmp = os.path.join(PYF_PATH, 'kernel_impl.temp') + self._gen_kentry(kerentry) + self._gen_kimpl_code(kerimpl, kertmp) + self._gen_tiling_data_header() + self._gen_replay_code(replayimpl, reptmp, ops_product) + + def _gen_tiling_data_header(self): + self.tiling_data_file = os.path.join(self.outdir, self.kernel + '_tiling_data.h') + gen_tiling(self.tiling_file, self.tiling_data_file) + + def _gen_kimpl_code(self, src, tmpfile): + with open(tmpfile, 'r') as fd: + temp = fd.read() + temp = temp.replace('__CCE_FILE__', self.impl) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(temp) + + def _gen_replay_code(self, src, tmpfile, ops_product: str): + with open(tmpfile, 'r') as fd: + temp = fd.read() + temp = temp.replace('__ARG_NUM__', str(self.argn)) + argdef = [] + kargs = [] + for i in range(0, self.argn): + argdef.append('{} *'.format(self.data_type)) + kargs.append('({} *)GetArg({})'.format(self.data_type, i)) + temp = temp.replace('__ARGS_DEF__', ', '.join(argdef)) + temp = temp.replace('__KERNEL_ARGS__', ', '.join(kargs)) + temp = temp.replace('__KERNEL_FUN__', self.entry) + core_type_infer = 'core_type' + code_channel = code_channel_infer.infer_code_channel(code_channel_infer.InfoCodeChanelParams(self.impl,\ + self.tiling_data_file, self.kernel, self.outdir, ops_product, None)) + if code_channel == code_channel_infer.CODE_VEC: + core_type_infer = '0' + elif code_channel == code_channel_infer.CODE_CUBE: + core_type_infer = '1' + temp = temp.replace('__CORE_TYPE__', core_type_infer) + # regist function + temp = temp.replace('__OPS_PRODUCT__', ops_product) + temp = temp.replace('__OPTYPE__', self.op_type) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(temp) + + def _gen_kentry(self, src): + kf = '' + pre_alloc_str = 'A' * 256 + if self.batch: + kf += keb.batch_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str), self.argn, self.data_type) + else: + kf += keb.mc_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str),\ + self.argn, self.data_type, self.blknum) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(kf) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp new file mode 100755 index 000000000..1d30dd865 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp @@ -0,0 +1,120 @@ +#include +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#include "register/op_check.h" +#define __ASCENDC_REPLAY_CODE__ +using namespace std; +using namespace optiling; +using namespace AscendCReplay; + +extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); +extern "C" int elf_append(char *elf, uint32_t elfSize, char *jit, int kernum, int blknum[], char *atext[], + int alen[], int atlen, const char* kernelname[]); + +#define KERNEL_N 1 +#define ARG_N (__ARG_NUM__) +#define MAX_L (1024 * 1024 * 100) +#define MAX_E (1024 * 1024) + +int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) +{ + // gen type 1 : direct call codes 0: load .o file + if (param.gentype < 0 || param.gentype > 1) { + printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); + return 0; + } else if (param.gentype == 1 && param.objptr == nullptr) { + printf("Error: call replay with direct call mode, but code obj addr is null\n"); + return 0; + } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { + printf("Error: call replay with object file mode, but object file path is null\n"); + return 0; + } + // core_type 0:MIX 1:CUBE 2:VEC + if (core_type < 0 || core_type > 2) { + printf("Error: call replay core type is %d !\n", core_type); + return 0; + } + g_coreType = __CORE_TYPE__; + g_taskRation = param.task_ration; + g_tilingKey = param.tiling_key; + + unsigned char *buf, *jit; + char *kernel[KERNEL_N * 32]; + int len[KERNEL_N * 32]; + int blknum[KERNEL_N]; + int max; + block_num = param.block_dim; + g_ubBase = block_num; + uint8_t *code = (uint8_t *)malloc(MAX_L); + uint8_t *pos = code; + struct timespec tp1, tp2; + + clock_gettime(CLOCK_MONOTONIC, &tp1); + if (block_num > 32) { + printf("Error: block_num > 32\n"); + return 0; + } + //__OP_FOPEN__ + for (int i = 0; i < KERNEL_N; i++) { + for (int j = 0; j < ARG_N; j++) + AddArg(j, ARG_STEP * (j + 1)); + for (block_idx = 0; block_idx < block_num; block_idx++) { + //__OP_SET_KERNEL__ + int code_idx = i * block_num + block_idx; +#ifdef FP_CEILING + SetCtrlFloatEnable(); +#else + SetCtrlFloatDisable(); +#endif + CodeInit(pos, false); + __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); + CodeEnd(); + kernel[code_idx] = (char *)pos; + len[code_idx] = CodeLen(); + pos += len[code_idx]; + printf("kernel %d core %ld code generated len %d\n", i, block_idx, len[code_idx]); + } + blknum[i] = block_num; + } + //__OP_FCLOSE__ + clock_gettime(CLOCK_MONOTONIC, &tp2); + buf = (unsigned char *)malloc(MAX_E); + int fd = open(param.entry_file, O_RDONLY); + if (fd < 0) { + printf("[error]: cannot find entry.o : %s\n", param.entry_file); + return 0; + } + uint32_t bufSize = read(fd, buf, MAX_E); + if (bufSize <= 0) { + printf("[error]: entry.o : %s is too small ! \n", param.entry_file); + } + close(fd); + jit = (unsigned char *)malloc(MAX_L); + printf("total code generated %ld\n", pos - code); + int sz = elf_append((char *)buf, bufSize, (char *)jit, KERNEL_N, blknum, kernel, len, pos - code, ¶m.kernel_name); + if (tp1.tv_sec != tp2.tv_sec) { + printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); + } else { + printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); + } + printf("new elf size %d\n", sz); + if (param.gentype == 0) { + fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); + (void)write(fd, jit, sz); + close(fd); + free(jit); + } else if (param.gentype == 1) { + *param.objptr = (char*)jit; + } + free(buf); + free(code); + return sz; +} + +REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py new file mode 100755 index 000000000..a96304261 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py @@ -0,0 +1,84 @@ +#!/usr/bin/env python +# coding=utf-8 +""" +Function: +The replay funtion entry +Copyright Information: +Huawei Technologies Co., Ltd. All Rights Reserved © 2020 +""" + +import sys +import os +import stat +import re +import const_var + + +def gen_tiling(tiling_header_file: str, tiling_file_out: str): + if not os.path.exists(tiling_header_file): + print("warning: no userdef tiling header file: ", tiling_header_file) + return + print("generate tiling def header file: ", tiling_file_out) + tmp_name = os.path.splitext(os.path.basename(tiling_header_file))[0].upper() + tiling_source = '#ifndef __{}_H__\n'.format(tmp_name) + tiling_source += '#define __{}_H__\n\n'.format(tmp_name) + tiling_source += '#include \n' + tiling_source += '#include \n\n' + tiling_source += '#include "kernel_tiling/kernel_tiling.h"\n\n' + end_source = "" + pattern = re.compile(r'[(](.*)[)]', re.S) + with open(tiling_header_file, 'r') as fd: + lines = fd.readlines() + for line in lines: + line = line.strip() + if (line.startswith('BEGIN_TILING_DATA_DEF')): + tiling_source += '#pragma pack(1)\n' + tiling_source += 'struct ' + struct_def = re.findall(pattern, line)[0] + tiling_source += struct_def + ' {\n' + elif (line.startswith('TILING_DATA_FIELD_DEF_ARR')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {}[{}] = {{}};\n'.format(fds[0].strip(), fds[2].strip(), fds[1].strip()) + elif (line.startswith('TILING_DATA_FIELD_DEF_STRUCT')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {};\n'.format(fds[0].strip(), fds[1].strip()) + elif (line.startswith('TILING_DATA_FIELD_DEF')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {} = 0;\n'.format(fds[0].strip(), fds[1].strip()) + elif (line.startswith('END_TILING_DATA_DEF')): + tiling_source += '};\n' + tiling_source += '#pragma pack()\n\n' + tiling_source += '#ifdef __NPU_TILING__\n' + tiling_source += \ + 'inline [aicore] void Init{stru}(const __gm__ uint8_t* tiling, {stru}* const_data)\n'\ + .format(stru=struct_def) + tiling_source += '{\n' + tiling_source += ' const __gm__ uint32_t *src = (const __gm__ uint32_t *)tiling;\n' + tiling_source += ' uint32_t *dst = (uint32_t *)const_data;\n' + tiling_source += ' for (auto i = 0; i < sizeof({}) / 4; i++) *(dst + i) = *(src + i);\n'\ + .format(struct_def) + tiling_source += '}\n' + tiling_source += '#else\n' + tiling_source += 'inline void Init{stru}(uint8_t* tiling, {stru}* const_data)\n'.format(stru=struct_def) + tiling_source += '{\n' + tiling_source += ' memcpy(const_data, tiling, sizeof({}));\n'.format(struct_def) + tiling_source += '}\n' + tiling_source += '#endif\n\n' + end_source = ''' +#define GET_TILING_DATA(tiling_data, tiling_arg) \\ +{stru} tiling_data; \\ +Init{stru}(tiling_arg, &tiling_data)\n +'''.format(stru=struct_def) + tiling_source += end_source + tiling_source += '#endif' + with os.fdopen(os.open(tiling_file_out, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(tiling_source) + + +if __name__ == '__main__': + if len(sys.argv) <= 2: + raise RuntimeError('arguments must greater than 2') + gen_tiling(sys.argv[1], sys.argv[2]) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/CMakeLists.txt new file mode 100644 index 000000000..b6be9b492 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/CMakeLists.txt @@ -0,0 +1,11 @@ +if(NOT EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/mindspore") + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/caffe_plugin") + add_subdirectory(caffe_plugin) + endif() + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/tf_plugin") + add_subdirectory(tf_plugin) + endif() + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/onnx_plugin") + add_subdirectory(onnx_plugin) + endif() +endif() diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt new file mode 100644 index 000000000..a6aba5c20 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt @@ -0,0 +1,14 @@ + +aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} plugin_srcs) +add_library(cust_tf_parsers SHARED ${plugin_srcs}) +target_compile_definitions(cust_tf_parsers PRIVATE google=ascend_private) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_tf_parsers PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_tf_parsers PRIVATE intf_pub graph) +install(TARGETS cust_tf_parsers + LIBRARY DESTINATION packages/vendors/${vendor_name}/framework/tensorflow +) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc new file mode 100644 index 000000000..2cd837ce5 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc @@ -0,0 +1,23 @@ +/* Copyright (C) 2020-2021. Huawei Technologies Co., Ltd. All +rights reserved. + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the Apache License Version 2.0. + * You may not use this file except in compliance with the License. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * Apache License for more details at + * http://www.apache.org/licenses/LICENSE-2.0 + */ + +#include "register/register.h" + +namespace domi { +// register op info to GE +REGISTER_CUSTOM_OP("AddCustom") + .FrameworkType(TENSORFLOW) // type: CAFFE, TENSORFLOW + .OriginOpType("Add") // name in tf module + .ParseParamsByOperatorFn(AutoMappingByOpFn); +} // namespace domi diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/CMakeLists.txt new file mode 100644 index 000000000..40dd51cfa --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/CMakeLists.txt @@ -0,0 +1,82 @@ + +aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} ops_srcs) + +opbuild(OPS_SRC ${ops_srcs} + OUT_DIR ${ASCEND_AUTOGEN_PATH} +) + +add_library(cust_op_proto SHARED ${ops_srcs} ${ASCEND_AUTOGEN_PATH}/op_proto.cc) +target_compile_definitions(cust_op_proto PRIVATE OP_PROTO_LIB) +target_compile_options(cust_op_proto PRIVATE + -fvisibility=hidden +) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_op_proto PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_op_proto PRIVATE + intf_pub + exe_graph + register + tiling_api + -Wl,--whole-archive + rt2_registry + -Wl,--no-whole-archive +) +set_target_properties(cust_op_proto PROPERTIES OUTPUT_NAME + cust_opsproto_rt2.0 +) +add_library(cust_optiling SHARED ${ops_srcs}) +target_compile_definitions(cust_optiling PRIVATE OP_TILING_LIB) +target_compile_options(cust_optiling PRIVATE + -fvisibility=hidden +) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_optiling PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_optiling PRIVATE + intf_pub + exe_graph + register + tiling_api + -Wl,--whole-archive + rt2_registry + -Wl,--no-whole-archive +) +set_target_properties(cust_optiling PROPERTIES OUTPUT_NAME + cust_opmaster_rt2.0 +) + +file(GLOB aclnn_src ${ASCEND_AUTOGEN_PATH}/aclnn_*.cpp) +file(GLOB aclnn_inc ${ASCEND_AUTOGEN_PATH}/aclnn_*.h) +add_library(cust_opapi SHARED ${aclnn_src}) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_opapi PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_opapi PRIVATE intf_pub ascendcl nnopbase) + +add_custom_target(optiling_compat ALL + COMMAND ln -sf lib/linux/${CMAKE_SYSTEM_PROCESSOR}/$ + ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so +) + +install(TARGETS cust_op_proto + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_proto/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) +install(FILES ${ASCEND_AUTOGEN_PATH}/op_proto.h + DESTINATION packages/vendors/${vendor_name}/op_proto/inc) +install(TARGETS cust_optiling + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) +install(FILES ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so + DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling) +install(TARGETS cust_opapi + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_api/lib) +install(FILES ${aclnn_inc} + DESTINATION packages/vendors/${vendor_name}/op_api/include) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom.cpp b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom.cpp new file mode 100644 index 000000000..b61a67f11 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom.cpp @@ -0,0 +1,117 @@ +#include "add_custom_tiling.h" +#include "register/op_def_registry.h" +#include "graph/utils/type_utils.h" +#include "tiling/platform/platform_ascendc.h" + +namespace optiling { +const uint32_t BLOCK_SIZE = 32; +const uint32_t BUFFER_NUM = 2; +static ge::graphStatus TilingFunc(gert::TilingContext* context) +{ + TilingData tiling; + uint64_t ubSize; + auto ascendcPlatform = platform_ascendc::PlatformAscendC(context->GetPlatformInfo()); + ascendcPlatform.GetCoreMemSize(platform_ascendc::CoreMemType::UB, ubSize); + auto coreNum = ascendcPlatform.GetCoreNum(); + + // Based on the input length and the number of inputs, the number of bytes of the input data type is obtained + uint32_t inputNum = context->GetInputShape(0)->GetStorageShape().GetShapeSize(); + uint32_t typeLength = 0; + ge::TypeUtils::GetDataTypeLength(context->GetInputDesc(0)->GetDataType(), typeLength); + uint32_t inputLength = inputNum * typeLength; + uint32_t inputBytes = inputLength / inputNum; + + // There are a total of 3 shared UB spaces in the input and output. If it's int8, there are 2 more TBUFs + uint32_t ubDataNumber = (inputBytes == 1) ? 5 : 3; + // The number of 32B data blocks that can be used for each data. DOUBLE BUFFER is already counted here + uint32_t tileBlockNum = (ubSize / BLOCK_SIZE / BUFFER_NUM) / ubDataNumber; + uint32_t tileDataNum = (tileBlockNum * BLOCK_SIZE) / inputBytes; + + // Input data for 32B alignment + uint32_t inputLengthAlgin32 = (((inputLength + BLOCK_SIZE - 1) / BLOCK_SIZE) * BLOCK_SIZE); + // There is at least 32B of data on each core, satisfying several settings for several cores. The maximum number of audits is the actual number of audits + coreNum = (coreNum < inputLengthAlgin32 / BLOCK_SIZE) ? coreNum : inputLengthAlgin32 / BLOCK_SIZE; + coreNum = (coreNum >= 1) ? coreNum : 1; + uint32_t everyCoreInputBlockNum = inputLengthAlgin32 / BLOCK_SIZE / coreNum; + uint32_t tailBlockNum = (inputLengthAlgin32 / BLOCK_SIZE) % coreNum; + + // Small chunks are calculated and sliced several times using the number of data on each core + uint32_t smallCoreDataNum = everyCoreInputBlockNum * BLOCK_SIZE / inputBytes; + uint32_t smallTileNum = everyCoreInputBlockNum / tileBlockNum; + uint32_t finalSmallTileNum = (everyCoreInputBlockNum % tileBlockNum) == 0 ? smallTileNum : smallTileNum + 1; + // Tail block calculation for small chunks of data + uint32_t smallTailDataNum = smallCoreDataNum - (tileDataNum * smallTileNum); + smallTailDataNum = smallTailDataNum == 0 ? tileDataNum : smallTailDataNum; + + // The total length of a large block of data is 32B larger than that of a small block of data + everyCoreInputBlockNum += 1; + uint32_t bigCoreDataNum = everyCoreInputBlockNum * BLOCK_SIZE / inputBytes; + uint32_t bigTileNum = everyCoreInputBlockNum / tileBlockNum; + uint32_t finalBigTileNum = (everyCoreInputBlockNum % tileBlockNum) == 0 ? bigTileNum : bigTileNum + 1; + uint32_t bigTailDataNum = bigCoreDataNum - tileDataNum * bigTileNum; + bigTailDataNum = bigTailDataNum == 0 ? tileDataNum : bigTailDataNum; + + tiling.set_smallCoreDataNum(smallCoreDataNum); + tiling.set_bigCoreDataNum(bigCoreDataNum); + tiling.set_tileDataNum(tileDataNum); + tiling.set_smallTailDataNum(smallTailDataNum); + tiling.set_bigTailDataNum(bigTailDataNum); + tiling.set_finalSmallTileNum(finalSmallTileNum); + tiling.set_finalBigTileNum(finalBigTileNum); + tiling.set_tailBlockNum(tailBlockNum); + + context->SetBlockDim(coreNum); + tiling.SaveToBuffer(context->GetRawTilingData()->GetData(), context->GetRawTilingData()->GetCapacity()); + context->GetRawTilingData()->SetDataSize(tiling.GetDataSize()); + size_t *currentWorkspace = context->GetWorkspaceSizes(1); + currentWorkspace[0] = 0; + return ge::GRAPH_SUCCESS; +} +} + +namespace ge { +static ge::graphStatus InferShape(gert::InferShapeContext* context) +{ + const gert::Shape* x1_shape = context->GetInputShape(0); + gert::Shape* y_shape = context->GetOutputShape(0); + *y_shape = *x1_shape; + return GRAPH_SUCCESS; +} +static graphStatus InferDataType(gert::InferDataTypeContext* context) +{ + const auto inputDataType = context->GetInputDataType(0); + context->SetOutputDataType(0, inputDataType); + return ge::GRAPH_SUCCESS; +} +} + +namespace ops { +class AddCustom : public OpDef { +public: + explicit AddCustom(const char* name) : OpDef(name) + { + this->Input("x") + .ParamType(REQUIRED) + .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) + .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) + .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); + this->Input("y") + .ParamType(REQUIRED) + .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) + .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) + .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); + this->Output("z") + .ParamType(REQUIRED) + .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) + .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) + .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); + + this->SetInferShape(ge::InferShape).SetInferDataType(ge::InferDataType); + this->AICore() + .SetTiling(optiling::TilingFunc) + .AddConfig("ascend310b") + .AddConfig("ascend910b"); + } +}; +OP_ADD(AddCustom); +} diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h new file mode 100644 index 000000000..28fd00a37 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h @@ -0,0 +1,22 @@ +/* + * Copyright (c) Huawei Technologies Co., Ltd. 2022-2023. All rights reserved. + */ +#ifndef ADD_CUSTOM_TILING_H +#define ADD_CUSTOM_TILING_H +#include "register/tilingdata_base.h" + +namespace optiling { +BEGIN_TILING_DATA_DEF(TilingData) + TILING_DATA_FIELD_DEF(uint32_t, smallCoreDataNum); + TILING_DATA_FIELD_DEF(uint32_t, bigCoreDataNum); + TILING_DATA_FIELD_DEF(uint32_t, finalBigTileNum); + TILING_DATA_FIELD_DEF(uint32_t, finalSmallTileNum); + TILING_DATA_FIELD_DEF(uint32_t, tileDataNum); + TILING_DATA_FIELD_DEF(uint32_t, smallTailDataNum); + TILING_DATA_FIELD_DEF(uint32_t, bigTailDataNum); + TILING_DATA_FIELD_DEF(uint32_t, tailBlockNum); +END_TILING_DATA_DEF; + +REGISTER_TILING_DATA_CLASS(AddCustom, TilingData) +} +#endif // ADD_CUSTOM_TILING_H \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt new file mode 100644 index 000000000..0d31a444c --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt @@ -0,0 +1,61 @@ +# set custom compile options +if ("${CMAKE_BUILD_TYPE}x" STREQUAL "Debugx") + add_ops_compile_options(ALL OPTIONS -g -O0) +endif() + +foreach(compute_unit ${ASCEND_COMPUTE_UNIT}) + + # generate aic-${compute_unit}-ops-info.json + add_ops_info_target(TARGET ops_info_gen_${compute_unit} + OUTPUT ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core/${compute_unit}/aic-${compute_unit}-ops-info.json + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/config/${compute_unit} + ) + + # generate ascendc impl py once + if (NOT TARGET ascendc_impl_gen) + add_ops_impl_target(TARGET ascendc_impl_gen + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl + ) + endif() + + # dynamic shape binary compile + if (${ENABLE_BINARY_PACKAGE}) + add_bin_compile_target(TARGET ascendc_bin_${compute_unit} + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} + ADP_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/dynamic + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/binary/${compute_unit} + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/kernel + COMPUTE_UNIT ${compute_unit} + ) + add_dependencies(ascendc_bin_${compute_unit} ascendc_impl_gen) + endif() + +endforeach() + +# generate npu_supported_ops.json +add_npu_support_target(TARGET npu_supported_ops + OPS_INFO_DIR ${ASCEND_AUTOGEN_PATH} + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core + INSTALL_DIR packages/vendors/${vendor_name}/framework/${ASCEND_FRAMEWORK_TYPE} +) + +if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) + add_subdirectory(testcases) +endif() + +# install kernel file +if (${ENABLE_SOURCE_PACKAGE}) + file(GLOB KERNEL_FILES + ${CMAKE_CURRENT_SOURCE_DIR}/*.cpp + ${CMAKE_CURRENT_SOURCE_DIR}/*.h + ${CMAKE_CURRENT_SOURCE_DIR}/*.py + ) + install(FILES ${KERNEL_FILES} + DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl/dynamic + ) +endif() diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp new file mode 100644 index 000000000..37c80a825 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp @@ -0,0 +1,134 @@ +/* + * Copyright (c) Huawei Technologies Co., Ltd. 2022-2023. All rights reserved. + * + * Function : z = x + y + * This sample is a very basic sample that implements vector add on Ascend plaform. + */ +#include "kernel_operator.h" +// tensor num for each queue +constexpr int32_t BUFFER_NUM = 2; + +template class KernelAdd { + using T = TYPE_X; +public: + __aicore__ inline KernelAdd() {} + __aicore__ inline void Init(GM_ADDR x, GM_ADDR y, GM_ADDR z, uint32_t smallCoreDataNum, + uint32_t bigCoreDataNum, uint32_t finalBigTileNum, + uint32_t finalSmallTileNum, uint32_t tileDataNum, + uint32_t smallTailDataNum, uint32_t bigTailDataNum, + uint32_t tailBlockNum) + { + ASSERT(AscendC::GetBlockNum() != 0 && "block dim can not be zero!"); + uint32_t coreNum = AscendC::GetBlockIdx(); + uint32_t globalBufferIndex = bigCoreDataNum * AscendC::GetBlockIdx(); + this->tileDataNum = tileDataNum; + if (coreNum < tailBlockNum) { + this->coreDataNum = bigCoreDataNum; + this->tileNum = finalBigTileNum; + this->tailDataNum = bigTailDataNum; + } + else { + this->coreDataNum = smallCoreDataNum; + this->tileNum = finalSmallTileNum; + this->tailDataNum = smallTailDataNum; + globalBufferIndex -= (bigCoreDataNum - smallCoreDataNum) * (AscendC::GetBlockIdx() - tailBlockNum); + } + xGm.SetGlobalBuffer((__gm__ TYPE_X*)x + globalBufferIndex, this->coreDataNum); + yGm.SetGlobalBuffer((__gm__ TYPE_Y*)y + globalBufferIndex, this->coreDataNum); + zGm.SetGlobalBuffer((__gm__ TYPE_Z*)z + globalBufferIndex, this->coreDataNum); + pipe.InitBuffer(inQueueX, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_X)); + pipe.InitBuffer(inQueueY, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_Y)); + pipe.InitBuffer(outQueueZ, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_Z)); + pipe.InitBuffer(tmp1, this->tileDataNum * sizeof(half)); + pipe.InitBuffer(tmp2, this->tileDataNum * sizeof(half)); + } + __aicore__ inline void Process() + { + int32_t loopCount = this->tileNum; + this->processDataNum = this->tileDataNum; + for (int32_t i = 0; i < loopCount; i++) { + if (i == this->tileNum - 1) { + this->processDataNum = this->tailDataNum; + } + CopyIn(i); + Compute(i); + CopyOut(i); + } + } + +private: + __aicore__ inline void CopyIn(int32_t progress) + { + AscendC::LocalTensor xLocal = inQueueX.AllocTensor(); + AscendC::LocalTensor yLocal = inQueueY.AllocTensor(); + AscendC::DataCopy(xLocal, xGm[progress * this->tileDataNum], this->processDataNum); + AscendC::DataCopy(yLocal, yGm[progress * this->tileDataNum], this->processDataNum); + inQueueX.EnQue(xLocal); + inQueueY.EnQue(yLocal); + } + __aicore__ inline void Compute(int32_t progress) + { + AscendC::LocalTensor xLocal = inQueueX.DeQue(); + AscendC::LocalTensor yLocal = inQueueY.DeQue(); + AscendC::LocalTensor zLocal = outQueueZ.AllocTensor(); + if constexpr (std::is_same_v) { + auto p1 = tmp1.Get(); + auto p2 = tmp2.Get(); + AscendC::Cast(p1, xLocal, AscendC::RoundMode::CAST_NONE, this->processDataNum); + AscendC::Cast(p2, yLocal, AscendC::RoundMode::CAST_NONE, this->processDataNum); + AscendC::Add(p2, p1, p2, this->processDataNum); + AscendC::Cast(p1.ReinterpretCast(), p2, AscendC::RoundMode::CAST_RINT, this->processDataNum); + AscendC::ShiftLeft(p1.ReinterpretCast(), p1.ReinterpretCast(), int16_t(8), this->processDataNum); + AscendC::ShiftRight(p1.ReinterpretCast(), p1.ReinterpretCast(), int16_t(8), this->processDataNum); + AscendC::Cast(p2, p1.ReinterpretCast(), AscendC::RoundMode::CAST_NONE, this->processDataNum); + AscendC::Cast(zLocal, p2, AscendC::RoundMode::CAST_NONE, this->processDataNum); + } + else { + AscendC::Add(zLocal, xLocal, yLocal, this->processDataNum); + } + outQueueZ.EnQue(zLocal); + inQueueX.FreeTensor(xLocal); + inQueueY.FreeTensor(yLocal); + } + __aicore__ inline void CopyOut(int32_t progress) + { + AscendC::LocalTensor zLocal = outQueueZ.DeQue(); + AscendC::DataCopy(zGm[progress * this->tileDataNum], zLocal, this->processDataNum); + outQueueZ.FreeTensor(zLocal); + } + +private: + AscendC::TPipe pipe; + AscendC::TQue inQueueX, inQueueY; + AscendC::TQue outQueueZ; + AscendC::TBuf tmp1, tmp2; + AscendC::GlobalTensor xGm; + AscendC::GlobalTensor yGm; + AscendC::GlobalTensor zGm; + uint32_t coreDataNum; + uint32_t tileNum; + uint32_t tileDataNum; + uint32_t tailDataNum; + uint32_t processDataNum; +}; + +extern "C" __global__ __aicore__ void add_custom(GM_ADDR x, GM_ADDR y, GM_ADDR z, GM_ADDR workspace, GM_ADDR tiling) +{ + GET_TILING_DATA(tiling_data, tiling); + KernelAdd op; + op.Init(x, y, z, tiling_data.smallCoreDataNum, + tiling_data.bigCoreDataNum, tiling_data.finalBigTileNum, + tiling_data.finalSmallTileNum, tiling_data.tileDataNum, + tiling_data.smallTailDataNum, tiling_data.bigTailDataNum, + tiling_data.tailBlockNum); + op.Process(); +} + +#ifndef ASCENDC_CPU_DEBUG +// call of kernel function +void add_custom_do(uint32_t blockDim, void* l2ctrl, void* stream, uint8_t* x, uint8_t* y, uint8_t* z, + uint8_t* workspace, uint8_t* tiling) +{ + add_custom<<>>(x, y, z, workspace, tiling); +} +#endif diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/install.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/install.sh new file mode 100644 index 000000000..9d45a8052 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/install.sh @@ -0,0 +1,318 @@ +#!/bin/bash +vendor_name=customize +targetdir=/usr/local/Ascend/opp +target_custom=0 + +sourcedir=$PWD/packages +vendordir=vendors/$vendor_name + +QUIET="y" + +while true +do + case $1 in + --quiet) + QUIET="y" + shift + ;; + --install-path=*) + INSTALL_PATH=$(echo $1 | cut -d"=" -f2-) + INSTALL_PATH=${INSTALL_PATH%*/} + shift + ;; + --*) + shift + ;; + *) + break + ;; + esac +done + +log() { + cur_date=`date +"%Y-%m-%d %H:%M:%S"` + echo "[runtime] [$cur_date] "$1 +} + +if [ -n "${INSTALL_PATH}" ]; then + if [[ ! "${INSTALL_PATH}" = /* ]]; then + log "[ERROR] use absolute path for --install-path argument" + exit 1 + fi + if [ ! -d ${INSTALL_PATH} ]; then + mkdir ${INSTALL_PATH} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${INSTALL_PATH} failed" + exit 1 + fi + fi + targetdir=${INSTALL_PATH} +elif [ -n "${ASCEND_CUSTOM_OPP_PATH}" ]; then + if [ ! -d ${ASCEND_CUSTOM_OPP_PATH} ]; then + mkdir -p ${ASCEND_CUSTOM_OPP_PATH} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${ASCEND_CUSTOM_OPP_PATH} failed" + fi + fi + targetdir=${ASCEND_CUSTOM_OPP_PATH} +else + if [ "x${ASCEND_OPP_PATH}" == "x" ]; then + log "[ERROR] env ASCEND_OPP_PATH no exist" + exit 1 + fi + targetdir="${ASCEND_OPP_PATH}" +fi + +if [ ! -d $targetdir ];then + log "[ERROR] $targetdir no exist" + exit 1 +fi + +upgrade() +{ + if [ ! -d ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 files" + return 0 + fi + + if [ ! -d ${targetdir}/$vendordir/$1 ];then + log "[INFO] create ${targetdir}/$vendordir/$1." + mkdir -p ${targetdir}/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/$1 failed" + return 1 + fi + else + has_same_file=-1 + for file_a in ${sourcedir}/$vendordir/$1/*; do + file_b=${file_a##*/}; + if [ "ls ${targetdir}/$vendordir/$1" = "" ]; then + log "[INFO] ${targetdir}/$vendordir/$1 is empty !!" + return 1 + fi + grep -q $file_b <<<`ls ${targetdir}/$vendordir/$1`; + if [[ $? -eq 0 ]]; then + echo -n "${file_b} " + has_same_file=0 + fi + done + if [ 0 -eq $has_same_file ]; then + if test $QUIET = "n"; then + echo "[INFO]: has old version in ${targetdir}/$vendordir/$1, \ + you want to Overlay Installation , please enter:[o]; \ + or replace directory installation , please enter: [r]; \ + or not install , please enter:[n]." + + while true + do + read orn + if [ "$orn" = n ]; then + return 0 + elif [ "$orn" = m ]; then + break; + elif [ "$0rn" = r ]; then + [ -n "${targetdir}/$vendordir/$1/" ] && rm -rf "${targetdir}/$vendordir/$1"/* + break; + else + echo "[ERROR] input error, please input again!" + fi + done + fi + fi + log "[INFO] replace or merge old ops $1 files .g....." + fi + + log "copy new ops $1 files ......" + if [ -d ${targetdir}/$vendordir/$1/ ]; then + chmod -R +w "$targetdir/$vendordir/$1/" >/dev/null 2>&1 + fi + cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 files failed" + return 1 + fi + + return 0 +} +upgrade_proto() +{ + if [ ! -f ${sourcedir}/$vendordir/custom.proto ]; then + log "[INFO] no need to upgrade custom.proto files" + return 0 + fi + if [ ! -d ${targetdir}/$vendordir/framework/caffe ];then + log "[INFO] create ${targetdir}/$vendordir/framework/caffe." + mkdir -p ${targetdir}/$vendordir/framework/caffe + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/framework/caffe failed" + return 1 + fi + else + if [ -f ${targetdir}/$vendordir/framework/caffe/custom.proto ]; then + # 有老版本,判断是否要覆盖式安装 + if test $QUIET = "n"; then + echo "[INFO] ${targetdir}/$vendordir/framework/caffe has old version"\ + "custom.proto file. Do you want to replace? [y/n] " + + while true + do + read yn + if [ "$yn" = n ]; then + return 0 + elif [ "$yn" = y ]; then + break; + else + echo "[ERROR] input error, please input again!" + fi + done + fi + fi + log "[INFO] replace old caffe.proto files ......" + fi + chmod -R +w "$targetdir/$vendordir/framework/caffe/" >/dev/null 2>&1 + cp -rf ${sourcedir}/$vendordir/custom.proto ${targetdir}/$vendordir/framework/caffe/ + if [ $? -ne 0 ];then + log "[ERROR] copy new custom.proto failed" + return 1 + fi + log "[INFO] copy custom.proto success" + + return 0 +} + +upgrade_file() +{ + if [ ! -e ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 file" + return 0 + fi + + log "copy new $1 files ......" + cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 file failed" + return 1 + fi + + return 0 +} + +delete_optiling_file() +{ + if [ ! -d ${targetdir}/vendors ];then + log "[INFO] $1 not exist, no need to uninstall" + return 0 + fi + sys_info=$(uname -m) + if [ ! -d ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/lib/linux/${sys_info} ];then + rm -rf ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/liboptiling.so + fi + return 0 +} + +log "[INFO] copy uninstall sh success" + +if [ ! -d ${targetdir}/vendors ];then + log "[INFO] create ${targetdir}/vendors." + mkdir -p ${targetdir}/vendors + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/vendors failed" + return 1 + fi +fi +chmod u+w ${targetdir}/vendors + +echo "[ops_custom]upgrade framework" +upgrade framework +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op proto" +upgrade op_proto +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade version.info" +upgrade_file version.info +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op impl" +delete_optiling_file op_impl +upgrade op_impl +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op api" +upgrade op_api +if [ $? -ne 0 ];then + exit 1 +fi + +upgrade_proto +if [ $? -ne 0 ];then + exit 1 +fi + +# set the set_env.bash +if [ -n "${INSTALL_PATH}" ] && [ -d ${INSTALL_PATH} ]; then + _ASCEND_CUSTOM_OPP_PATH=${targetdir}/${vendordir} + bin_path="${_ASCEND_CUSTOM_OPP_PATH}/bin" + set_env_variable="#!/bin/bash\nexport ASCEND_CUSTOM_OPP_PATH=${_ASCEND_CUSTOM_OPP_PATH}:\${ASCEND_CUSTOM_OPP_PATH}" + if [ ! -d ${bin_path} ]; then + mkdir -p ${bin_path} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${bin_path} failed" + exit 1 + fi + fi + echo -e ${set_env_variable} > ${bin_path}/set_env.bash + if [ $? -ne 0 ]; then + log "[ERROR] write ASCEND_CUSTOM_OPP_PATH to set_env.bash failed" + exit 1 + else + log "[INFO] using requirements: when custom module install finished or before you run the custom module, \ + execute the command [ source ${bin_path}/set_env.bash ] to set the environment path" + fi +else + config_file=${targetdir}/vendors/config.ini + if [ ! -f ${config_file} ]; then + touch ${config_file} + chmod 640 ${config_file} + echo "load_priority=$vendor_name" > ${config_file} + if [ $? -ne 0 ];then + echo "echo load_priority failed" + exit 1 + fi + else + found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" + found_vendor=$(echo $found_vendors | sed "s/\<$vendor_name\>//g" | tr ',' ' ') + vendor=$(echo $found_vendor | tr -s ' ' ',') + if [ "$vendor" != "" ]; then + sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" + fi + fi +fi + +chmod u-w ${targetdir}/vendors + +if [ -d ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/ ]; then + chmod -R 440 ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/* >/dev/null 2>&1 +fi +if [ -f ${targetdir}/ascend_install.info ]; then + chmod -R 440 ${targetdir}/ascend_install.info +fi +if [ -f ${targetdir}/scene.info ]; then + chmod -R 440 ${targetdir}/scene.info +fi +if [ -f ${targetdir}/version.info ]; then + chmod -R 440 ${targetdir}/version.info +fi + +echo "SUCCESS" +exit 0 + diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/upgrade.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/upgrade.sh new file mode 100644 index 000000000..34801efc4 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/upgrade.sh @@ -0,0 +1,151 @@ +#!/bin/bash +vendor_name=customize +targetdir=/usr/local/Ascend/opp +target_custom=0 + +sourcedir=$PWD/packages +vendordir=vendors/$vendor_name + +log() { + cur_date=`date +"%Y-%m-%d %H:%M:%S"` + echo "[runtime] [$cur_date] "$1 +} + +if [[ "x${ASCEND_OPP_PATH}" == "x" ]];then + log "[ERROR] env ASCEND_OPP_PATH no exist" + exit 1 +fi + +targetdir=${ASCEND_OPP_PATH} + +if [ ! -d $targetdir ];then + log "[ERROR] $targetdir no exist" + exit 1 +fi + +upgrade() +{ + if [ ! -d ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 files" + return 0 + fi + + if [ ! -d ${targetdir}/$vendordir/$1 ];then + log "[INFO] create ${targetdir}/$vendordir/$1." + mkdir -p ${targetdir}/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/$1 failed" + return 1 + fi + else + vendor_installed_dir=$(ls "$targetdir/vendors" 2> /dev/null) + for i in $vendor_installed_dir;do + vendor_installed_file=$(ls "$vendor_installed_dir/$vendor_name/$i" 2> /dev/null) + if [ "$i" = "$vendor_name" ] && [ "$vendor_installed_file" != "" ]; then + echo "[INFO]: $vendor_name custom opp package has been installed on the path $vendor_installed_dir, \ + you want to Overlay Installation , please enter:[o]; \ + or replace directory installation , please enter: [r]; \ + or not install , please enter:[n]." + fi + while true + do + read mrn + if [ "$mrn" = m ]; then + break + elif [ "$mrn" = r ]; then + [ -n "$vendor_installed_file"] && rm -rf "$vendor_installed_file" + break + elif [ "$mrn" = n ]; then + return 0 + else + echo "[WARNING]: Input error, please input m or r or n to choose!" + fi + done + done + log "[INFO] replace old ops $1 files ......" + fi + + log "copy new ops $1 files ......" + cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 files failed" + return 1 + fi + + return 0 +} + +upgrade_file() +{ + if [ ! -e ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 file" + return 0 + fi + + log "copy new $1 files ......" + cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 file failed" + return 1 + fi + + return 0 +} + +log "[INFO] copy uninstall sh success" + +echo "[ops_custom]upgrade framework" +upgrade framework +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op proto" +upgrade op_proto +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op impl" +upgrade op_impl +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op api" +upgrade op_api +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade version.info" +upgrade_file version.info +if [ $? -ne 0 ];then + exit 1 +fi + +config_file=${targetdir}/vendors/config.ini +found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" +found_vendor=$(echo $found_vendors | sed "s/\<$vendor_name\>//g" | tr ',' ' ') +vendor=$(echo $found_vendor | tr -s ' ' ',') +if [ "$vendor" != "" ]; then + sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" +fi + +changemode() +{ + if [ -d ${targetdir} ];then + chmod -R 550 ${targetdir}>/dev/null 2>&1 + fi + + return 0 +} +echo "[ops_custom]changemode..." +#changemode +if [ $? -ne 0 ];then + exit 1 +fi + +echo "SUCCESS" +exit 0 + diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt new file mode 100644 index 000000000..c89578fcd --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt @@ -0,0 +1,21 @@ +# Copyright 2024 Tencent Inc. All rights reserved. +# +# ============================================================================== +cmake_minimum_required(VERSION 3.14) +set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") +file(GLOB_RECURSE atb_plugin_operations_SRCS "./*.cpp") +set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") +message(STATUS "atb_plugin_operations_SRCS: ${atb_plugin_operations_SRCS}") +include_directories( + $ENV{ATB_HOME_PATH}/include + $ENV{ASCEND_HOME_PATH}/include + ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 + ${CMAKE_CURRENT_SOURCE_DIR} + ) +link_directories( + $ENV{ATB_HOME_PATH}/lib + $ENV{ASCEND_HOME_PATH}/lib64 + ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 + ) +add_library(atb_plugin_operations STATIC ${atb_plugin_operations_SRCS}) +target_link_libraries(atb_plugin_operations PRIVATE ${ACL_SHARED_LIBS} ${ACL_SHARED_LIBS} ascendcl) \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp b/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp new file mode 100644 index 000000000..468b84aff --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp @@ -0,0 +1,142 @@ +#include "aclnn_add_operation.h" +#include "aclnn_add_custom.h" + +AddOperation::AddOperation(const std::string &name, AddAttrParam param){ + attrParam = param; + opName_ = name; +} + +atb::SVector GetCopyTensorStride(atb::Dims &tensorDims) +{ + atb::SVector tmpStrides(tensorDims.dimNum, 1); + if (tensorDims.dimNum > 8) { // 8: tensor最大维度数量 + printf("tensor's dimNum is larger than 8, GetCopyTensorStride failed."); + return tmpStrides; + } + for (int64_t i = static_cast(tensorDims.dimNum) - 2; i >= 0; i--) { + tmpStrides[i] = (tensorDims.dims[i + 1] * tmpStrides[i + 1]); + } + return tmpStrides; +} + +std::shared_ptr AddOperation::CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx) +{ + auto aclnnTensor = std::make_shared(); + aclnnTensor->tensorIdx = static_cast(tensorIdx); + aclnnTensor->needUpdateTensorDataPtr = true; + aclnnTensor->atbTensor = atbTensor; + aclnnTensor->strides = GetCopyTensorStride(atbTensor.desc.shape); + + // 创建Aclnn tensor + aclnnTensor->tensor = aclCreateTensor(atbTensor.desc.shape.dims, + atbTensor.desc.shape.dimNum, + atbTensor.desc.dtype, + aclnnTensor->strides.data(), + 0, + atbTensor.desc.format, + atbTensor.desc.shape.dims, + atbTensor.desc.shape.dimNum, + atbTensor.deviceData); + return aclnnTensor; +} + +atb::Status AddOperation::UpdateAclnnVariantPack(const atb::VariantPack &variantPack) +{ + // 更新inTensor的device地址 + for (size_t i = 0; i < aclInTensors_.size(); ++i) { + int ret = -1; + if (!aclInTensors_[i]->needUpdateTensorDataPtr) { + continue; + } + aclInTensors_[i]->atbTensor = variantPack.inTensors.at(i); + ret = aclSetInputTensorAddr(aclExecutor_, + aclInTensors_[i]->tensorIdx, + aclInTensors_[i]->tensor, + aclInTensors_[i]->atbTensor.deviceData); + + if (ret != 0) { + printf("set input fail"); + return atb::ERROR_CANN_ERROR; + } + } + + // 更新outTensor的device地址 + for (size_t i = 0; i < aclOutTensors_.size(); ++i) { + int ret = -1; + if (!aclOutTensors_[i]->needUpdateTensorDataPtr) { + continue; + } + aclOutTensors_[i]->atbTensor = variantPack.outTensors.at(i); + ret = aclSetOutputTensorAddr(aclExecutor_, + aclOutTensors_[i]->tensorIdx, + aclOutTensors_[i]->tensor, + aclOutTensors_[i]->atbTensor.deviceData); + + if (ret != 0) { + printf("set output fail"); + return atb::ERROR_CANN_ERROR; + } + } + return atb::NO_ERROR; +} + +atb::Status AddOperation::Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) { + + aclInTensors_.resize(GetInputNum()); + for (size_t i = 0; i < aclInTensors_.size(); ++i) { + auto aclnnTensor = CreateAclnnTensor(variantPack.inTensors.at(i), i); + if (aclnnTensor->tensor == nullptr) { + printf("creat input tensor %ld fail", i); + return atb::ERROR_INTERNAL_ERROR; + } + aclInTensors_[i] = aclnnTensor; + } + + aclOutTensors_.resize(GetOutputNum()); + for (size_t i = 0; i < aclOutTensors_.size(); ++i) { + auto aclnnTensor = CreateAclnnTensor(variantPack.outTensors.at(i), i); + if (aclnnTensor->tensor == nullptr) { + printf("creat output tensor %ld fail", i); + return atb::ERROR_INTERNAL_ERROR; + } + aclOutTensors_[i] = aclnnTensor; + } + + + auto ret = aclnnAddCustomGetWorkspaceSize(aclInTensors_.at(0)->tensor, + aclInTensors_.at(1)->tensor, + aclOutTensors_.at(0)->tensor, + &workspaceSize_, + &aclExecutor_); + + workspaceSize = workspaceSize_; + return ret; + +} + +atb::Status AddOperation::Execute(const atb::VariantPack &variantPack, uint8_t *workspace, uint64_t workspaceSize, atb::Context *context) { + + + + aclrtStream stream = context->GetExecuteStream(); + if (!stream) { + printf("get stream fail"); + return atb::ERROR_INVALID_PARAM; + } + // 更新数据传入的地址 + int ret = UpdateAclnnVariantPack(variantPack); + if (ret != 0) { + printf("UpdateAclnnVariantPack fail"); + return atb::ERROR_CANN_ERROR; + } + ret = aclnnAddCustom(workspace, workspaceSize_, aclExecutor_, stream); + + return ret; +} + +atb::Status AddOperation::InferShape( + const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const +{ + outTensorDesc.at(0) = inTensorDesc.at(0); + return atb::NO_ERROR; +} diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h b/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h new file mode 100644 index 000000000..8670088d3 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h @@ -0,0 +1,56 @@ +#include +#include +#include +#include +#include +#include "atb/infer_op_params.h" + + +struct AddAttrParam +{ + // add没属性,此处空 +}; + +struct AclnnTensor +{ +public: + atb::Tensor atbTensor; // + aclTensor *tensor = nullptr; + int tensorIdx = -1; // aclTensor在aclExecutor中的index + bool needUpdateTensorDataPtr = false; + atb::SVector strides = {}; +}; + +class AddOperation: public atb::Operation{ +public: + AddOperation(const std::string &name, AddAttrParam param); + atb::Status Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) override; + atb::Status Execute(const atb::VariantPack &variantPack, uint8_t *workspace, + uint64_t workspaceSize, atb::Context *context) override; + atb::Status InferShape( + const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const; + std::shared_ptr CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx); + atb::Status UpdateAclnnVariantPack(const atb::VariantPack &variantPack); + + uint32_t GetInputNum() const + { + return 2; // 算子入参个数 + } + + uint32_t GetOutputNum() const + { + return 1; // 算子出参个数 + } + std::string GetName() const + { + return opName_; + } + + aclOpExecutor *aclExecutor_ = nullptr; + AddAttrParam attrParam; + std::string opName_; + uint64_t workspaceSize_; + + atb::SVector> aclInTensors_; + atb::SVector> aclOutTensors_; +}; \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/build.sh b/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/build.sh new file mode 100755 index 000000000..ba28cc143 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/build.sh @@ -0,0 +1,33 @@ +#!/bin/bash + +# 定义构建目录 +BUILD_DIR="build" + +# 创建构建目录 +mkdir -p "$BUILD_DIR" +cd "$BUILD_DIR" + +# 运行 CMake 配置和编译 +cmake .. +make + +# 查找生成的 .a 文件 +A_FILE=$(find . -name "*.a" -type f) + +# 检查是否找到了 .a 文件 +if [ -z "$A_FILE" ]; then + echo "未找到 .a 文件,编译可能失败。" + exit 1 +fi + +# 复制头文件到 /usr/include +HEADER_FILES=$(find .. -name "*.h" -type f) +for header in $HEADER_FILES; do + cp "$header" /usr/include/ +done + +# 复制 .a 文件到 /usr/local/lib +cp "$A_FILE" /usr/local/lib/ + +echo "构建完成,头文件和 .a 文件已复制到目标目录。" + diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/CMakeLists.txt new file mode 100644 index 000000000..b9c2cec46 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/CMakeLists.txt @@ -0,0 +1,40 @@ +cmake_minimum_required(VERSION 3.16) +project("test_model") +option(USE_CXX11_ABI "USE_CXX11_ABI" OFF) +set(CMAKE_BUILD_TYPE Debug) +if(USE_CXX11_ABI) + set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=1") +else() + set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") +endif() + +set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -std=c++17") + +set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") + +include_directories( + $ENV{ATB_HOME_PATH}/include + $ENV{ASCEND_HOME_PATH}/include + ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 + ${CMAKE_CURRENT_SOURCE_DIR} + ./src/ + ) + +message(".h path = ${CUST_PKG_PATH}/include") +link_directories( + $ENV{ATB_HOME_PATH}/lib + $ENV{ASCEND_HOME_PATH}/lib64 + ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 + ) + + +link_libraries(atb_plugin_operations) +add_executable(test_model ./src/main.cpp) + +# 列出所有的头文件目录 +target_link_libraries(test_model PRIVATE atb ascendcl cust_opapi opapi nnopbase pthread) # 添加自定义算子so及适配so + + + + + diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/gen_data.py b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/gen_data.py new file mode 100644 index 000000000..83ecafbce --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/gen_data.py @@ -0,0 +1,20 @@ +import torch +import os +def gen_golden_data_simple(): + dtype = "float" + input1 = torch.randn(133, 4095, dtype=torch.float) + input2 = torch.randn(133, 4095, dtype=torch.float) + + + golden = input1 + input2 + os.system("mkdir -p input") + os.system("mkdir -p output") + input1.numpy().tofile('./script/input/input0.bin') + input2.numpy().tofile('./script/input/input1.bin') + golden.numpy().tofile("./script/output/golden0.bin") + + with open("./output/meta", "w") as fp: + print(dtype, file=fp) + +if __name__ == "__main__": + gen_golden_data_simple() diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/run.sh b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/run.sh new file mode 100644 index 000000000..450b717f2 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/run.sh @@ -0,0 +1,52 @@ +#!/bin/bash + + +# 步骤1: 运行gen_data.py生成输入bin文件和golden标杆输出数据 +echo "正在生成输入数据和golden标杆数据..." +mkdir -p script/input +mkdir -p script/output +python3 script/gen_data.py +if [ $? -ne 0 ]; then + echo "生成数据失败,脚本终止。" + exit 1 +fi + +# 步骤2: 创建构建目录并进入 +mkdir -p build +cd build +if [ $? -ne 0 ]; then + echo "无法进入构建目录,脚本终止。" + exit 1 +fi + +# 步骤3: 使用CMake配置项目 +echo "正在配置CMake项目..." +cmake .. +if [ $? -ne 0 ]; then + echo "CMake配置失败,脚本终止。" + exit 1 +fi + +# 步骤4: 编译代码 +echo "正在编译代码..." +make +if [ $? -ne 0 ]; then + echo "编译失败,脚本终止。" + exit 1 +fi + +mv test_model ../ +cd .. + +# 步骤5: 运行可执行文件生成实际输出文件 +echo "正在运行可执行文件生成实际输出..." +./test_model +if [ $? -ne 0 ]; then + echo "运行可执行文件失败,脚本终止。" + exit 1 +fi + +# 步骤6: 调用verify_result.py进行golden标杆数据和实际输出数据的比对 +echo "正在验证结果..." +python3 script/verify_result.py script/output/output_0.bin script/output/golden0.bin + diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/verify_result.py b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/verify_result.py new file mode 100644 index 000000000..629845e13 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/verify_result.py @@ -0,0 +1,28 @@ +import os +import sys +import numpy + +loss = 1e-3 # 容忍偏差,一般fp16要求绝对误差和相对误差均不超过千分之一 +minimum = 10e-10 + +def verify_result(real_result, golden): + with open("output/meta", "r") as fp: + dtype_str = fp.readline().strip() + dtype = eval(dtype_str) + real_result = numpy.fromfile(real_result, dtype=dtype) # 从bin文件读取实际运算结果 + golden = numpy.fromfile(golden, dtype=dtype) # 从bin文件读取预期运算结果 + print("=" * 50, real_result[:5], golden[:5], "=" * 50, sep='\n', end='\n', file=sys.stderr) + result = numpy.abs(real_result - golden) # 计算运算结果和预期结果偏差 + deno = numpy.maximum(numpy.abs(real_result), numpy.abs(golden)) # 获取最大值并组成新数组 + result_atol = numpy.less_equal(result, loss) # 计算绝对误差 + result_rtol = numpy.less_equal(result / numpy.add(deno, minimum), loss) # 计算相对误差 + if not result_rtol.all() and not result_atol.all(): + if numpy.sum(result_rtol == False) > real_result.size * loss and numpy.sum(result_atol == False) > real_result.size * loss: # 误差超出预期时返回打印错误,返回对比失败 + print("[ERROR] result error") + return False + print("test Operation success!") + return True + + +if __name__ == '__main__': + verify_result(sys.argv[1],sys.argv[2]) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.cpp b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.cpp new file mode 100644 index 000000000..258c0cb3c --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.cpp @@ -0,0 +1,217 @@ +/* + * Copyright (c) Huawei Technologies Co., Ltd. 2023. All rights reserved. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +#include "main.h" + +bool SetInputData(std::vector &inputData){ + char *xPath = "./script/input/input0.bin"; + char *yPath = "./script/input/input1.bin"; + InputData inputX; + InputData inputY; + inputX.data = ReadBinFile(xPath,inputX.size); + inputY.data = ReadBinFile(yPath,inputY.size); + inputData.push_back(inputX); + inputData.push_back(inputY); + return true; +} + +bool SetOperationInputDesc(atb::SVector &intensorDescs){ + atb::TensorDesc xDesc; + xDesc.dtype = ACL_FLOAT; + xDesc.format = ACL_FORMAT_ND; + xDesc.shape.dimNum = 2; + xDesc.shape.dims[0] = 133; + xDesc.shape.dims[1] = 4095; + + atb::TensorDesc yDesc; + yDesc.dtype = ACL_FLOAT; + yDesc.format = ACL_FORMAT_ND; + yDesc.shape.dimNum = 2; + yDesc.shape.dims[0] = 133; + yDesc.shape.dims[1] = 4095; + + intensorDescs.at(0) = xDesc; + intensorDescs.at(1) = yDesc; +} + + + +static void SetCurrentDevice() +{ + const int deviceId = 0; + std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; + int ret = aclrtSetDevice(deviceId); + if (ret != 0) { + std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; + return; + } + std::cout << "[INFO]: aclrtSetDevice success" << std::endl; +} + + +static void FreeTensor(atb::Tensor &tensor) +{ + if (tensor.deviceData) { + int ret = aclrtFree(tensor.deviceData); + if (ret != 0) { + std::cout << "[ERROR]: aclrtFree fail" << std::endl; + } + tensor.deviceData = nullptr; + tensor.dataSize = 0; + } + if (tensor.hostData) { + int ret = aclrtFreeHost(tensor.hostData); + if (ret != 0) { + std::cout << "[ERROR]: aclrtFreeHost fail, ret = " << ret << std::endl; + } + tensor.hostData = nullptr; + tensor.dataSize = 0; + } +} + +static void FreeTensors(atb::SVector &inTensors, atb::SVector &outTensors) +{ + for (size_t i = 0; i < inTensors.size(); ++i) { + FreeTensor(inTensors.at(i)); + } + for (size_t i = 0; i < outTensors.size(); ++i) { + FreeTensor(outTensors.at(i)); + } +} +bool SaveMemoryToBinFile(void* memoryAddress, size_t memorySize, size_t i) { + // 创建 output 目录(如果不存在) + std::filesystem::create_directories("output"); + + // 生成文件名 + std::string filename = "script/output/output_" + std::to_string(i) + ".bin"; + + // 打开文件以二进制写入模式 + std::ofstream file(filename, std::ios::binary); + if (!file) { + std::cerr << "无法打开文件: " << filename << std::endl; + return false; + } + + // 写入数据 + file.write(static_cast(memoryAddress), memorySize); + if (!file) { + std::cerr << "写入文件时出错: " << filename << std::endl; + file.close(); + return false; + } + + // 关闭文件 + file.close(); + std::cout << "数据已成功保存到: " << filename << std::endl; + return true; +} + +int main(int argc, const char *argv[]) +{ + const int deviceId = 0; + std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; + int ret = aclrtSetDevice(deviceId); + if (ret != 0) { + std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; + return 1; + } + std::cout << "[INFO]: aclrtSetDevice success" << std::endl; + atb::Context *context = nullptr; + ret = atb::CreateContext(&context); + void *stream = nullptr; + ret = aclrtCreateStream(&stream); + if (ret != 0) { + std::cout << "[ERROR]: AsdRtStreamCreate fail, ret:" << ret << std::endl; + return 1; + } + context->SetExecuteStream(stream); + + std::vector input; + SetInputData(input); + + AddAttrParam addAttrParam; + AddOperation *op = new AddOperation("Add",addAttrParam); + std::cout << "[INFO]: complete CreateOp!" << std::endl; + + if(input.size() != op->GetInputNum()) std::cout << "[ERROR]: Operation actual input num is not equal to GetInputNum()"; + + atb::SVector intensorDescs; + atb::SVector outtensorDescs; + intensorDescs.resize(op->GetInputNum()); + outtensorDescs.resize(op->GetOutputNum()); + SetOperationInputDesc(intensorDescs); + atb::Status st = op->InferShape(intensorDescs,outtensorDescs); + if (st != 0) { + std::cout << "[ERROR]: Operation InferShape fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation InferShape success" << std::endl; + atb::VariantPack variantPack; + variantPack.inTensors.resize(op->GetInputNum()); + variantPack.outTensors.resize(op->GetOutputNum()); + for(size_t i=0;iGetInputNum();i++){ + variantPack.inTensors.at(i).desc = intensorDescs.at(i); + variantPack.inTensors.at(i).hostData = input[i].data; + variantPack.inTensors.at(i).dataSize = input[i].size; + CheckAcl(aclrtMalloc(&variantPack.inTensors.at(i).deviceData, input[i].size, ACL_MEM_MALLOC_HUGE_FIRST)); + CheckAcl(aclrtMemcpy(variantPack.inTensors.at(i).deviceData, input[i].size, input[i].data, input[i].size, ACL_MEMCPY_HOST_TO_DEVICE)); + } + std::cout << "[INFO]: Operation Input prepare sucess" << std::endl; + for(size_t i=0;iGetOutputNum();i++){ + int64_t *dims = new int64_t[outtensorDescs.at(i).shape.dimNum]; + for(size_t j=0;jSetup(variantPack, workspaceSize, context); + if (st != 0) { + std::cout << "[ERROR]: Operation setup fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation setup success" << std::endl; + void *workspace = nullptr; + if (workspaceSize > 0) { + aclrtMalloc(&workspace, workspaceSize, ACL_MEM_MALLOC_HUGE_FIRST); + } + + std::cout << "[INFO]: Operation execute start" << std::endl; + st = op->Execute(variantPack, (uint8_t*)workspace, workspaceSize, context); + if (st != 0) { + std::cout << "[ERROR]: Operation execute fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation execute success" << std::endl; + for(size_t i = 0; i < op->GetOutputNum(); i++){ + CheckAcl(aclrtMemcpy(variantPack.outTensors.at(i).hostData, variantPack.outTensors.at(i).dataSize, variantPack.outTensors.at(0).deviceData, + variantPack.outTensors.at(i).dataSize, ACL_MEMCPY_DEVICE_TO_HOST)); + SaveMemoryToBinFile(variantPack.outTensors.at(i).hostData,variantPack.outTensors.at(i).dataSize,i); + } + + FreeTensors(variantPack.inTensors, variantPack.outTensors); + st = atb::DestroyContext(context); + CheckAcl(aclrtDestroyStream(stream)); + CheckAcl(aclrtResetDevice(0)); + CheckAcl(aclFinalize()); + return atb::ErrorType::NO_ERROR; +} diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.h b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.h new file mode 100644 index 000000000..44a94e9ec --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.h @@ -0,0 +1,55 @@ +#include +#include +#include +#include +#include +#include +#include +#include "securec.h" +#include "atb/atb_infer.h" +#include "aclnn_add_operation.h" +#include + +struct InputData{ + void* data; + uint64_t size; +}; +aclError CheckAcl(aclError ret) +{ + if (ret != ACL_ERROR_NONE) { + std::cerr << __FILE__ << ":" << __LINE__ << " aclError:" << ret << std::endl; + } + return ret; +} +void* ReadBinFile(const char* filename, size_t& size) { + std::ifstream file(filename, std::ios::binary | std::ios::ate); + if (!file) { + std::cerr << "无法打开文件: " << filename << std::endl; + return nullptr; + } + + // 获取文件大小 + size = file.tellg(); + file.seekg(0, std::ios::beg); + + // 分配内存 + void* buffer; + int ret = aclrtMallocHost(&buffer,size); + if (!buffer) { + std::cerr << "内存分配失败" << std::endl; + file.close(); + return nullptr; + } + + // 读取文件内容到内存 + file.read(static_cast(buffer), size); + if (!file) { + std::cerr << "读取文件失败" << std::endl; + delete[] static_cast(buffer); + file.close(); + return nullptr; + } + + file.close(); + return buffer; +} \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AtbAddSample/readme.md b/atb_operator/AtbAddSample/AtbAddSample/readme.md new file mode 100644 index 000000000..d205e5429 --- /dev/null +++ b/atb_operator/AtbAddSample/AtbAddSample/readme.md @@ -0,0 +1,81 @@ +### 概述 + +本样例基于AscendC自定义Add算子,开发了ATB插件并进行了插件调用测试. + +### 项目结构介绍 + +├── AddAscendC //Add AscendC自定义算子工程 + +├── AddOperationATBPlugin //AddOperation ATB插件代码 + +├── AddOperationTest //AddOperation 测试代码 + +### 样例运行 + +#### Add AscendC自定义算子部署 + +- 运行生成算子run包 + + ```shell + cd AddAscendC + bash build.sh + ``` + +- 算子run包部署 + + ```shell + cd AddAscendC/build_out + ./custom_opp_hce_aarch64.run + ``` + +#### AddOperation ATB插件部署 + +- 运行编译脚本完成部署(脚本会生成静态库.a文件,同时将头文件拷贝到/usr/include,.a文件拷贝到/usr/local/lib下) + + ``` + cd AddOperationATBPlugin + bash build.sh + ``` + +#### AddOperation测试 + +- 运行脚本完成算子测试 + + ```shell + cd AddOperationTest + bash run.sh + ``` + +### AddOperation算子介绍 + +#### 功能 + +实现两个输入张量相加 + +#### 定义 + +``` +struct AddParam { +}; +``` + +#### 参数列表 + +该算子参数为空 + +#### 输入 + +| **参数** | **维度** | **数据类型** | **格式** | 描述 | +| -------- | -------------------------- | --------------------- | -------- | ---------- | +| x | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | +| y | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | + +#### 输出 + +| **参数** | **维度** | **数据类型** | **格式** | 描述 | +| -------- | -------------------------- | --------------------- | -------- | ---------------------------------------- | +| output | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输出tensor。数据类型和shape与x保持一致。 | + +#### 规格约束 + +暂无 \ No newline at end of file -- Gitee From e9a0f967079df094ac0b30372090a2f9e7952bf0 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E5=88=98=E5=96=9C=E5=BC=BA?= Date: Mon, 31 Mar 2025 06:22:05 +0000 Subject: [PATCH 03/26] update --- .../AtbAddSample/AddAscendC/CMakeLists.txt | 69 ++ .../AtbAddSample/AddAscendC/CMakePresets.json | 63 ++ .../AtbAddSample/AddAscendC/README.md | 240 +++++ atb_operator/AtbAddSample/AddAscendC/build.sh | 37 + .../AddAscendC/cmake/config.cmake | 25 + .../AtbAddSample/AddAscendC/cmake/func.cmake | 228 +++++ .../AtbAddSample/AddAscendC/cmake/intf.cmake | 26 + .../AddAscendC/cmake/makeself.cmake | 17 + .../AddAscendC/cmake/util/__init__.py | 8 + .../cmake/util/ascendc_bin_param_build.py | 129 +++ .../cmake/util/ascendc_impl_build.py | 446 ++++++++++ .../cmake/util/ascendc_ops_config.py | 114 +++ .../cmake/util/ascendc_replay_build.py | 65 ++ .../cmake/util/batch_replay_impl.temp | 117 +++ .../cmake/util/code_channel_infer.py | 115 +++ .../AddAscendC/cmake/util/const_var.py | 33 + .../cmake/util/gen_impl_and_mrege_json.sh | 57 ++ .../AddAscendC/cmake/util/gen_ops_filter.sh | 62 ++ .../AddAscendC/cmake/util/gen_version_info.sh | 6 + .../AddAscendC/cmake/util/insert_op_info.py | 36 + .../cmake/util/insert_simplified_keys.py | 248 ++++++ .../AddAscendC/cmake/util/kernel_entry.py | 115 +++ .../AddAscendC/cmake/util/kernel_impl.temp | 10 + .../AddAscendC/cmake/util/makeself/COPYING | 339 ++++++++ .../AddAscendC/cmake/util/makeself/README.md | 246 ++++++ .../AddAscendC/cmake/util/makeself/VERSION | 1 + .../cmake/util/makeself/make-release.sh | 9 + .../cmake/util/makeself/makeself-header.sh | 660 ++++++++++++++ .../AddAscendC/cmake/util/makeself/makeself.1 | 110 +++ .../cmake/util/makeself/makeself.lsm | 16 + .../cmake/util/makeself/makeself.sh | 822 ++++++++++++++++++ .../cmake/util/makeself/run-tests.sh | 8 + .../cmake/util/merge_aicpu_info_json.sh | 31 + .../AddAscendC/cmake/util/opdesc_parser.py | 260 ++++++ .../cmake/util/parse_ini_to_json.py | 338 +++++++ .../AddAscendC/cmake/util/preset_parse.py | 23 + .../AddAscendC/cmake/util/replay_codegen.py | 105 +++ .../AddAscendC/cmake/util/replay_impl.temp | 120 +++ .../cmake/util/tiling_data_def_build.py | 84 ++ .../AddAscendC/framework/CMakeLists.txt | 11 + .../framework/tf_plugin/CMakeLists.txt | 14 + .../tf_plugin/tensorflow_add_custom_plugin.cc | 23 + .../AddAscendC/op_host/CMakeLists.txt | 82 ++ .../AddAscendC/op_host/add_custom.cpp | 117 +++ .../AddAscendC/op_host/add_custom_tiling.h | 22 + .../AddAscendC/op_kernel/CMakeLists.txt | 61 ++ .../AddAscendC/op_kernel/add_custom.cpp | 134 +++ .../AddAscendC/scripts/install.sh | 318 +++++++ .../AddAscendC/scripts/upgrade.sh | 151 ++++ .../AddOperationATBPlugin/CMakeLists.txt | 21 + .../aclnn_add_operation.cpp | 142 +++ .../aclnn_add_operation.h | 56 ++ .../AddOperationATBPlugin/build.sh | 33 + .../AddOperationTest/CMakeLists.txt | 40 + .../AddOperationTest/script/gen_data.py | 20 + .../AddOperationTest/script/run.sh | 52 ++ .../AddOperationTest/script/verify_result.py | 28 + .../AddOperationTest/src/main.cpp | 217 +++++ .../AtbAddSample/AddOperationTest/src/main.h | 55 ++ atb_operator/AtbAddSample/readme.md | 81 ++ 60 files changed, 7116 insertions(+) create mode 100644 atb_operator/AtbAddSample/AddAscendC/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AddAscendC/CMakePresets.json create mode 100644 atb_operator/AtbAddSample/AddAscendC/README.md create mode 100755 atb_operator/AtbAddSample/AddAscendC/build.sh create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/config.cmake create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/func.cmake create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/intf.cmake create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/makeself.cmake create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/__init__.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/const_var.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp create mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING create mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/README.md create mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION create mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh create mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh create mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 create mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh create mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/preset_parse.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp create mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py create mode 100644 atb_operator/AtbAddSample/AddAscendC/framework/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc create mode 100644 atb_operator/AtbAddSample/AddAscendC/op_host/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AddAscendC/op_host/add_custom.cpp create mode 100644 atb_operator/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h create mode 100644 atb_operator/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp create mode 100644 atb_operator/AtbAddSample/AddAscendC/scripts/install.sh create mode 100644 atb_operator/AtbAddSample/AddAscendC/scripts/upgrade.sh create mode 100644 atb_operator/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp create mode 100644 atb_operator/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h create mode 100755 atb_operator/AtbAddSample/AddOperationATBPlugin/build.sh create mode 100644 atb_operator/AtbAddSample/AddOperationTest/CMakeLists.txt create mode 100644 atb_operator/AtbAddSample/AddOperationTest/script/gen_data.py create mode 100644 atb_operator/AtbAddSample/AddOperationTest/script/run.sh create mode 100644 atb_operator/AtbAddSample/AddOperationTest/script/verify_result.py create mode 100644 atb_operator/AtbAddSample/AddOperationTest/src/main.cpp create mode 100644 atb_operator/AtbAddSample/AddOperationTest/src/main.h diff --git a/atb_operator/AtbAddSample/AddAscendC/CMakeLists.txt b/atb_operator/AtbAddSample/AddAscendC/CMakeLists.txt new file mode 100644 index 000000000..584132d80 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/CMakeLists.txt @@ -0,0 +1,69 @@ +cmake_minimum_required(VERSION 3.16.0) +project(opp) +if(ENABLE_CROSS_COMPILE) + if(${CMAKE_SYSTEM_PROCESSOR} STREQUAL x86_64) + set(CROSS_COMPILE_PLATFORM aarch64) + else() + set(CROSS_COMPILE_PLATFORM x86_64) + endif() + set(PLATFORM ${CMAKE_SYSTEM_PROCESSOR}) + set(CMAKE_COMPILE_COMPILER_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/linux/${CROSS_COMPILE_PLATFORM}/) + set(CMAKE_COMPILE_RUNTIME_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/${CROSS_COMPILE_PLATFORM}/) + set(CMAKE_SYSTEM_PROCESSOR ${CROSS_COMPILE_PLATFORM}) + set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) + set(CMAKE_CXX_COMPILER ${CMAKE_CROSS_PLATFORM_COMPILER}) +else() + set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) +endif() + +include(cmake/config.cmake) +include(cmake/func.cmake) +include(cmake/intf.cmake) + +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/framework) + add_subdirectory(framework) +endif() +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_host) + add_subdirectory(op_host) +endif() +if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_kernel) + add_subdirectory(op_kernel) +endif() +if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) + add_subdirectory(testcases) +endif() + +# modify vendor_name in install.sh and upgrade.sh +add_custom_command(OUTPUT ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh + COMMAND mkdir -p ${CMAKE_BINARY_DIR}/scripts + COMMAND cp -r ${CMAKE_SOURCE_DIR}/scripts/* ${CMAKE_BINARY_DIR}/scripts/ + COMMAND sed -i "s/vendor_name=customize/vendor_name=${vendor_name}/g" ${CMAKE_BINARY_DIR}/scripts/* +) +add_custom_target(modify_vendor ALL DEPENDS ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh) +install(DIRECTORY ${CMAKE_BINARY_DIR}/scripts/ DESTINATION . FILE_PERMISSIONS OWNER_EXECUTE OWNER_READ GROUP_READ) + +install(FILES ${CMAKE_SOURCE_DIR}/custom.proto DESTINATION packages OPTIONAL) + +get_system_info(SYSTEM_INFO) + +# gen version.info +add_custom_target(gen_version_info ALL + COMMAND bash ${CMAKE_CURRENT_SOURCE_DIR}/cmake/util/gen_version_info.sh ${ASCEND_CANN_PACKAGE_PATH} ${CMAKE_CURRENT_BINARY_DIR} +) + +install(FILES ${CMAKE_CURRENT_BINARY_DIR}/version.info + DESTINATION packages/vendors/${vendor_name}/) + +# CPack config +set(CPACK_PACKAGE_NAME ${CMAKE_PROJECT_NAME}) +set(CPACK_PACKAGE_VERSION ${CMAKE_PROJECT_VERSION}) +set(CPACK_PACKAGE_DESCRIPTION "CPack opp project") +set(CPACK_PACKAGE_DESCRIPTION_SUMMARY "CPack opp project") +set(CPACK_PACKAGE_DIRECTORY ${CMAKE_INSTALL_PREFIX}) +set(CPACK_PACKAGE_FILE_NAME "custom_opp_${SYSTEM_INFO}.run") +set(CPACK_GENERATOR External) +set(CPACK_CMAKE_GENERATOR "Unix Makefiles") +set(CPACK_EXTERNAL_ENABLE_STAGING TRUE) +set(CPACK_EXTERNAL_PACKAGE_SCRIPT ${CMAKE_SOURCE_DIR}/cmake/makeself.cmake) +set(CPACK_EXTERNAL_BUILT_PACKAGES ${CPACK_PACKAGE_DIRECTORY}/_CPack_Packages/Linux/External/${CPACK_PACKAGE_FILE_NAME}/${CPACK_PACKAGE_FILE_NAME}) +include(CPack) diff --git a/atb_operator/AtbAddSample/AddAscendC/CMakePresets.json b/atb_operator/AtbAddSample/AddAscendC/CMakePresets.json new file mode 100644 index 000000000..ddec9431b --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/CMakePresets.json @@ -0,0 +1,63 @@ +{ + "version": 1, + "cmakeMinimumRequired": { + "major": 3, + "minor": 19, + "patch": 0 + }, + "configurePresets": [ + { + "name": "default", + "displayName": "Default Config", + "description": "Default build using Unix Makefiles generator", + "generator": "Unix Makefiles", + "binaryDir": "${sourceDir}/build_out", + "cacheVariables": { + "CMAKE_BUILD_TYPE": { + "type": "STRING", + "value": "Release" + }, + "ENABLE_SOURCE_PACKAGE": { + "type": "BOOL", + "value": "True" + }, + "ENABLE_BINARY_PACKAGE": { + "type": "BOOL", + "value": "True" + }, + "ASCEND_COMPUTE_UNIT": { + "type": "STRING", + "value": "ascend310b;ascend910b" + }, + "ENABLE_TEST": { + "type": "BOOL", + "value": "True" + }, + "vendor_name": { + "type": "STRING", + "value": "customize" + }, + "ASCEND_CANN_PACKAGE_PATH": { + "type": "PATH", + "value": "/usr/local/Ascend/ascend-toolkit/latest" + }, + "ASCEND_PYTHON_EXECUTABLE": { + "type": "STRING", + "value": "python3" + }, + "CMAKE_INSTALL_PREFIX": { + "type": "PATH", + "value": "${sourceDir}/build_out" + }, + "ENABLE_CROSS_COMPILE": { + "type": "BOOL", + "value": "False" + }, + "CMAKE_CROSS_PLATFORM_COMPILER": { + "type": "PATH", + "value": "/usr/bin/aarch64-linux-gnu-g++" + } + } + } + ] +} diff --git a/atb_operator/AtbAddSample/AddAscendC/README.md b/atb_operator/AtbAddSample/AddAscendC/README.md new file mode 100644 index 000000000..da59a83e7 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/README.md @@ -0,0 +1,240 @@ +<<<<<<< HEAD +## 概述 +本样例基于AddCustom算子工程,介绍了单算子工程及单算子调用。 + +## 算子描述 +======= +# 概述 +本样例基于AddCustom算子工程,介绍了单算子工程及aclnn接口说明。 + +## 算子工程介绍 +### 算子描述 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd +Add算子实现了两个数据相加,返回相加结果的功能,该算子实现了非32B对齐shape下的算子功能,完成了算子的泛化实现。 +对应的数学表达式为: +``` +z = x + y +``` +<<<<<<< HEAD +## 算子规格描述 +======= +### 算子规格描述 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd + + + + + + + + + + + +
算子类型(OpType)Add
算子输入nameshapedata typeformat
x-float32,float16,int32,int8ND
y-float32,float16,int32,int8ND
算子输出z-float32,float16,int32,int8ND
核函数名add_custom
+ + +<<<<<<< HEAD +## 算子工程介绍 +======= +### 算子工程文件介绍 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd +其中,算子工程目录AddCustom包含算子实现的模板文件、编译脚本等,如下所示: +``` +├── AddCustom //Add自定义算子工程 +│ ├── cmake +│ ├── framework // 算子插件实现文件目录,单算子模型文件的生成不依赖算子适配插件,无需关注 +│ ├── op_host // host侧实现文件 +│ ├── op_kernel // kernel侧实现文件 +│ ├── scripts // 自定义算子工程打包相关脚本所在目录 +│ ├── build.sh // 编译入口脚本 +│ ├── CMakeLists.txt // 算子工程的CMakeLists.txt +│ └── CMakePresets.json // 编译配置项 +``` +<<<<<<< HEAD +CANN软件包中提供了工程创建工具msopgen,AddCustom算子工程可通过AddCustom.json自动创建,具体请参考[Ascend C算子开发](https://hiascend.com/document/redirect/CannCommunityOpdevAscendC)>算子开发>算子开发工程>基于自定义算子工程的算子开发>创建算子工程 章节。 +## 编译运行样例算子 +======= + +### 编译运行样例算子 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd +针对自定义算子工程,编译运行包含如下步骤: +- 编译自定义算子工程生成算子安装包; +- 安装自定义算子到算子库中; +- 调用执行自定义算子; + +详细操作如下所示。 +<<<<<<< HEAD +### 1. 获取源码包 +编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 +### 2. 编译算子工程 +======= +#### 1. 获取源码包 +编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 +#### 2. 编译算子工程 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd + 编译自定义算子工程,构建生成自定义算子包。 + + - 执行如下命令,切换到算子工程AddCustom目录。 + + ```bash + cd ${git_clone_path}/samples/atb_operator/AtbAddSample/AddOp + ``` + + - 修改CMakePresets.json中ASCEND_CANN_PACKAGE_PATH为CANN软件包安装后的实际路径。 + + + ```json + { + …… + "configurePresets": [ + { + …… + "ASCEND_CANN_PACKAGE_PATH": { + "type": "PATH", + "value": "/usr/local/Ascend/ascend-toolkit/latest" //请替换为CANN软件包安装后的实际路径。eg:/home/HwHiAiUser/Ascend/ascend-toolkit/latest + }, + …… + } + ] + } + ``` + - 在算子工程AddOp目录下执行如下命令,进行算子工程编译。 + + ```bash + ./build.sh + ``` +编译成功后,会在当前目录下创建build_out目录,并在build_out目录下生成自定义算子安装包custom_opp_\_\.run,例如“custom_opp_ubuntu_x86_64.run”。 + +备注:如果要使用dump调试功能,需要移除op_host内和CMakeLists.txt内的Atlas 200/500 A2 推理产品的配置。 + +<<<<<<< HEAD +### 3. 部署算子包 +======= +#### 3. 部署算子包 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd + +执行如下命令,在自定义算子安装包所在路径下,安装自定义算子包。 + ```bash + cd build_out + ./custom_opp__.run + ``` +命令执行成功后,自定义算子包中的相关文件将部署至当前环境的OPP算子库的vendors/customize目录中。 + +<<<<<<< HEAD +### 4. 配置环境变量 +======= +#### 4. 配置环境变量 +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd + + 请根据当前环境上CANN开发套件包的[安装方式](https://hiascend.com/document/redirect/CannCommunityInstSoftware),选择对应配置环境变量的命令。 + - 默认路径,root用户安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=/usr/local/Ascend/ascend-toolkit/latest + ``` + - 默认路径,非root用户安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=$HOME/Ascend/ascend-toolkit/latest + ``` + - 指定路径install_path,安装CANN软件包 + ```bash + export ASCEND_INSTALL_PATH=${install_path}/ascend-toolkit/latest + ``` + +## aclnn接口说明 + +自定义算子编译部署后,会自动生成单算子API,可以直接在应用程序中调用。算子API的形式一般定义为“两段式接口”,形如: + ```cpp + aclnnStatus aclnnAddCustomGetWorkspaceSize(const aclTensor *x, const aclTensor *y, const alcTensor *out, uint64_t workspaceSize, aclOpExecutor **executor); + aclnnStatus aclnnAddCustom(void *workspace, int64_t workspaceSize, aclOpExecutor **executor, aclrtStream stream); + ``` +其中aclnnAddCustomGetWorkspaceSize为第一段接口,主要用于计算本次API调用计算过程中需要多少的workspace内存。获取到本次API计算需要的workspace大小之后,按照workspaceSize大小申请Device侧内存,然后调用第二段接口aclnnAddCustom执行计算。具体参考[AscendCL单算子调用](https://hiascend.com/document/redirect/CannCommunityAscendCInVorkSingleOp)>单算子API执行 章节。 + +<<<<<<< HEAD +======= +### 功能描述 +* 算子功能:完成加法计算。 +* 计算公式: +``` +z = x + y +``` + +#### 参数说明 +##### aclnnAddCustomGetWorkspaceSize: + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
接口参数描述
aclnnAddCustomGetWorkspaceSizex输入x的Tensor,支持flaot/half/int8/int32类型,ND排布格式
y输入y的Tensor,支持flaot/half/int8/int32类型,ND排布格式
out输出z的Tensor,支持flaot/half/int8/int32类型,ND排布格式
workspaceSize返回需要在Device侧申请的workspace大小
executor返回op执行器,包含了算子计算流程
+ + +| 返回值 | 描述 | +| ---- | ---- | +| ACL_SUCCESS | 成功 | +| ACL_ERROR | 失败 | + +##### aclnnAddCustom + + + + + + + + + + + + + + + + + + + + + + + +
接口参数描述
aclnnAddCustomworkspace输入Device侧申请的workspace内存
workspaceSizeworkspace大小
executor输入op执行器,包含了算子计算流程
stream输入执行流
+ +| 返回值 | 描述 | +| ---- | ---- | +| ACL_SUCCESS | 成功 | +| ACL_ERROR | 失败 | + +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd + +## 更新说明 +| 时间 | 更新事项 | +|----|------| +<<<<<<< HEAD +| 2025/03/27 | 修改不同用户环境配置 | +======= +| 2025/03/27 | 新增README | +>>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd diff --git a/atb_operator/AtbAddSample/AddAscendC/build.sh b/atb_operator/AtbAddSample/AddAscendC/build.sh new file mode 100755 index 000000000..4be96d7d8 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/build.sh @@ -0,0 +1,37 @@ +#!/bin/bash +script_path=$(realpath $(dirname $0)) + + +mkdir -p build_out +rm -rf build_out/* +cd build_out + +cmake_version=$(cmake --version | grep "cmake version" | awk '{print $3}') +if [ "$cmake_version" \< "3.19.0" ] ; then + opts=$(python3 $script_path/cmake/util/preset_parse.py $script_path/CMakePresets.json) + echo $opts + cmake .. $opts +else + cmake .. --preset=default +fi +target=package +if [ "$1"x != ""x ]; then target=$1; fi + +cmake --build . --target $target -j16 +if [ $? -ne 0 ]; then exit 1; fi + +if [ $target = "package" ]; then + if test -d ./op_kernel/binary ; then + ./cust*.run + if [ $? -ne 0 ]; then exit 1; fi + cmake --build . --target binary -j16 + if [ $? -ne 0 ]; then exit 1; fi + cmake --build . --target $target -j16 + fi +fi + +# for debug +# cd build_out +# make +# cpack +# verbose append -v diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/config.cmake b/atb_operator/AtbAddSample/AddAscendC/cmake/config.cmake new file mode 100755 index 000000000..886119daa --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/config.cmake @@ -0,0 +1,25 @@ + +set(CMAKE_CXX_FLAGS_DEBUG "") +set(CMAKE_CXX_FLAGS_RELEASE "") + +if (NOT DEFINED vendor_name) + set(vendor_name customize CACHE STRING "") +endif() +if (NOT DEFINED ASCEND_CANN_PACKAGE_PATH) + set(ASCEND_CANN_PACKAGE_PATH /usr/local/Ascend/latest CACHE PATH "") +endif() +if (NOT DEFINED ASCEND_PYTHON_EXECUTABLE) + set(ASCEND_PYTHON_EXECUTABLE python3 CACHE STRING "") +endif() +if (NOT DEFINED ASCEND_COMPUTE_UNIT) + message(FATAL_ERROR "ASCEND_COMPUTE_UNIT not set in CMakePreset.json ! +") +endif() +set(ASCEND_TENSOR_COMPILER_PATH ${ASCEND_CANN_PACKAGE_PATH}/compiler) +set(ASCEND_CCEC_COMPILER_PATH ${ASCEND_TENSOR_COMPILER_PATH}/ccec_compiler/bin) +set(ASCEND_AUTOGEN_PATH ${CMAKE_BINARY_DIR}/autogen) +set(ASCEND_FRAMEWORK_TYPE tensorflow) +file(MAKE_DIRECTORY ${ASCEND_AUTOGEN_PATH}) +set(CUSTOM_COMPILE_OPTIONS "custom_compile_options.ini") +execute_process(COMMAND rm -rf ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} + COMMAND touch ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS}) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/func.cmake b/atb_operator/AtbAddSample/AddAscendC/cmake/func.cmake new file mode 100755 index 000000000..ad187e7d6 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/func.cmake @@ -0,0 +1,228 @@ + +function(get_system_info SYSTEM_INFO) + if (UNIX) + execute_process(COMMAND grep -i ^id= /etc/os-release OUTPUT_VARIABLE TEMP) + string(REGEX REPLACE "\n|id=|ID=|\"" "" SYSTEM_NAME ${TEMP}) + set(${SYSTEM_INFO} ${SYSTEM_NAME}_${CMAKE_SYSTEM_PROCESSOR} PARENT_SCOPE) + elseif (WIN32) + message(STATUS "System is Windows. Only for pre-build.") + else () + message(FATAL_ERROR "${CMAKE_SYSTEM_NAME} not support.") + endif () +endfunction() + +function(opbuild) + message(STATUS "Opbuild generating sources") + cmake_parse_arguments(OPBUILD "" "OUT_DIR;PROJECT_NAME;ACCESS_PREFIX" "OPS_SRC" ${ARGN}) + execute_process(COMMAND ${CMAKE_COMPILE} -g -fPIC -shared -std=c++11 ${OPBUILD_OPS_SRC} -D_GLIBCXX_USE_CXX11_ABI=0 + -I ${ASCEND_CANN_PACKAGE_PATH}/include -L ${ASCEND_CANN_PACKAGE_PATH}/lib64 -lexe_graph -lregister -ltiling_api + -o ${OPBUILD_OUT_DIR}/libascend_all_ops.so + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("build ops lib info: ${EXEC_INFO}") + message("build ops lib error: ${EXEC_ERROR}") + message(FATAL_ERROR "opbuild run failed!") + endif() + set(proj_env "") + set(prefix_env "") + if (NOT "${OPBUILD_PROJECT_NAME}x" STREQUAL "x") + set(proj_env "OPS_PROJECT_NAME=${OPBUILD_PROJECT_NAME}") + endif() + if (NOT "${OPBUILD_ACCESS_PREFIX}x" STREQUAL "x") + set(prefix_env "OPS_DIRECT_ACCESS_PREFIX=${OPBUILD_ACCESS_PREFIX}") + endif() + execute_process(COMMAND ${proj_env} ${prefix_env} ${ASCEND_CANN_PACKAGE_PATH}/toolkit/tools/opbuild/op_build + ${OPBUILD_OUT_DIR}/libascend_all_ops.so ${OPBUILD_OUT_DIR} + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("opbuild ops info: ${EXEC_INFO}") + message("opbuild ops error: ${EXEC_ERROR}") + endif() + message(STATUS "Opbuild generating sources - done") +endfunction() + +function(add_ops_info_target) + cmake_parse_arguments(OPINFO "" "TARGET;OPS_INFO;OUTPUT;INSTALL_DIR" "" ${ARGN}) + get_filename_component(opinfo_file_path "${OPINFO_OUTPUT}" DIRECTORY) + add_custom_command(OUTPUT ${OPINFO_OUTPUT} + COMMAND mkdir -p ${opinfo_file_path} + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/parse_ini_to_json.py + ${OPINFO_OPS_INFO} ${OPINFO_OUTPUT} + ) + add_custom_target(${OPINFO_TARGET} ALL + DEPENDS ${OPINFO_OUTPUT} + ) + install(FILES ${OPINFO_OUTPUT} + DESTINATION ${OPINFO_INSTALL_DIR} + ) +endfunction() + +function(add_ops_compile_options OP_TYPE) + cmake_parse_arguments(OP_COMPILE "" "OP_TYPE" "COMPUTE_UNIT;OPTIONS" ${ARGN}) + file(APPEND ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} + "${OP_TYPE},${OP_COMPILE_COMPUTE_UNIT},${OP_COMPILE_OPTIONS}\n") +endfunction() + +function(add_ops_impl_target) + cmake_parse_arguments(OPIMPL "" "TARGET;OPS_INFO;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) + add_custom_command(OUTPUT ${OPIMPL_OUT_DIR}/.impl_timestamp + COMMAND mkdir -m 700 -p ${OPIMPL_OUT_DIR}/dynamic + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py + ${OPIMPL_OPS_INFO} + \"${OPIMPL_OPS_BATCH}\" \"${OPIMPL_OPS_ITERATE}\" + ${OPIMPL_IMPL_DIR} + ${OPIMPL_OUT_DIR}/dynamic + ${ASCEND_AUTOGEN_PATH} + + COMMAND rm -rf ${OPIMPL_OUT_DIR}/.impl_timestamp + COMMAND touch ${OPIMPL_OUT_DIR}/.impl_timestamp + DEPENDS ${OPIMPL_OPS_INFO} + ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py + ) + add_custom_target(${OPIMPL_TARGET} ALL + DEPENDS ${OPIMPL_OUT_DIR}/.impl_timestamp) + if (${ENABLE_SOURCE_PACKAGE}) + install(DIRECTORY ${OPIMPL_OUT_DIR}/dynamic + DESTINATION ${OPIMPL_INSTALL_DIR} + ) + endif() +endfunction() + +function(add_ops_replay_targets) + cmake_parse_arguments(OPREPLAY "" "OPS_INFO;COMPUTE_UNIT;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) + # ccec compile options + set(ccec_base_opts -c -O2 --cce-aicore-only -mllvm -cce-aicore-function-stack-size=16000 + -mllvm -cce-aicore-record-overflow=false -std=c++17) + set(ccec_extopts_ascend310p --cce-aicore-arch=dav-m200 -mllvm -cce-aicore-fp-ceiling=2) + set(ccec_extopts_ascend910 --cce-aicore-arch=dav-c100) + set(ccec_extopts_ascend910b --cce-aicore-arch=dav-c220-cube) + file(MAKE_DIRECTORY ${OPREPLAY_OUT_DIR}) + execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_replay_build.py + ${OPREPLAY_OPS_INFO} + "${OPREPLAY_OPS_BATCH}" "${OPREPLAY_OPS_ITERATE}" + ${OPREPLAY_IMPL_DIR} + ${OPREPLAY_OUT_DIR} + ${OPREPLAY_COMPUTE_UNIT} + ) + file(GLOB replay_kernel_entries ${OPREPLAY_OUT_DIR}/*.cce) + if (NOT "${replay_kernel_entries}x" STREQUAL "x") + foreach(replay_kernel_file ${replay_kernel_entries}) + get_filename_component(replay_kernel_file_name "${replay_kernel_file}" NAME) + string(REPLACE "_entry.cce" "" op_kerne_name ${replay_kernel_file_name}) + file(GLOB replay_lib_src ${OPREPLAY_OUT_DIR}/${op_kerne_name}*.cpp) + set(OP_TILING_DATA_H_PATH ${OPREPLAY_OUT_DIR}/${op_kerne_name}_tiling_data.h) + add_library(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} SHARED ${replay_lib_src}) + if(EXISTS ${OP_TILING_DATA_H_PATH}) + target_compile_options(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE + -include ${OP_TILING_DATA_H_PATH} + ) + endif() + target_compile_definitions(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE + ${op_kerne_name}=${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} + ) + target_compile_options(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE + -D__ASCENDC_REPLAY__ + ) + target_link_libraries(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE intf_pub + tikreplaylib::${OPREPLAY_COMPUTE_UNIT} + register + ) + add_custom_command(OUTPUT ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o + COMMAND ccec ${ccec_base_opts} ${ccec_extopts_${OPREPLAY_COMPUTE_UNIT}} ${replay_kernel_file} + -o ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o + DEPENDS ${replay_kernel_file} + ) + add_custom_target(replay_kernel_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} ALL + DEPENDS ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o + ) + install(TARGETS replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_replay + ) + install(FILES ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o + DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_replay + ) + endforeach() + endif() +endfunction() + +function(add_npu_support_target) + cmake_parse_arguments(NPUSUP "" "TARGET;OPS_INFO_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) + get_filename_component(npu_sup_file_path "${NPUSUP_OUT_DIR}" DIRECTORY) + add_custom_command(OUTPUT ${NPUSUP_OUT_DIR}/npu_supported_ops.json + COMMAND mkdir -p ${NPUSUP_OUT_DIR} + COMMAND ${CMAKE_SOURCE_DIR}/cmake/util/gen_ops_filter.sh + ${NPUSUP_OPS_INFO_DIR} + ${NPUSUP_OUT_DIR} + ) + add_custom_target(npu_supported_ops ALL + DEPENDS ${NPUSUP_OUT_DIR}/npu_supported_ops.json + ) + install(FILES ${NPUSUP_OUT_DIR}/npu_supported_ops.json + DESTINATION ${NPUSUP_INSTALL_DIR} + ) +endfunction() + +function(add_bin_compile_target) + cmake_parse_arguments(BINCMP "" "TARGET;OPS_INFO;COMPUTE_UNIT;IMPL_DIR;ADP_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/src) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/gen) + execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_bin_param_build.py + ${BINCMP_OPS_INFO} ${BINCMP_OUT_DIR}/gen ${BINCMP_COMPUTE_UNIT} + RESULT_VARIABLE EXEC_RESULT + OUTPUT_VARIABLE EXEC_INFO + ERROR_VARIABLE EXEC_ERROR + ) + if (${EXEC_RESULT}) + message("ops binary compile scripts gen info: ${EXEC_INFO}") + message("ops binary compile scripts gen error: ${EXEC_ERROR}") + message(FATAL_ERROR "ops binary compile scripts gen failed!") + endif() + if (NOT TARGET binary) + add_custom_target(binary) + endif() + add_custom_target(${BINCMP_TARGET} + COMMAND cp -r ${BINCMP_IMPL_DIR}/*.* ${BINCMP_OUT_DIR}/src + ) + add_custom_target(${BINCMP_TARGET}_gen_ops_config + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/insert_simplified_keys.py -p ${BINCMP_OUT_DIR}/bin + COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_ops_config.py -p ${BINCMP_OUT_DIR}/bin + -s ${BINCMP_COMPUTE_UNIT} + ) + add_dependencies(binary ${BINCMP_TARGET}_gen_ops_config) + file(GLOB bin_scripts ${BINCMP_OUT_DIR}/gen/*.sh) + foreach(bin_script ${bin_scripts}) + get_filename_component(bin_file ${bin_script} NAME_WE) + string(REPLACE "-" ";" bin_sep ${bin_file}) + list(GET bin_sep 0 op_type) + list(GET bin_sep 1 op_file) + list(GET bin_sep 2 op_index) + if (NOT TARGET ${BINCMP_TARGET}_${op_file}_copy) + file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file}) + add_custom_target(${BINCMP_TARGET}_${op_file}_copy + COMMAND cp ${BINCMP_ADP_DIR}/${op_file}.py ${BINCMP_OUT_DIR}/src/${op_type}.py + ) + install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} + DESTINATION ${BINCMP_INSTALL_DIR}/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) + install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json + DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT}/ OPTIONAL + ) + endif() + add_custom_target(${BINCMP_TARGET}_${op_file}_${op_index} + COMMAND export HI_PYTHON=${ASCEND_PYTHON_EXECUTABLE} && bash ${bin_script} ${BINCMP_OUT_DIR}/src/${op_type}.py ${BINCMP_OUT_DIR}/bin/${op_file} + WORKING_DIRECTORY ${BINCMP_OUT_DIR} + ) + add_dependencies(${BINCMP_TARGET}_${op_file}_${op_index} ${BINCMP_TARGET} ${BINCMP_TARGET}_${op_file}_copy) + add_dependencies(${BINCMP_TARGET}_gen_ops_config ${BINCMP_TARGET}_${op_file}_${op_index}) + endforeach() + install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json + DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT} OPTIONAL + ) +endfunction() diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/intf.cmake b/atb_operator/AtbAddSample/AddAscendC/cmake/intf.cmake new file mode 100755 index 000000000..2f362c396 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/intf.cmake @@ -0,0 +1,26 @@ + +add_library(intf_pub INTERFACE) +target_compile_options(intf_pub INTERFACE + -fPIC + -fvisibility=hidden + -fvisibility-inlines-hidden + $<$:-O2> + $<$:-O0 -g> + $<$:-std=c++11> + $<$,$>:-ftrapv -fstack-check> + $<$:-pthread -Wfloat-equal -Wshadow -Wformat=2 -Wno-deprecated -Wextra> + $,-fstack-protector-strong,-fstack-protector-all> +) +target_compile_definitions(intf_pub INTERFACE + _GLIBCXX_USE_CXX11_ABI=0 + $<$:_FORTIFY_SOURCE=2> +) +target_include_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/include) +target_link_options(intf_pub INTERFACE + $<$,EXECUTABLE>:-pie> + $<$:-s> + -Wl,-z,relro + -Wl,-z,now + -Wl,-z,noexecstack +) +target_link_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/lib64) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/makeself.cmake b/atb_operator/AtbAddSample/AddAscendC/cmake/makeself.cmake new file mode 100755 index 000000000..48c565bfb --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/makeself.cmake @@ -0,0 +1,17 @@ +execute_process(COMMAND chmod +x ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh) +execute_process(COMMAND ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh + --header ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself-header.sh + --help-header ./help.info + --gzip --complevel 4 --nomd5 --sha256 + ./ ${CPACK_PACKAGE_FILE_NAME} "version:1.0" ./install.sh + WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} + RESULT_VARIABLE EXEC_RESULT + ERROR_VARIABLE EXEC_ERROR +) +if (NOT "${EXEC_RESULT}x" STREQUAL "0x") + message(FATAL_ERROR "CPack Command error: ${EXEC_RESULT}\n${EXEC_ERROR}") +endif() +execute_process(COMMAND cp ${CPACK_EXTERNAL_BUILT_PACKAGES} ${CPACK_PACKAGE_DIRECTORY}/ + COMMAND echo "Copy ${CPACK_EXTERNAL_BUILT_PACKAGES} to ${CPACK_PACKAGE_DIRECTORY}/" + WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} +) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/__init__.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/__init__.py new file mode 100755 index 000000000..c4ddc893a --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/__init__.py @@ -0,0 +1,8 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- + +import sys +import os + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) +sys.path.append(PYF_PATH) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py new file mode 100755 index 000000000..decf34544 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py @@ -0,0 +1,129 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import json +import hashlib +import const_var +import opdesc_parser + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + + +class BinParamBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + super().__init__(op_type) + self.soc = '' + self.out_path = '' + + def set_soc_version(self: any, soc: str): + self.soc = soc + + def set_out_path(self: any, out_path: str): + self.out_path = out_path + + def gen_input_json(self: any): + key_map = {} + count = len(self.input_dtype[0].split(',')) + for i in range(0, count): + inputs = [] + outputs = [] + attrs = [] + op_node = {} + for idx in range(0, len(self.input_name)): + idtypes = self.input_dtype[idx].split(',') + ifmts = self.input_fmt[idx].split(',') + itype = self.input_type[idx] + para = {} + para['name'] = self.input_name[idx] + para['index'] = idx + para['dtype'] = idtypes[i] + para['format'] = ifmts[i] + para['paramType'] = itype + para['shape'] = [-2] + if itype == 'dynamic': + inputs.append([para]) + else: + inputs.append(para) + for idx in range(0, len(self.output_name)): + odtypes = self.output_dtype[idx].split(',') + ofmts = self.output_fmt[idx].split(',') + otype = self.output_type[idx] + para = {} + para['name'] = self.output_name[idx] + para['index'] = idx + para['dtype'] = odtypes[i] + para['format'] = ofmts[i] + para['paramType'] = otype + para['shape'] = [-2] + if otype == 'dynamic': + outputs.append([para]) + else: + outputs.append(para) + for attr in self.attr_list: + att = {} + att['name'] = attr + atype = self.attr_val.get(attr).get('type').lower() + atype = atype.replace('list', 'list_') + att['dtype'] = atype + att['value'] = const_var.ATTR_DEF_VAL.get(atype) + attrs.append(att) + op_node['bin_filename'] = '' + op_node['inputs'] = inputs + op_node['outputs'] = outputs + if len(attrs) > 0: + op_node['attrs'] = attrs + param = {} + param['op_type'] = self.op_type + param['op_list'] = [op_node] + objstr = json.dumps(param, indent=' ') + md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() + while key_map.get(md5sum) is not None: + objstr += '1' + md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() + key_map[md5sum] = md5sum + bin_file = self.op_type + '_' + md5sum + op_node['bin_filename'] = bin_file + param_file = os.path.join(self.out_path, bin_file + '_param.json') + param_file = os.path.realpath(param_file) + with os.fdopen(os.open(param_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + json.dump(param, fd, indent=' ') + self._write_buld_cmd(param_file, bin_file, i) + + + def _write_buld_cmd(self: any, param_file: str, bin_file: str, index: int): + hard_soc = const_var.SOC_MAP_EXT.get(self.soc) + if not hard_soc: + hard_soc = soc.capitalize() + name_com = [self.op_type, self.op_file, str(index)] + compile_file = os.path.join(self.out_path, '-'.join(name_com) + '.sh') + compile_file = os.path.realpath(compile_file) + with os.fdopen(os.open(compile_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + fd.write('#!/bin/bash\n') + fd.write('echo "[{}] Generating {} ..."\n'.format(hard_soc, bin_file)) + cmd = const_var.BIN_CMD.format(fun=self.op_intf, soc=hard_soc, param=param_file, impl='""') + fd.write(cmd) + chk = const_var.CHK_CMD.format(res_file=bin_file + '.json') + fd.write(chk) + chk = const_var.CHK_CMD.format(res_file=bin_file + '.o') + fd.write(chk) + fd.write('echo "[{}] Generating {} Done"\n'.format(hard_soc, bin_file)) + + +def gen_bin_param_file(cfgfile: str, out_dir: str, soc: str): + op_descs = opdesc_parser.get_op_desc(cfgfile, [], [], BinParamBuilder, None) + for op_desc in op_descs: + op_desc.set_soc_version(soc) + op_desc.set_out_path(out_dir) + op_desc.gen_input_json() + + +if __name__ == '__main__': + if len(sys.argv) <= 3: + raise RuntimeError('arguments must greater than 3') + gen_bin_param_file(sys.argv[1], sys.argv[2], sys.argv[3]) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py new file mode 100755 index 000000000..7fe177da1 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py @@ -0,0 +1,446 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import re +import stat +import opdesc_parser +import const_var + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + +IMPL_HEAD = ''' +import os, sys +import ctypes +import json +import shutil +from tbe.common.platform import get_soc_spec +from tbe.common.utils import para_check +from tbe.tikcpp import compile_op, replay_op, check_op_cap, generalize_op_params, get_code_channel, OpInfo +from tbe.common.buildcfg import get_default_build_config +from impl.util.platform_adapter import tbe_register +from tbe.common.buildcfg import get_current_build_config +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + +DTYPE_MAP = {"float32": ["DT_FLOAT", "float"], + "float16": ["DT_FLOAT16", "half"], + "int8": ["DT_INT8", "int8_t"], + "int16": ["DT_INT16", "int16_t"], + "int32": ["DT_INT32", "int32_t"], + "int64": ["DT_INT64", "int64_t"], + "uint1": ["DT_UINT1", "uint8_t"], + "uint8": ["DT_UINT8", "uint8_t"], + "uint16": ["DT_UINT16", "uint16_t"], + "uint32": ["DT_UINT32", "uint32_t"], + "uint64": ["DT_UINT64", "uint64_t"], + "bool": ["DT_BOOL", "bool"], + "double": ["DT_DOUBLE", "double"], + "dual": ["DT_DUAL", "unknown"], + "dual_sub_int8": ["DT_DUAL_SUB_INT8", "unknown"], + "dual_sub_uint8": ["DT_DUAL_SUB_UINT8", "unknown"], + "string": ["DT_STRING", "unknown"], + "complex64": ["DT_COMPLEX64", "unknown"], + "complex128": ["DT_COMPLEX128", "unknown"], + "qint8": ["DT_QINT8", "unknown"], + "qint16": ["DT_QINT16", "unknown"], + "qint32": ["DT_QINT32", "unknown"], + "quint8": ["DT_QUINT8", "unknown"], + "quint16": ["DT_QUINT16", "unknown"], + "resource": ["DT_RESOURCE", "unknown"], + "string_ref": ["DT_STRING_REF", "unknown"], + "int4": ["DT_INT4", "int8_t"], + "bfloat16": ["DT_BF16", "bfloat16_t"]} + +def get_dtype_fmt_options(__inputs__, __outputs__): + options = [] + for x in __inputs__ + __outputs__: + x_n = x.get("param_name").upper() + x_fmt = x.get("format") + x_dtype = x.get("dtype") + options.append("-DDTYPE_{n}={t}".format(n=x_n, t=DTYPE_MAP.get(x_dtype)[1])) + options.append("-DORIG_DTYPE_{n}={ot}".format(n=x_n, ot=DTYPE_MAP.get(x_dtype)[0])) + options.append("-DFORMAT_{n}=FORMAT_{f}".format(n=x_n, f=x_fmt)) + return options + +def load_dso(so_path): + try: + ctypes.CDLL(so_path) + except OSError as error : + print(error) + raise RuntimeError("cannot open %s" %(so_path)) + else: + print("load so succ ", so_path) + +def get_shortsoc_compile_option(compile_option_list: list, shortsoc:str): + compile_options = [] + if shortsoc in compile_option_list: + compile_options = compile_option_list[shortsoc] + elif '__ALLSOC__' in compile_option_list: + compile_options = compile_option_list['__ALLSOC__'] + return compile_options + +''' + +IMPL_API = ''' +@tbe_register.register_operator("{}") +@para_check.check_op_params({}) +def {}({}, kernel_name="{}", impl_mode=""): + if get_current_build_config("enable_op_prebuild"): + return + __inputs__, __outputs__, __attrs__ = _build_args({}) + options = get_dtype_fmt_options(__inputs__, __outputs__) + options += ["-x", "cce"] + ccec = os.environ.get('CCEC_REAL_PATH') + if ccec is None: + ccec = shutil.which("ccec") + if ccec != None: + ccec_path = os.path.dirname(ccec) + tikcpp_path = os.path.realpath(os.path.join(ccec_path, "..", "..", "tikcpp")) + else: + tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") + options.append("-I" + tikcpp_path) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) + options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) + options.append("-I" + os.path.join(PYF_PATH, "..", "ascendc", "common")) + if impl_mode == "high_performance": + options.append("-DHIGH_PERFORMANCE=1") + elif impl_mode == "high_precision": + options.append("-DHIGH_PRECISION=1") + if get_default_build_config("enable_deterministic_mode") == 1: + options.append("-DDETEMINISTIC_MODE=1") + + custom_compile_options = {}, + custom_all_compile_options = {}, + soc_version = get_soc_spec("SOC_VERSION") + soc_short = get_soc_spec("SHORT_SOC_VERSION").lower() + custom_compile_options_soc = get_shortsoc_compile_option(custom_compile_options[0], soc_short) + custom_all_compile_options_soc = get_shortsoc_compile_option(custom_all_compile_options[0], soc_short) + options += custom_all_compile_options_soc + options += custom_compile_options_soc + + origin_func_name = "{}" + ascendc_src_dir = "{}" + ascendc_src_file = "{}" + src = os.path.join(PYF_PATH, "..", "ascendc", ascendc_src_dir, ascendc_src_file) + if not os.path.exists(src): + src = os.path.join(PYF_PATH, ascendc_src_file) +''' + +REPLAY_OP_API = ''' + print("start replay Acend C Operator {}, kernel name is {}") + tikreplay_codegen_path = tikcpp_path + "/tikreplaylib/lib" + tikreplay_stub_path = tikcpp_path + "/tikreplaylib/lib/" + soc_version + print("start load libtikreplaylib_codegen.so and libtikreplaylib_stub.so") + codegen_so_path = tikreplay_codegen_path + "/libtikreplaylib_codegen.so" + replaystub_so_path = tikreplay_stub_path + "/libtikreplaylib_stub.so" + if PYF_PATH.endswith("dynamic"): + op_replay_path = os.path.join(PYF_PATH, "..", "..", "op_replay") + else: + op_replay_path = os.path.join(PYF_PATH, "..", "op_replay") + replayapi_so_path = os.path.join(op_replay_path, "libreplay_{}_" + soc_short + ".so") + load_dso(codegen_so_path) + load_dso(replaystub_so_path) + load_dso(replayapi_so_path) + op_type = "{}" + entry_obj = os.path.join(op_replay_path, "{}_entry_" + soc_short + ".o") + code_channel = get_code_channel(src, kernel_name, op_type, options) + op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ + attrs = __attrs__, impl_mode = impl_mode) + res, msg = replay_op(op_info, entry_obj, code_channel, src, options) + if not res: + print("call replay op failed for %s and get into call compile op" %(msg)) + compile_op(src, origin_func_name, op_info, options, code_channel, '{}') +''' + +COMPILE_OP_API = ''' + print("start compile Ascend C operator {}. kernel name is {}") + op_type = "{}" + code_channel = get_code_channel(src, kernel_name, op_type, options) + op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ + attrs = __attrs__, impl_mode = impl_mode, origin_inputs=[{}], origin_outputs = [{}]) + compile_op(src, origin_func_name, op_info, options, code_channel, '{}') +''' + +SUP_API = ''' +def {}({}, impl_mode=""): + __inputs__, __outputs__, __attrs__ = _build_args({}) + ret_str = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) + ret_dict = json.loads(ret_str) + err_code = ret_dict.get("ret_code") + sup = "Unknown" + reason = "Unknown reason" + if err_code is not None: + if err_code is 0: + sup = "True" + reason = "" + elif err_code is 1: + sup = "False" + reason = ret_dict.get("reason") + else: + sup = "Unknown" + reason = ret_dict.get("reason") + return sup, reason +''' +CAP_API = ''' +def {}({}, impl_mode=""): + __inputs__, __outputs__, __attrs__ = _build_args({}) + result = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) + return result.decode("utf-8") +''' +GLZ_API = ''' +@tbe_register.register_param_generalization("{}") +def {}_generalization({}, generalize_config=None): + __inputs__, __outputs__, __attrs__ = _build_args({}) + ret_str = generalize_op_params("{}", __inputs__, __outputs__, __attrs__, generalize_config) + return [json.loads(ret_str)] +''' + +ATTR_DEFAULT = {'bool': 'False', 'int': '0', 'float': '0.0', 'listInt': '[]', + 'listFloat': '[]', 'listBool': '[]', 'listListInt': '[[]]', 'str': ''} + + +def optype_snake(origin_str): + temp_str = origin_str[0].lower() + origin_str[1:] + new_str = re.sub(r'([A-Z])', r'_\1', temp_str).lower() + return new_str + + +class AdpBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + self.argsname = [] + self.argsdefv = [] + self.op_compile_option:str = '{}' + super().__init__(op_type) + + + def write_adapt(self: any, impl_path, path: str, op_compile_option_all:list = None): + self._build_paradefault() + if impl_path != "": + src_file = os.path.join(impl_path, self.op_file + '.cpp') + if not os.path.exists(src_file): + return + out_path = os.path.abspath(path) + if self.dynamic_shape and not out_path.endswith('dynamic'): + out_path = os.path.join(path, 'dynamic') + os.makedirs(out_path, mode=0o700, exist_ok=True) + adpfile = os.path.join(out_path, self.op_file + '.py') + self._gen_op_compile_option(op_compile_option_all) + with os.fdopen(os.open(adpfile, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + self._write_head(fd) + self._write_argparse(fd) + self._write_impl(fd) + if self.op_chk_support: + self._write_cap('check_supported', fd) + self._write_cap('get_op_support_info', fd) + if self.op_fmt_sel: + self._write_cap('op_select_format', fd) + self._write_cap('get_op_specific_info', fd) + if self.op_range_limit == 'limited' or self.op_range_limit == 'dynamic': + self._write_glz(fd) + + + def _gen_op_compile_option(self:any, op_compile_option_all:list =None): + if op_compile_option_all is not None: + if self.op_type in op_compile_option_all: + self.op_compile_option = op_compile_option_all[self.op_type] + elif "__all__" in op_compile_option_all: + self.op_compile_option = op_compile_option_all["__all__"] + + + def _ip_argpack(self: any, default: bool = True) -> list: + args = [] + for i in range(len(self.input_name)): + arg = self.input_name[i] + if default and self.argsdefv[i] is not None: + arg += '=' + self.argsdefv[i] + args.append(arg) + return args + + def _op_argpack(self: any, default: bool = True) -> list: + args = [] + argidx = len(self.input_name) + for i in range(len(self.output_name)): + arg = self.output_name[i] + if default and self.argsdefv[i + argidx] is not None: + arg += '=' + self.argsdefv[i + argidx] + args.append(arg) + return args + + def _attr_argpack(self: any, default: bool = True) -> list: + args = [] + argidx = len(self.input_name) + len(self.output_name) + for i in range(len(self.attr_list)): + att = self.attr_list[i] + arg = att + if default and self.argsdefv[i + argidx] is not None: + if self.attr_val.get(att).get('type') == 'str': + arg += '="' + self.argsdefv[i + argidx] + '"' + elif self.attr_val.get(att).get('type') == 'bool': + arg += '=' + self.argsdefv[i + argidx].capitalize() + else: + arg += '=' + self.argsdefv[i + argidx] + args.append(arg) + return args + + def _build_paralist(self: any, default: bool = True) -> str: + args = [] + args.extend(self._ip_argpack(default)) + args.extend(self._op_argpack(default)) + args.extend(self._attr_argpack(default)) + return ', '.join(args) + + def _io_parachk(self: any, types: list, type_name: str) -> list: + chk = [] + for iot in types: + if iot == 'optional': + ptype = 'OPTION' + else: + ptype = iot.upper() + chk.append('para_check.{}_{}'.format(ptype, type_name)) + return chk + + def _attr_parachk(self: any) -> list: + chk = [] + for att in self.attr_list: + if self.attr_val.get(att).get('paramType') == 'optional': + pt = 'OPTION' + else: + pt = self.attr_val.get(att).get('paramType').upper() + att_type = self.attr_val.get(att).get('type').upper() + att_type = att_type.replace('LIST', 'LIST_') + chk.append('para_check.{}_ATTR_{}'.format(pt, att_type)) + return chk + + def _build_parachk(self: any) -> str: + chk = [] + chk.extend(self._io_parachk(self.input_type, 'INPUT')) + chk.extend(self._io_parachk(self.output_type, 'OUTPUT')) + chk.extend(self._attr_parachk()) + chk.append('para_check.KERNEL_NAME') + return ', '.join(chk) + + def _build_paradefault(self: any): + optional = False + argtypes = [] + argtypes.extend(self.input_type) + argtypes.extend(self.output_type) + for atype in argtypes: + if atype == 'optional': + optional = True + if optional: + self.argsdefv.append('None') + else: + self.argsdefv.append(None) + for attr in self.attr_list: + atype = self.attr_val.get(attr).get('paramType') + if atype == 'optional': + optional = True + attrval = self.attr_val.get(attr).get('defaultValue') + if attrval is not None: + optional = True + if type == "bool": + attrval = attrval.capitalize() + elif type == "str": + attrval = "\"" + attrval + "\"" + self.argsdefv.append(attrval) + continue + if optional: + self.argsdefv.append(ATTR_DEFAULT.get(self.attr_val.get(attr).get('type'))) + else: + self.argsdefv.append(None) + + def _write_head(self: any, fd: object): + fd.write(IMPL_HEAD) + + def _write_argparse(self: any, fd: object): + args = self._build_paralist(False) + fd.write('def _build_args({}):\n'.format(args)) + fd.write(' __inputs__ = []\n') + fd.write(' for arg in [{}]:\n'.format(', '.join(self.input_name))) + fd.write(' if arg != None:\n') + fd.write(' if isinstance(arg, (list, tuple)):\n') + fd.write(' if len(arg) == 0:\n') + fd.write(' continue\n') + fd.write(' __inputs__.append(arg[0])\n') + fd.write(' else:\n') + fd.write(' __inputs__.append(arg)\n') + fd.write(' __outputs__ = []\n') + fd.write(' for arg in [{}]:\n'.format(', '.join(self.output_name))) + fd.write(' if arg != None:\n') + fd.write(' if isinstance(arg, (list, tuple)):\n') + fd.write(' if len(arg) == 0:\n') + fd.write(' continue\n') + fd.write(' __outputs__.append(arg[0])\n') + fd.write(' else:\n') + fd.write(' __outputs__.append(arg)\n') + fd.write(' __attrs__ = []\n') + for attr in self.attr_list: + fd.write(' if {} != None:\n'.format(attr)) + fd.write(' attr = {}\n') + fd.write(' attr["name"] = "{}"\n'.format(attr)) + fd.write(' attr["dtype"] = "{}"\n'.format(self.attr_val.get(attr).get('type'))) + fd.write(' attr["value"] = {}\n'.format(attr)) + fd.write(' __attrs__.append(attr)\n') + fd.write(' return __inputs__, __outputs__, __attrs__\n') + + def _write_impl(self: any, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + pchk = self._build_parachk() + if len(self.kern_name) > 0: + kern_name = self.kern_name + else: + kern_name = self.op_intf + src = self.op_file + '.cpp' + fd.write(IMPL_API.format(self.op_type, pchk, self.op_intf, argsdef, kern_name, argsval,\ + self.custom_compile_options, self.custom_all_compile_options, self.op_intf,\ + optype_snake(self.op_type), src)) + if self.op_replay_flag: + fd.write(REPLAY_OP_API.format(self.op_type, kern_name, self.op_file, self.op_type, self.op_file,\ + self.op_compile_option)) + else: + fd.write(COMPILE_OP_API.format(self.op_type, kern_name, self.op_type, ', '.join(self.input_name),\ + ', '.join(self.output_name), self.op_compile_option)) + + def _write_cap(self: any, cap_name: str, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + if cap_name == 'check_supported': + fd.write(SUP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) + else: + fd.write(CAP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) + + def _write_glz(self: any, fd: object): + argsdef = self._build_paralist() + argsval = self._build_paralist(False) + fd.write(GLZ_API.format(self.op_type, self.op_intf, argsdef, argsval, self.op_type)) + + +def write_scripts(cfgfile: str, cfgs: dict, dirs: dict, ops: list = None, op_compile_option:list = None): + batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') + iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') + file_map = {} + op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, AdpBuilder,\ + ops, dirs.get(const_var.AUTO_GEN_DIR)) + for op_desc in op_descs: + op_desc.write_adapt(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), op_compile_option) + file_map[op_desc.op_type] = op_desc.op_file + return file_map + +if __name__ == '__main__': + if len(sys.argv) <= 6: + raise RuntimeError('arguments must greater equal than 6') + rep_cfg = {} + rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] + rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] + cfg_dir = {} + cfg_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] + cfg_dir[const_var.CFG_OUT_DIR] = sys.argv[5] + cfg_dir[const_var.AUTO_GEN_DIR] = sys.argv[6] + write_scripts(cfgfile=sys.argv[1], cfgs=rep_cfg, dirs=cfg_dir) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py new file mode 100755 index 000000000..7a97180be --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py @@ -0,0 +1,114 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import glob +import json +import argparse +import const_var + + +def load_json(json_file: str): + with open(json_file, encoding='utf-8') as file: + json_content = json.load(file) + return json_content + + +def get_specified_suffix_file(root_dir, suffix): + specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) + all_suffix_files = glob.glob(specified_suffix, recursive=True) + return all_suffix_files + + +def add_simplified_config(op_type, key, core_type, objfile, config): + simple_cfg = config.get('binary_info_config.json') + op_cfg = simple_cfg.get(op_type) + if not op_cfg: + op_cfg = {} + op_cfg['dynamicRankSupport'] = True + op_cfg['simplifiedKeyMode'] = 0 + op_cfg['binaryList'] = [] + simple_cfg[op_type] = op_cfg + bin_list = op_cfg.get('binaryList') + bin_list.append({'coreType': core_type, 'simplifiedKey': key, 'binPath': objfile}) + + +def add_op_config(op_file, bin_info, config): + op_cfg = config.get(op_file) + if not op_cfg: + op_cfg = {} + op_cfg['binList'] = [] + config[op_file] = op_cfg + op_cfg.get('binList').append(bin_info) + + +def gen_ops_config(json_file, soc, config): + core_type_map = {"MIX": 0, "AiCore": 1, "VectorCore": 2} + contents = load_json(json_file) + if ('binFileName' not in contents) or ('supportInfo' not in contents): + return + json_base_name = os.path.basename(json_file) + op_dir = os.path.basename(os.path.dirname(json_file)) + support_info = contents.get('supportInfo') + bin_name = contents.get('binFileName') + bin_suffix = contents.get('binFileSuffix') + core_type = core_type_map.get(contents.get("coreType")) + bin_file_name = bin_name + bin_suffix + op_type = bin_name.split('_')[0] + op_file = op_dir + '.json' + bin_info = {} + keys = support_info.get('simplifiedKey') + if keys: + bin_info['simplifiedKey'] = keys + for key in keys: + add_simplified_config(op_type, key, core_type, os.path.join(soc, op_dir, bin_file_name), config) + bin_info['staticKey'] = support_info.get('staticKey') + bin_info['int64Mode'] = support_info.get('int64Mode') + bin_info['inputs'] = support_info.get('inputs') + bin_info['outputs'] = support_info.get('outputs') + if support_info.get('attrs'): + bin_info['attrs'] = support_info.get('attrs') + bin_info['binInfo'] = {'jsonFilePath': os.path.join(soc, op_dir, json_base_name)} + add_op_config(op_file, bin_info, config) + + +def gen_all_config(root_dir, soc): + suffix = 'json' + config = {} + config['binary_info_config.json'] = {} + all_json_files = get_specified_suffix_file(root_dir, suffix) + for _json in all_json_files: + gen_ops_config(_json, soc, config) + for cfg_key in config.keys(): + cfg_file = os.path.join(root_dir, cfg_key) + with os.fdopen(os.open(cfg_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: + json.dump(config.get(cfg_key), fd, indent=' ') + + +def args_prase(): + parser = argparse.ArgumentParser() + parser.add_argument('-p', + '--path', + nargs='?', + required=True, + help='Parse the path of the json file.') + parser.add_argument('-s', + '--soc', + nargs='?', + required=True, + help='Parse the soc_version of ops.') + return parser.parse_args() + + +def main(): + args = args_prase() + gen_all_config(args.path, args.soc) + + +if __name__ == '__main__': + main() diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py new file mode 100755 index 000000000..1cac7d911 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py @@ -0,0 +1,65 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import opdesc_parser +import replay_codegen +import const_var +from replay_codegen import ReplayCodeGenParams + +PYF_PATH = os.path.dirname(os.path.realpath(__file__)) + + +class ReplayBuilder(opdesc_parser.OpDesc): + def __init__(self: any, op_type: str): + super().__init__(op_type) + + def gen_replay_source(self: any, impl_path: str, out_path: str, ops_product: str): + if not self.op_replay_flag: + print('{} replay not enabled'.format(self.op_type)) + return + argn = len(self.input_name) + len(self.output_name) + 1 + if self.op_replay_batch: + print('{} replay in batch mode'.format(self.op_type)) + else: + print('{} replay in normal mode'.format(self.op_type)) + if impl_path.endswith('op_kernel'): + implf = os.path.join(impl_path, self.op_file + '.cpp') + tiling_file = os.path.join(impl_path, "../op_host", self.op_file + '_tiling.h') + else: + if self.dynamic_shape: + dyn_path = 'dynamic' + else: + dyn_path = '' + implf = os.path.join(impl_path, dyn_path, self.op_file + '.cpp') + tiling_file = os.path.join(impl_path, "../../op_tiling", self.op_file + '_tiling.h') + rep_conf = replay_codegen.ReplayCodeGen(ReplayCodeGenParams(self.op_type, implf, tiling_file, self.op_file, \ + self.op_intf, argn, self.op_replay_batch, self.max_block_dim, self.max_shape_size)) + rep_conf.set_batch(self.op_replay_batch) + rep_conf.set_outdir(out_path) + rep_conf.gen_replay(ops_product) + + +def gen_replay(cfgfile: str, cfgs: dict, dirs: dict, ops_product: str, ops: list = None): + batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') + iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') + op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, ReplayBuilder, ops) + for op_desc in op_descs: + op_desc.gen_replay_source(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), ops_product) + + +if __name__ == '__main__': + if len(sys.argv) <= 6: + raise RuntimeError('arguments must greater than 6') + rep_cfg = {} + rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] + rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] + rep_dir = {} + rep_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] + rep_dir[const_var.CFG_OUT_DIR] = sys.argv[5] + gen_replay(sys.argv[1], rep_cfg, rep_dir, sys.argv[6]) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp b/atb_operator/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp new file mode 100755 index 000000000..0e8834664 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp @@ -0,0 +1,117 @@ +#include +#include +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#include "register/op_check.h" +#define __ASCENDC_REPLAY_CODE__ +#include + +using namespace std; +using namespace optiling; +using namespace AscendCReplay; + +extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); +extern "C" int elf_batch_append(char *elf, uint32_t elfSize, char *jit, int kernum, char *atext[], int alen[], + int atlen, const char* kernelname[]); + +#define KERNEL_N 1 +#define ARG_N (__ARG_NUM__) +#define MAX_L (1024 * 1024 * 100) +#define MAX_E (1024 * 1024) + +int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) +{ + // gen type 1 : direct call codes 0: load .o file + if (param.gentype < 0 || param.gentype > 1) { + printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); + return 0; + } else if (param.gentype == 1 && param.objptr == nullptr) { + printf("Error: call replay with direct call mode, but code obj addr is null\n"); + return 0; + } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { + printf("Error: call replay with object file mode, but object file path is null\n"); + return 0; + } + // core_type 0:MIX 1:CUBE 2:VEC + if (core_type < 0 || core_type > 2) { + printf("Error: call replay core type is %d !\n", core_type); + return 0; + } + g_coreType = __CORE_TYPE__; + g_taskRation = param.task_ration; + g_tilingKey = param.tiling_key; + + unsigned char *buf, *jit; + char *kernel[KERNEL_N]; + int len[KERNEL_N]; + block_idx = 0; + block_num = param.block_dim; + g_ubBase = block_num; + uint8_t *code = (uint8_t *)malloc(MAX_L); + uint8_t *pos = code; + struct timespec tp1, tp2; + + clock_gettime(CLOCK_MONOTONIC, &tp1); + if (block_num > 32) { + printf("Error: block_num > 32\n"); + return 0; + } + //__OP_FOPEN__ + for (int i = 0; i < KERNEL_N; i++) { + //__OP_SET_KERNEL__ + for (int j = 0; j < ARG_N; j++) + AddArg(j, ARG_STEP * (j + 1)); +#ifdef FP_CEILING + SetCtrlFloatEnable(); +#else + SetCtrlFloatDisable(); +#endif + CodeInit(pos, true); + __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); + CodeEnd(); + kernel[i] = (char *)pos; + len[i] = CodeLen(); + pos += len[i]; + } + //__OP_FCLOSE__ + clock_gettime(CLOCK_MONOTONIC, &tp2); + buf = (unsigned char *)malloc(MAX_E); + int fd = open(param.entry_file, O_RDONLY); + if (fd < 0) { + printf("[error]: cannot find entry.o : %s\n", param.entry_file); + return 0; + } + uint32_t bufSize = read(fd, buf, MAX_E); + if (bufSize <= 0) { + printf("[error]: entry.o : %s is too small ! \n", param.entry_file); + } + close(fd); + jit = (unsigned char *)malloc(MAX_L); + printf("total code generated %ld\n", pos - code); + int sz = elf_batch_append((char *)buf, bufSize, (char *)jit, KERNEL_N, kernel, len, pos - code, ¶m.kernel_name); + if (tp1.tv_sec != tp2.tv_sec) { + printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); + } else { + printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); + } + printf("new elf size %d\n", sz); + if (param.gentype == 0) { + fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); + (void)write(fd, jit, sz); + close(fd); + free(jit); + } else if (param.gentype == 1) { + *param.objptr = (char*)jit; + } + free(buf); + free(code); + return sz; +} + +REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py new file mode 100755 index 000000000..137c9f39f --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py @@ -0,0 +1,115 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" +import os +import stat +import ctypes +import collections +import shutil +import subprocess +import copy + +"""CODE_* is used to cube/vector api is called in operator code +CODE_MIX means both cube and vector api is called +CODE_CUBE means only cube api is called +CODE_VEC means only vector api is called +""" +CODE_MIX = 0 +CODE_CUBE = 1 +CODE_VEC = 2 + + +def _is_v220(op_product: str): + """return if current soc version is V220 + + Returns: + res: True means V220 + """ + if op_product in ["ascend910b"]: + return True + return False + + +InfoCodeChanelParams = collections.namedtuple('InfoCodeChanelParams',\ +['src_file', 'tiling_header', 'kernel_name', 'outdir', 'op_product', 'compile_options']) + + +def infer_code_channel(params: InfoCodeChanelParams): + """get code channel for v220, return CODE_MIX if soc version is not V220 + + Args: + src_file (str): AscendC operator code file + src_file (str): AscendC operator tiling header file + kernel_name (str): kernel function name + optype (str): operator type + compile_options (list): compile options for ccec cmd + + Raises: + Exception: if not exist L1/L0/UB if code, it's not a aicore code + + Returns: + res (int): CODE_MIX/CODE_CUBE/CODE_VEC + """ + if not _is_v220(params.op_product): + return CODE_MIX + return CODE_VEC + if params.compile_options is None: + compile_options = [] + else: + compile_options = params.compile_options + ccec = shutil.which("ccec") + if ccec is not None: + ccec_path = os.path.dirname(ccec) + tikcpp_path = os.path.realpath(os.path.join(ccec_path, "..", "..", "tikcpp")) + else: + tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") + compile_options.append("-I" + tikcpp_path) + compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) + compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) + compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) + compile_options += ["-include", params.tiling_header] + arch = "dav-c220-cube" + sub_core_type = "AIC" + optional_core = "AiCore" + compile_cmd = [shutil.which("ccec"), '-c', '-O3'] + compile_cmd += compile_options + temp_file_name_tag = "_" + str(os.getpid()) + "_temp.o" + dst_file = os.path.join(kernel_meta_dir, kernel_name + temp_file_name_tag) + compile_cmd += [params.src_file, "--cce-aicore-arch={}".format(arch), + "--cce-aicore-only", "-o", dst_file, + "-mllvm", "-cce-aicore-function-stack-size=16000", + "-mllvm", "-cce-aicore-record-overflow=true", + "-mllvm", "-cce-aicore-addr-transform"] + compile_cmd += ["-std=c++17"] + print('get_code_channel: ', ' '.join(compile_cmd)) + proc = subprocess.Popen( + compile_cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) + (out, _) = proc.communicate() + if proc.returncode != 0: + print('get_code_channel coretype compile error: ', out.decode()) + msg = "compile %s error :%s\n" % (params.src_file, out.decode()) + raise Exception(f"get_code_channel coretype error, msg is{msg}") + objdump_cmd = ['objdump', '-s', '-j', '.text', '{}'.format(dst_file)] + + proc = subprocess.Popen( + objdump_cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) + (out, _) = proc.communicate() + if proc.returncode != 0: + print('get_code_channel objdump error: ', out.decode()) + msg = "get_code_channel objdump %s error :%s\n" % (src_file, out.decode()) + raise Exception(f"get_code_channel objdump error, msg is{msg}") + os.remove(dst_file) + lines = out.decode('utf-8').split('\n') + for line in lines: + insts = line.strip().split() + if len(insts) < 5: + continue + for inst in insts[1:]: + if len(inst) != 8: + continue + if inst[6] == 'f': + return CODE_MIX + return CODE_VEC diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/const_var.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/const_var.py new file mode 100755 index 000000000..8b32c3b91 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/const_var.py @@ -0,0 +1,33 @@ + +#!/usr/bin/env python +# coding=utf-8 +""" +Function: +The replay funtion entry +Copyright Information: +Huawei Technologies Co., Ltd. All Rights Reserved © 2020 +""" + +import os +import stat + + +REPLAY_BATCH = 'batch' +REPLAY_ITERATE = 'iterate' +CFG_IMPL_DIR = 'impl_dir' +CFG_OUT_DIR = 'out_dir' +AUTO_GEN_DIR = 'auto_gen_dir' +WFLAGS = os.O_WRONLY | os.O_CREAT | os.O_TRUNC +WMODES = stat.S_IWUSR | stat.S_IRUSR +SOC_MAP_EXT = {'ascend310p': 'Ascend310P3', 'ascend310b': 'Ascend310B1', + 'ascend910': 'Ascend910A', 'ascend910b': 'Ascend910B1'} +BIN_CMD = 'opc $1 --main_func={fun} --input_param={param} --soc_version={soc} \ +--output=$2 --impl_mode={impl} --simplified_key_mode=0 --op_mode=dynamic\n' +CHK_CMD = ''' +if ! test -f $2/{res_file} ; then + echo "$2/{res_file} not generated!" + exit 1 +fi +''' +ATTR_DEF_VAL = {'str' : '', 'int': 0, 'float': 0.0, 'bool': False, 'list_bool': [], + 'list_int': [], 'list_float': [], 'list_list_int': [[]]} diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh new file mode 100755 index 000000000..55e12e5ed --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh @@ -0,0 +1,57 @@ +#!/usr/bin/bash + +project_path=$1 +build_path=$2 +vendor_name=customize +if [[ ! -d "$project_path" ]]; then + echo "[ERROR] No projcet path is provided" + exit 1 +fi + +if [[ ! -d "$build_path" ]]; then + echo "[ERROR] No build path is provided" + exit 1 +fi + +# copy ai_core operators implements +# tbe_impl_files_num=$(ls $project_path/tbe/impl/* 2> /dev/null | wc -l) +# if [[ "$tbe_impl_files_num" -gt 0 ]];then +# cp -r ${project_path}/tbe/impl/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/ai_core/tbe/customize_impl +# cp -r ${project_path}/tbe/impl/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/vector_core/tbe/customize_impl +# fi + +# copy aicpu kernel so operators +if [[ -d "${project_path}/cpukernel/aicpu_kernel_lib" ]]; then + cp -f ${project_path}/cpukernel/aicpu_kernel_lib/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/cpu/aicpu_kernel/impl + rm -rf ${project_path}/cpukernel/aicpu_kernel_lib +fi + +# merge aicpu.ini and aicore.ini to generate npu_supported_ops.json +# mkdir -p ${build_path}/framework/op_info_cfg +# mkdir -p ${build_path}/framework/op_info_cfg/aicpu_kernel +# mkdir -p ${build_path}/framework/op_info_cfg/ai_core + +# if [[ -d "${project_path}/tbe/op_info_cfg/ai_core" ]]; then +# bash ${project_path}/cmake/util/gen_ops_filter.sh ${project_path}/tbe/op_info_cfg/ai_core ${build_path}/framework/op_info_cfg/ai_core +# fi + +# if [[ -d "${project_path}/cpukernel/op_info_cfg/aicpu_kernel" ]]; then +# bash ${project_path}/cmake/util/gen_ops_filter.sh ${project_path}/cpukernel/op_info_cfg/aicpu_kernel ${build_path}/framework/op_info_cfg/aicpu_kernel +# fi + +# aicpu_filter_file=${build_path}/framework/op_info_cfg/aicpu_kernel/npu_supported_ops.json +# aicore_filter_file=${build_path}/framework/op_info_cfg/ai_core/npu_supported_ops.json +# if [[ -f "${aicpu_filter_file}" ]] && [[ ! -f "${aicore_filter_file}" ]]; then +# cp $aicpu_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow +# fi +# if [[ -f "${aicore_filter_file}" ]] && [[ ! -f "${aicpu_filter_file}" ]]; then +# cp $aicore_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow +# fi + +# if [[ -f "${aicore_filter_file}" ]] && [[ -f "${aicpu_filter_file}" ]]; then +# chmod u+w ${aicpu_filter_file} +# python3 ${project_path}/cmake/util/insert_op_info.py ${aicore_filter_file} ${aicpu_filter_file} +# chmod u-w ${aicpu_filter_file} +# cp $aicpu_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow +# fi + diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh new file mode 100755 index 000000000..d4c27d17f --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh @@ -0,0 +1,62 @@ +#!/bin/bash +# Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +# Description: Generate npu_supported_ops.json +# ============================================================================== + +if [[ -z "$1" ]]; then + echo -e "[ERROR] No source dir provided" + exit 1 +fi + +if [[ -z "$2" ]]; then + echo -e "[ERROR] No destination dir provided" + exit 1 +fi + +src=$1 +dest_file=$2/npu_supported_ops.json + +if [ -f "$dest_file" ];then + chmod u+w $dest_file +fi + +echo $* + +add_ops() { + name=$1 + isHeavy=$2 + file=$3 + grep -w "\"$name\"" ${file} >/dev/null + if [ $? == 0 ];then + return + fi + echo " \"${name}\": {" >> ${file} + echo " \"isGray\": false," >> ${file} + echo " \"isHeavy\": ${isHeavy}" >> ${file} + echo " }," >> ${file} +} + +echo "{" > ${dest_file} +ini_files=$(find ${src} -name "*.ini") +for file in ${ini_files} ; do + name=$(grep '^\[' ${file} | sed 's/\[//g' | sed 's/]//g' | sed 's/\r//g') + grep 'heavyOp.flag' ${file} >/dev/null + if [ $? == 0 ];then + isHeavy=$(grep 'heavyOp.flag' ${file} | awk -F= '{print $2}') + else + isHeavy="false" + fi + for op in ${name} ; do + add_ops ${op} "false" ${dest_file} + done +done +echo "}" >> ${dest_file} +file_count=$(cat ${dest_file} | wc -l) +line=$(($file_count-1)) +sed -i "${line}{s/,//g}" ${dest_file} + +chmod 640 "${dest_file}" +echo -e "[INFO] Succed generated ${dest_file}" + +exit 0 + diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh new file mode 100755 index 000000000..a06cfc78d --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh @@ -0,0 +1,6 @@ +ascend_install_dir=$1 +gen_file_dir=$2 + +# create version.info +compiler_version=$(grep "Version" -w ${ascend_install_dir}/compiler/version.info | awk -F = '{print $2}') +echo "custom_opp_compiler_version=${compiler_version}" > ${gen_file_dir}/version.info \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py new file mode 100755 index 000000000..28ba08757 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py @@ -0,0 +1,36 @@ +# -*- coding: utf-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" +import json +import os +import sys +import stat +import const_var + + +if __name__ == '__main__': + if len(sys.argv) != 3: + print(sys.argv) + print('argv error, inert_op_info.py your_op_file lib_op_file') + sys.exit(2) + + with open(sys.argv[1], 'r') as load_f: + insert_operator = json.load(load_f) + + all_operators = {} + if os.path.exists(sys.argv[2]): + if os.path.getsize(sys.argv[2]) != 0: + with open(sys.argv[2], 'r') as load_f: + all_operators = json.load(load_f) + + for k in insert_operator.keys(): + if k in all_operators.keys(): + print('replace op:[', k, '] success') + else: + print('insert op:[', k, '] success') + all_operators[k] = insert_operator[k] + + with os.fdopen(os.open(sys.argv[2], const_var.WFLAGS, const_var.WMODES), 'w') as json_file: + json_file.write(json.dumps(all_operators, indent=4)) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py new file mode 100755 index 000000000..ace727b90 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py @@ -0,0 +1,248 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os +import re +import glob +import json +import argparse +import const_var + + +DATA_TPYE_DICT = { + 'float32': 0, + 'float16': 1, + 'int8': 2, + 'int16': 6, + 'uint16': 7, + 'uint8': 4, + 'int32': 3, + 'int64': 9, + 'uint32': 8, + 'uint64': 10, + 'bool': 12, + 'double': 11, + 'string': 13, + 'dual': 14, + 'dual': 15, + 'complex64': 16, + 'complex128': 17, + 'qint8': 18, + 'qint16': 19, + 'qint32': 20, + 'quint8': 21, + 'quint16': 22, + 'resource': 23, + 'string': 24, + 'dual': 25, + 'variant': 26, + 'bf16': 27, + 'bfloat16': 27, + 'undefined': 28, + 'int4': 29, + 'uint1': 30, + 'int2': 31 +} + +FORMAT_DICT = { + 'NCHW': 0, + 'NHWC': 1, + 'ND': 2, + 'NC1HWC0': 3, + 'FRACTAL_Z': 4, + 'NC1C0HWPAD': 5, + 'NHWC1C0': 6, + 'FSR_NCHW': 7, + 'FRACTAL_DECONV': 8, + 'C1HWNC0': 9, + 'FRACTAL_DECONV_TRANSPOSE': 10, + 'FRACTAL_DECONV_SP_STRIDE_TRANS': 11, + 'NC1HWC0_C04': 12, + 'FRACTAL_Z_C04': 13, + 'CHWN': 14, + 'FRACTAL_DECONV_SP_STRIDE8_TRANS': 15, + 'HWCN': 16, + 'NC1KHKWHWC0': 17, + 'BN_WEIGHT': 18, + 'FILTER_HWCK': 19, + 'HASHTABLE_LOOKUP_LOOKUPS': 20, + 'HASHTABLE_LOOKUP_KEYS': 21, + 'HASHTABLE_LOOKUP_VALUE': 22, + 'HASHTABLE_LOOKUP_OUTPUT': 23, + 'HASHTABLE_LOOKUP_HITS': 24, + 'C1HWNCoC0': 25, + 'MD': 26, + 'NDHWC': 27, + 'FRACTAL_ZZ': 28, + 'FRACTAL_NZ': 29, + 'NCDHW': 30, + 'DHWCN': 31, + 'NDC1HWC0': 32, + 'FRACTAL_Z_3D': 33, + 'CN': 34, + 'NC': 35, + 'DHWNC': 36, + 'FRACTAL_Z_3D_TRANSPOSE': 37, + 'FRACTAL_ZN_LSTM': 38, + 'FRACTAL_Z_G': 39, + 'RESERVED': 40, + 'ALL': 41, + 'NULL': 42, + 'ND_RNN_BIAS': 43, + 'FRACTAL_ZN_RNN': 44, + 'NYUV': 45, + 'NYUV_A': 46 +} + + +def load_json(json_file: str): + with open(json_file, encoding='utf-8') as file: + json_content = json.load(file) + return json_content + + +def get_specified_suffix_file(root_dir, suffix): + specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) + all_suffix_files = glob.glob(specified_suffix, recursive=True) + return all_suffix_files + + +def get_deterministic_value(support_info): + deterministic_key = 'deterministic' + if deterministic_key not in support_info: + return 0 + deterministic_value = support_info.get(deterministic_key) + if deterministic_value == 'true': + return 1 + else: + return 0 + + +def get_precision_value(support_info): + precision_key = 'implMode' + precision_value = support_info.get(precision_key) + if precision_value == 'high_performance': + _value = 1 + elif precision_value == 'high_precision': + _value = 2 + else: + _value = 0 + return _value + + +def get_overflow_value(support_info): + return 0 + + +def get_parameters(info): + if info: + if 'dtype' in info: + data_type = info['dtype'] + data_type_value = DATA_TPYE_DICT.get(data_type) + else: + data_type_value = 0 + if 'format' in info: + _format = info['format'] + _format_value = FORMAT_DICT.get(_format) + else: + _format_value = 0 + else: + data_type_value = 0 + _format_value = 0 + return str(data_type_value), str(_format_value) + + +def get_dynamic_parameters(info): + # 动态输入时只需获取第一个参数 + return get_parameters(info[0]) + + +def get_all_parameters(support_info, _type): + result_list = list() + info_lists = support_info.get(_type) + if info_lists: + for _info in info_lists: + # 输入为列表时是动态输入 + if isinstance(_info, (list, tuple)): + data_type_value, _format_value = get_dynamic_parameters(_info) + else: + data_type_value, _format_value = get_parameters(_info) + result_list.append("{},{}".format(data_type_value, _format_value)) + return result_list + + +def get_all_input_parameters(support_info): + result = get_all_parameters(support_info, 'inputs') + return '/'.join(result) + + +def insert_content_into_file(input_file, content): + with open(input_file, 'r+') as file: + lines = file.readlines() + for index, line in enumerate(lines): + match_result = re.search(r'"staticKey":', line) + if match_result: + count = len(line) - len(line.lstrip()) + new_content = "{}{}".format(' ' * count, content) + # 插入到前一行,防止插入最后时还需要考虑是否添加逗号 + lines.insert(index, new_content) + break + file.seek(0) + file.write(''.join(lines)) + + +def insert_simplified_keys(json_file): + contents = load_json(json_file) + # 不存在'binFileName'或者'supportInfo'字段时,非需要替换的解析json文件 + if ('binFileName' not in contents) or ('supportInfo' not in contents): + return + support_info = contents.get('supportInfo') + bin_file_name = contents.get('binFileName') + bin_suffix = contents.get('binFileSuffix') + # 'simplifiedKey'字段已经存在时,直接返回,不重复生成 + if 'simplifiedKey' in support_info: + return + op_type = bin_file_name.split('_')[0] + deterministic = str(get_deterministic_value(support_info)) + precision = str(get_precision_value(support_info)) + overflow = str(get_overflow_value(support_info)) + input_parameters = get_all_input_parameters(support_info) + key = '{}/d={},p={},o={}/{}/'.format( + op_type, + deterministic, + precision, + overflow, + input_parameters) + result = '"simplifiedKey": "' + key + '",\n' + insert_content_into_file(json_file, result) + + +def insert_all_simplified_keys(root_dir): + suffix = 'json' + all_json_files = get_specified_suffix_file(root_dir, suffix) + for _json in all_json_files: + insert_simplified_keys(_json) + + +def args_prase(): + parser = argparse.ArgumentParser() + parser.add_argument('-p', + '--path', + nargs='?', + required=True, + help='Parse the path of the json file.') + return parser.parse_args() + + +def main(): + args = args_prase() + insert_all_simplified_keys(args.path) + + +if __name__ == '__main__': + main() diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py new file mode 100755 index 000000000..2b77c970d --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py @@ -0,0 +1,115 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + + +def gen_fun_def(title, kernel, argn, arg_type, arg_name): + entry = [] + entry.append(title) + entry.append(kernel) + entry.append('(') + args = [] + for i in range(0, argn): + args.append(arg_type + ' ' + arg_name + str(i)) + entry.append(', '.join(args)) + entry.append(')') + return ' '.join(entry) + + +def gen_batch_kernel_body(fname, argn, arg_name): + body = [] + body.append('{') + fun = [] + fun.append(fname) + fun.append('(') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + fun.append(', '.join(args)) + fun.append(');') + body.append(' '.join(fun)) + body.append('}') + return '\n'.join(body) + + +def gen_mc_kernel_body(kn, argn, arg_name, blknum): + body = [] + body.append('{') + body.append(' switch(block_idx) {') + for blk in range(0, blknum): + fun = [] + fun.append('{}_blk{:02d}'.format(kn, blk)) + fun.append('(') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + fun.append(', '.join(args)) + fun.append(')') + body.append(' case {}: {}; break;'.format(blk, ' '.join(fun))) + body.append(' default: break;') + body.append(' }') + body.append('}') + return '\n'.join(body) + + +def gen_proc_body(argn, arg_name): + body = [] + body.append('{') + args = [] + for i in range(0, argn): + args.append(arg_name + str(i)) + body.append('uint64_t __x = (uint64_t)' + ' + (uint64_t)'.join(args) + ';') + body.append('__asm__ ("NOP");') + body.append('__asm__ ("NOP");') + body.append('__asm__ ("NOP");') + body.append('}') + return '\n'.join(body) + + +def batch_code_gen(kn, argn, argt): + codes = [] + kernel_name = kn + proc_name = kernel_name + '_percore' + arg_num = int(argn) + data_type = argt + arg_type = '__gm__ ' + data_type + '* __restrict__' + arg_name = 'arg' + kernel_title = 'extern \"C\" __global__ __aicore__ void' + proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' + codes.append('#ifndef __aicore__') + codes.append('#define __aicore__ [aicore]') + codes.append('#endif') + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') + codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) + codes.append(gen_batch_kernel_body(proc_name, arg_num, arg_name)) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) + codes.append(gen_proc_body(arg_num, arg_name)) + return '\n'.join(codes) + '\n' + + +def mc_code_gen(kn, argn, argt, blknum): + codes = [] + kernel_name = kn + core_num = int(blknum) + arg_num = int(argn) + data_type = argt + arg_type = '__gm__ ' + data_type + '* __restrict__' + arg_name = 'arg' + kernel_title = 'extern \"C\" __global__ __aicore__ void' + proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' + codes.append('#ifndef __aicore__') + codes.append('#define __aicore__ [aicore]') + codes.append('#endif') + for i in range(0, core_num): + proc_name = '{}_blk{:02d}'.format(kernel_name, i) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') + codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) + codes.append(gen_mc_kernel_body(kernel_name, arg_num, arg_name, core_num)) + for i in range(0, core_num): + proc_name = '{}_blk{:02d}'.format(kernel_name, i) + codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) + codes.append(gen_proc_body(arg_num, arg_name)) + return '\n'.join(codes) + '\n' diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp b/atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp new file mode 100755 index 000000000..5079a1043 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp @@ -0,0 +1,10 @@ +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#define __ASCENDC_REPLAY_CODE__ +#include "__CCE_FILE__" diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING new file mode 100644 index 000000000..d159169d1 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING @@ -0,0 +1,339 @@ + GNU GENERAL PUBLIC LICENSE + Version 2, June 1991 + + Copyright (C) 1989, 1991 Free Software Foundation, Inc., + 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + Everyone is permitted to copy and distribute verbatim copies + of this license document, but changing it is not allowed. + + Preamble + + The licenses for most software are designed to take away your +freedom to share and change it. By contrast, the GNU General Public +License is intended to guarantee your freedom to share and change free +software--to make sure the software is free for all its users. This +General Public License applies to most of the Free Software +Foundation's software and to any other program whose authors commit to +using it. (Some other Free Software Foundation software is covered by +the GNU Lesser General Public License instead.) You can apply it to +your programs, too. + + When we speak of free software, we are referring to freedom, not +price. Our General Public Licenses are designed to make sure that you +have the freedom to distribute copies of free software (and charge for +this service if you wish), that you receive source code or can get it +if you want it, that you can change the software or use pieces of it +in new free programs; and that you know you can do these things. + + To protect your rights, we need to make restrictions that forbid +anyone to deny you these rights or to ask you to surrender the rights. +These restrictions translate to certain responsibilities for you if you +distribute copies of the software, or if you modify it. + + For example, if you distribute copies of such a program, whether +gratis or for a fee, you must give the recipients all the rights that +you have. You must make sure that they, too, receive or can get the +source code. And you must show them these terms so they know their +rights. + + We protect your rights with two steps: (1) copyright the software, and +(2) offer you this license which gives you legal permission to copy, +distribute and/or modify the software. + + Also, for each author's protection and ours, we want to make certain +that everyone understands that there is no warranty for this free +software. If the software is modified by someone else and passed on, we +want its recipients to know that what they have is not the original, so +that any problems introduced by others will not reflect on the original +authors' reputations. + + Finally, any free program is threatened constantly by software +patents. We wish to avoid the danger that redistributors of a free +program will individually obtain patent licenses, in effect making the +program proprietary. To prevent this, we have made it clear that any +patent must be licensed for everyone's free use or not licensed at all. + + The precise terms and conditions for copying, distribution and +modification follow. + + GNU GENERAL PUBLIC LICENSE + TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION + + 0. This License applies to any program or other work which contains +a notice placed by the copyright holder saying it may be distributed +under the terms of this General Public License. The "Program", below, +refers to any such program or work, and a "work based on the Program" +means either the Program or any derivative work under copyright law: +that is to say, a work containing the Program or a portion of it, +either verbatim or with modifications and/or translated into another +language. (Hereinafter, translation is included without limitation in +the term "modification".) Each licensee is addressed as "you". + +Activities other than copying, distribution and modification are not +covered by this License; they are outside its scope. The act of +running the Program is not restricted, and the output from the Program +is covered only if its contents constitute a work based on the +Program (independent of having been made by running the Program). +Whether that is true depends on what the Program does. + + 1. You may copy and distribute verbatim copies of the Program's +source code as you receive it, in any medium, provided that you +conspicuously and appropriately publish on each copy an appropriate +copyright notice and disclaimer of warranty; keep intact all the +notices that refer to this License and to the absence of any warranty; +and give any other recipients of the Program a copy of this License +along with the Program. + +You may charge a fee for the physical act of transferring a copy, and +you may at your option offer warranty protection in exchange for a fee. + + 2. You may modify your copy or copies of the Program or any portion +of it, thus forming a work based on the Program, and copy and +distribute such modifications or work under the terms of Section 1 +above, provided that you also meet all of these conditions: + + a) You must cause the modified files to carry prominent notices + stating that you changed the files and the date of any change. + + b) You must cause any work that you distribute or publish, that in + whole or in part contains or is derived from the Program or any + part thereof, to be licensed as a whole at no charge to all third + parties under the terms of this License. + + c) If the modified program normally reads commands interactively + when run, you must cause it, when started running for such + interactive use in the most ordinary way, to print or display an + announcement including an appropriate copyright notice and a + notice that there is no warranty (or else, saying that you provide + a warranty) and that users may redistribute the program under + these conditions, and telling the user how to view a copy of this + License. (Exception: if the Program itself is interactive but + does not normally print such an announcement, your work based on + the Program is not required to print an announcement.) + +These requirements apply to the modified work as a whole. If +identifiable sections of that work are not derived from the Program, +and can be reasonably considered independent and separate works in +themselves, then this License, and its terms, do not apply to those +sections when you distribute them as separate works. But when you +distribute the same sections as part of a whole which is a work based +on the Program, the distribution of the whole must be on the terms of +this License, whose permissions for other licensees extend to the +entire whole, and thus to each and every part regardless of who wrote it. + +Thus, it is not the intent of this section to claim rights or contest +your rights to work written entirely by you; rather, the intent is to +exercise the right to control the distribution of derivative or +collective works based on the Program. + +In addition, mere aggregation of another work not based on the Program +with the Program (or with a work based on the Program) on a volume of +a storage or distribution medium does not bring the other work under +the scope of this License. + + 3. You may copy and distribute the Program (or a work based on it, +under Section 2) in object code or executable form under the terms of +Sections 1 and 2 above provided that you also do one of the following: + + a) Accompany it with the complete corresponding machine-readable + source code, which must be distributed under the terms of Sections + 1 and 2 above on a medium customarily used for software interchange; or, + + b) Accompany it with a written offer, valid for at least three + years, to give any third party, for a charge no more than your + cost of physically performing source distribution, a complete + machine-readable copy of the corresponding source code, to be + distributed under the terms of Sections 1 and 2 above on a medium + customarily used for software interchange; or, + + c) Accompany it with the information you received as to the offer + to distribute corresponding source code. (This alternative is + allowed only for noncommercial distribution and only if you + received the program in object code or executable form with such + an offer, in accord with Subsection b above.) + +The source code for a work means the preferred form of the work for +making modifications to it. For an executable work, complete source +code means all the source code for all modules it contains, plus any +associated interface definition files, plus the scripts used to +control compilation and installation of the executable. However, as a +special exception, the source code distributed need not include +anything that is normally distributed (in either source or binary +form) with the major components (compiler, kernel, and so on) of the +operating system on which the executable runs, unless that component +itself accompanies the executable. + +If distribution of executable or object code is made by offering +access to copy from a designated place, then offering equivalent +access to copy the source code from the same place counts as +distribution of the source code, even though third parties are not +compelled to copy the source along with the object code. + + 4. You may not copy, modify, sublicense, or distribute the Program +except as expressly provided under this License. Any attempt +otherwise to copy, modify, sublicense or distribute the Program is +void, and will automatically terminate your rights under this License. +However, parties who have received copies, or rights, from you under +this License will not have their licenses terminated so long as such +parties remain in full compliance. + + 5. You are not required to accept this License, since you have not +signed it. However, nothing else grants you permission to modify or +distribute the Program or its derivative works. These actions are +prohibited by law if you do not accept this License. Therefore, by +modifying or distributing the Program (or any work based on the +Program), you indicate your acceptance of this License to do so, and +all its terms and conditions for copying, distributing or modifying +the Program or works based on it. + + 6. Each time you redistribute the Program (or any work based on the +Program), the recipient automatically receives a license from the +original licensor to copy, distribute or modify the Program subject to +these terms and conditions. You may not impose any further +restrictions on the recipients' exercise of the rights granted herein. +You are not responsible for enforcing compliance by third parties to +this License. + + 7. If, as a consequence of a court judgment or allegation of patent +infringement or for any other reason (not limited to patent issues), +conditions are imposed on you (whether by court order, agreement or +otherwise) that contradict the conditions of this License, they do not +excuse you from the conditions of this License. If you cannot +distribute so as to satisfy simultaneously your obligations under this +License and any other pertinent obligations, then as a consequence you +may not distribute the Program at all. For example, if a patent +license would not permit royalty-free redistribution of the Program by +all those who receive copies directly or indirectly through you, then +the only way you could satisfy both it and this License would be to +refrain entirely from distribution of the Program. + +If any portion of this section is held invalid or unenforceable under +any particular circumstance, the balance of the section is intended to +apply and the section as a whole is intended to apply in other +circumstances. + +It is not the purpose of this section to induce you to infringe any +patents or other property right claims or to contest validity of any +such claims; this section has the sole purpose of protecting the +integrity of the free software distribution system, which is +implemented by public license practices. Many people have made +generous contributions to the wide range of software distributed +through that system in reliance on consistent application of that +system; it is up to the author/donor to decide if he or she is willing +to distribute software through any other system and a licensee cannot +impose that choice. + +This section is intended to make thoroughly clear what is believed to +be a consequence of the rest of this License. + + 8. If the distribution and/or use of the Program is restricted in +certain countries either by patents or by copyrighted interfaces, the +original copyright holder who places the Program under this License +may add an explicit geographical distribution limitation excluding +those countries, so that distribution is permitted only in or among +countries not thus excluded. In such case, this License incorporates +the limitation as if written in the body of this License. + + 9. The Free Software Foundation may publish revised and/or new versions +of the General Public License from time to time. Such new versions will +be similar in spirit to the present version, but may differ in detail to +address new problems or concerns. + +Each version is given a distinguishing version number. If the Program +specifies a version number of this License which applies to it and "any +later version", you have the option of following the terms and conditions +either of that version or of any later version published by the Free +Software Foundation. If the Program does not specify a version number of +this License, you may choose any version ever published by the Free Software +Foundation. + + 10. If you wish to incorporate parts of the Program into other free +programs whose distribution conditions are different, write to the author +to ask for permission. For software which is copyrighted by the Free +Software Foundation, write to the Free Software Foundation; we sometimes +make exceptions for this. Our decision will be guided by the two goals +of preserving the free status of all derivatives of our free software and +of promoting the sharing and reuse of software generally. + + NO WARRANTY + + 11. BECAUSE THE PROGRAM IS LICENSED FREE OF CHARGE, THERE IS NO WARRANTY +FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN +OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES +PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED +OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF +MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS +TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE +PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, +REPAIR OR CORRECTION. + + 12. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING +WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MAY MODIFY AND/OR +REDISTRIBUTE THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, +INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING +OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED +TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY +YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER +PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE +POSSIBILITY OF SUCH DAMAGES. + + END OF TERMS AND CONDITIONS + + How to Apply These Terms to Your New Programs + + If you develop a new program, and you want it to be of the greatest +possible use to the public, the best way to achieve this is to make it +free software which everyone can redistribute and change under these terms. + + To do so, attach the following notices to the program. It is safest +to attach them to the start of each source file to most effectively +convey the exclusion of warranty; and each file should have at least +the "copyright" line and a pointer to where the full notice is found. + + + Copyright (C) + + This program is free software; you can redistribute it and/or modify + it under the terms of the GNU General Public License as published by + the Free Software Foundation; either version 2 of the License, or + (at your option) any later version. + + This program is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + GNU General Public License for more details. + + You should have received a copy of the GNU General Public License along + with this program; if not, write to the Free Software Foundation, Inc., + 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. + +Also add information on how to contact you by electronic and paper mail. + +If the program is interactive, make it output a short notice like this +when it starts in an interactive mode: + + Gnomovision version 69, Copyright (C) year name of author + Gnomovision comes with ABSOLUTELY NO WARRANTY; for details type `show w'. + This is free software, and you are welcome to redistribute it + under certain conditions; type `show c' for details. + +The hypothetical commands `show w' and `show c' should show the appropriate +parts of the General Public License. Of course, the commands you use may +be called something other than `show w' and `show c'; they could even be +mouse-clicks or menu items--whatever suits your program. + +You should also get your employer (if you work as a programmer) or your +school, if any, to sign a "copyright disclaimer" for the program, if +necessary. Here is a sample; alter the names: + + Yoyodyne, Inc., hereby disclaims all copyright interest in the program + `Gnomovision' (which makes passes at compilers) written by James Hacker. + + , 1 April 1989 + Ty Coon, President of Vice + +This General Public License does not permit incorporating your program into +proprietary programs. If your program is a subroutine library, you may +consider it more useful to permit linking proprietary applications with the +library. If this is what you want to do, use the GNU Lesser General +Public License instead of this License. diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/README.md b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/README.md new file mode 100644 index 000000000..b41f01682 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/README.md @@ -0,0 +1,246 @@ +[![License: GPL v2](https://img.shields.io/badge/License-GPL%20v2-blue.svg)](https://www.gnu.org/licenses/old-licenses/gpl-2.0.en.html) +![Build Status](https://github.com/megastep/makeself/workflows/CI/badge.svg) + +# makeself - Make self-extractable archives on Unix + +[makeself.sh][1] is a small shell script that generates a self-extractable +compressed tar archive from a directory. The resulting file appears as a shell script +(many of those have a **.run** suffix), and can be launched as is. The archive +will then uncompress itself to a temporary directory and an optional arbitrary +command will be executed (for example an installation script). This is pretty +similar to archives generated with WinZip Self-Extractor in the Windows world. +Makeself archives also include checksums for integrity self-validation (CRC +and/or MD5/SHA256 checksums). + +The makeself.sh script itself is used only to create the archives from a +directory of files. The resultant archive is actually a compressed (using +gzip, bzip2, or compress) TAR archive, with a small shell script stub at the +beginning. This small stub performs all the steps of extracting the files, +running the embedded command, and removing the temporary files when done. +All the user has to do to install the software contained in such an +archive is to "run" the archive, i.e **sh nice-software.run**. I recommend +using the ".run" (which was introduced by some Makeself archives released by +Loki Software) or ".sh" suffix for such archives not to confuse the users, +so that they will know they are actually shell scripts (with quite a lot of binary data +attached to them though!). + +I am trying to keep the code of this script as portable as possible, i.e it is +not relying on any bash-specific features and only calls commands that are +installed on any functioning UNIX-compatible system. This script as well as +the archives it generates should run on any Unix flavor, with any compatible +Bourne shell, provided of course that the compression programs are available. + +As of version 2.1, Makeself has been rewritten and tested on the following +platforms : + + * Linux (all distributions) + * Sun Solaris (8 and above) + * HP-UX (tested on 11.0 and 11i on HPPA RISC) + * SCO OpenUnix and OpenServer + * IBM AIX 5.1L + * macOS (Darwin) + * SGI IRIX 6.5 + * FreeBSD + * UnicOS / Cray + * Cygwin (Windows) + +If you successfully run Makeself and/or archives created with it on another +system, then please [let me know][2]! + +Examples of publicly available archives made using makeself are : + + * Game patches and installers for [Id Software][3] games like Quake 3 for Linux or Return To Castle Wolfenstein ; + * All game patches released by [Loki Software][4] for the Linux version of popular games ; + * The [nVidia drivers][5] for Linux + * The installer for the Linux version of [Google Earth][6] + * The [VirtualBox][7] installers for Linux + * The [Makeself][1] distribution itself ;-) + * and countless others... + +**Important note for Apache users:** By default, most Web servers will think that Makeself archives are regular text files and thus they may show up as text in a Web browser. The correct way to prevent this is to add a MIME type for this file format, like so (in httpd.conf) : + +`AddType application/x-makeself .run` + +**Important note for certain GNU/Linux distributions:** Archives created with Makeself prior to v2.1.2 were using an old syntax for the _head_ and _tail_ Unix commands that is being progressively obsoleted in their GNU forms. Therefore you may have problems uncompressing some of these archives. A workaround for this is to set the environment variable $_POSIX2_VERSION to enable the old syntax, i.e. : + +`export _POSIX2_VERSION=199209` + +## Usage + +The syntax of makeself is the following: + +``` +makeself.sh [args] archive_dir file_name label startup_script [script_args] +``` + + * _args_ are optional options for Makeself. The available ones are : + + * **`--version`** : Prints the version number on stdout, then exits immediately + * **`--gzip`** : Use gzip for compression (the default on platforms on which gzip is commonly available, like Linux) + * **`--bzip2`** : Use bzip2 instead of gzip for better compression. The bzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. + * **`--pbzip2`** : Use pbzip2 instead of gzip for better and faster compression on machines having multiple CPUs. The pbzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. + * **`--xz`** : Use xz instead of gzip for better compression. The xz command must be available in the command path. It is recommended that the archive prefix be set to something like '.xz.run' for the archive, so that potential users know that they'll need xz to extract it. + * **`--lzo`** : Use lzop instead of gzip for better compression. The lzop command must be available in the command path. It is recommended that the archive prefix be set to something like `.lzo.run` for the archive, so that potential users know that they'll need lzop to extract it. + * **`--lz4`** : Use lz4 instead of gzip for better compression. The lz4 command must be available in the command path. It is recommended that the archive prefix be set to something like '.lz4.run' for the archive, so that potential users know that they'll need lz4 to extract it. + * **`--zstd`** : Use zstd instead of gzip for better compression. The zstd command must be available in the command path. It is recommended that the archive prefix be set to something like '.zstd.run' for the archive, so that potential users know that they'll need zstd to extract it. + * **`--pigz`** : Use pigz for compression. + * **`--base64`** : Encode the archive to ASCII in Base64 format instead of compressing (base64 command required). + * **`--gpg-encrypt`** : Encrypt the archive using `gpg -ac -z $COMPRESS_LEVEL`. This will prompt for a password to encrypt with. Assumes that potential users have `gpg` installed. + * **`--ssl-encrypt`** : Encrypt the archive using `openssl aes-256-cbc -a -salt`. This will prompt for a password to encrypt with. Assumes that the potential users have the OpenSSL tools installed. + * **`--compress`** : Use the UNIX `compress` command to compress the data. This should be the default on all platforms that don't have gzip available. + * **`--nocomp`** : Do not use any compression for the archive, which will then be an uncompressed TAR. + * **`--complevel`** : Specify the compression level for gzip, bzip2, pbzip2, zstd, xz, lzo or lz4. (defaults to 9) + * **`--threads`** : Specify the number of threads to be used by compressors that support parallelization. Omit to use compressor's default. Most useful (and required) for opting into xz's threading, usually with `--threads=0` for all available cores. pbzip2 and pigz are parallel by default, and setting this value allows limiting the number of threads they use. + * **`--notemp`** : The generated archive will not extract the files to a temporary directory, but in a new directory created in the current directory. This is better to distribute software packages that may extract and compile by themselves (i.e. launch the compilation through the embedded script). + * **`--current`** : Files will be extracted to the current directory, instead of in a subdirectory. This option implies `--notemp` above. + * **`--follow`** : Follow the symbolic links inside of the archive directory, i.e. store the files that are being pointed to instead of the links themselves. + * **`--append`** _(new in 2.1.x)_: Append data to an existing archive, instead of creating a new one. In this mode, the settings from the original archive are reused (compression type, label, embedded script), and thus don't need to be specified again on the command line. + * **`--header`** : Makeself uses a separate file to store the header stub, called `makeself-header.sh`. By default, it is assumed that it is stored in the same location as makeself.sh. This option can be used to specify its actual location if it is stored someplace else. + * **`--cleanup`** : Specify a script that is run when execution is interrupted or finishes successfully. The script is executed with the same environment and initial `script_args` as `startup_script`. + * **`--copy`** : Upon extraction, the archive will first extract itself to a temporary directory. The main application of this is to allow self-contained installers stored in a Makeself archive on a CD, when the installer program will later need to unmount the CD and allow a new one to be inserted. This prevents "Filesystem busy" errors for installers that span multiple CDs. + * **`--nox11`** : Disable the automatic spawning of a new terminal in X11. + * **`--nowait`** : When executed from a new X11 terminal, disable the user prompt at the end of the script execution. + * **`--nomd5`** and **`--nocrc`** : Disable the creation of a MD5 / CRC checksum for the archive. This speeds up the extraction process if integrity checking is not necessary. + * **`--sha256`** : Adds a SHA256 checksum for the archive. This is in addition to the MD5 / CRC checksums unless `--nomd5` is also used. + * **`--lsm` _file_** : Provide and LSM file to makeself, that will be embedded in the generated archive. LSM files are describing a software package in a way that is easily parseable. The LSM entry can then be later retrieved using the `--lsm` argument to the archive. An example of a LSM file is provided with Makeself. + * **`--tar-format opt`** : Specify the tar archive format (default is ustar); you may use any value accepted by your tar command (such as posix, v7, etc). + * **`--tar-extra opt`** : Append more options to the tar command line. + + For instance, in order to exclude the `.git` directory from the packaged archive directory using the GNU `tar`, one can use `makeself.sh --tar-extra "--exclude=.git" ...` + + * **`--keep-umask`** : Keep the umask set to shell default, rather than overriding when executing self-extracting archive. + * **`--packaging-date date`** : Use provided string as the packaging date instead of the current date. + * **`--license`** : Append a license file. + * **`--nooverwrite`** : Do not extract the archive if the specified target directory already exists. + * **`--help-header file`** : Add a header to the archive's `--help` output. + * `archive_dir` is the name of the directory that contains the files to be archived + * `file_name` is the name of the archive to be created + * `label` is an arbitrary text string describing the package. It will be displayed while extracting the files. + * `startup_script` is the command to be executed _from within_ the directory of extracted files. Thus, if you wish to execute a program contained in this directory, you must prefix your command with `./`. For example, `./program` will be fine. The `script_args` are additional arguments for this command. + +Here is an example, assuming the user has a package image stored in a **/home/joe/mysoft**, and he wants to generate a self-extracting package named +**mysoft.sh**, which will launch the "setup" script initially stored in /home/joe/mysoft : + +`makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup +` + +Here is also how I created the [makeself.run][9] archive which contains the Makeself distribution : + +`makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" ` + +Archives generated with Makeself can be passed the following arguments: + + * **`--keep`** : Prevent the files to be extracted in a temporary directory that will be removed after the embedded script's execution. The files will then be extracted in the current working directory and will stay here until you remove them. + * **`--verbose`** : Will prompt the user before executing the embedded command + * **`--target dir`** : Allows to extract the archive in an arbitrary place. + * **`--nox11`** : Do not spawn a X11 terminal. + * **`--confirm`** : Prompt the user for confirmation before running the embedded command. + * **`--info`** : Print out general information about the archive (does not extract). + * **`--lsm`** : Print out the LSM entry, if it is present. + * **`--list`** : List the files in the archive. + * **`--check`** : Check the archive for integrity using the embedded checksums. Does not extract the archive. + * **`--nochown`** : By default, a `chown -R` command is run on the target directory after extraction, so that all files belong to the current user. This is mostly needed if you are running as root, as tar will then try to recreate the initial user ownerships. You may disable this behavior with this flag. + * **`--tar`** : Run the tar command on the contents of the archive, using the following arguments as parameter for the command. + * **`--noexec`** : Do not run the embedded script after extraction. + * **`--noexec-cleanup`** : Do not run the embedded cleanup script. + * **`--nodiskspace`** : Do not check for available disk space before attempting to extract. + * **`--cleanup-args`** : Specify arguments to be passed to the cleanup script. Wrap value in quotes to specify multiple arguments. + +Any subsequent arguments to the archive will be passed as additional arguments to the embedded command. You must explicitly use the `--` special command-line construct before any such options to make sure that Makeself will not try to interpret them. + +## Startup Script + +The startup script must be a regular Shell script. + +Within the startup script, you can use the `$USER_PWD` variable to get the path of the folder from which the self-extracting script is executed. This is especially useful to access files that are located in the same folder as the script, as shown in the example below. + +`my-self-extracting-script.sh --fooBarFileParameter foo.bar` + +## Building and Testing + +Clone the git repo and execute `git submodule update --init --recursive` to obtain all submodules. + +* To make a release: `make` +* To run all tests: `make test` + +## Maven Usage + +Makeself is now supported by the following maven plugin [makeself-maven-plugin](https://github.com/hazendaz/makeself-maven-plugin). Please refer to project for usage and report any bugs in regards to maven plugin on that project. + +## License + +Makeself itself is covered by the [GNU General Public License][8] (GPL) version 2 and above. Archives generated by Makeself don't have to be placed under this license (although I encourage it ;-)), since the archive itself is merely data for Makeself. + +## Contributing + +I will gladly consider merging your pull requests on the [GitHub][10] repository. However, please keep the following in mind: + + * One of the main purposes of Makeself is portability. Do not submit patches that will break supported platforms. The more platform-agnostic, the better. + * Please explain clearly what the purpose of the patch is, and how you achieved it. + +## Download + +Get the latest official distribution [here][9] (version 2.4.2). + +The latest development version can be grabbed from [GitHub][10]. Feel free to submit any patches there through the fork and pull request process. + +## Version history + + * **v1.0:** Initial public release + * **v1.1:** The archive can be passed parameters that will be passed on to the embedded script, thanks to John C. Quillan + * **v1.2:** Cosmetic updates, support for bzip2 compression and non-temporary archives. Many ideas thanks to Francois Petitjean. + * **v1.3:** More patches from Bjarni R. Einarsson and Francois Petitjean: Support for no compression (`--nocomp`), script is no longer mandatory, automatic launch in an xterm, optional verbose output, and -target archive option to indicate where to extract the files. + * **v1.4:** Many patches from Francois Petitjean: improved UNIX compatibility, automatic integrity checking, support of LSM files to get info on the package at run time.. + * **v1.5.x:** A lot of bugfixes, and many other patches, including automatic verification through the usage of checksums. Version 1.5.5 was the stable release for a long time, even though the Web page didn't get updated ;-). Makeself was also officially made a part of the [Loki Setup installer][11], and its source is being maintained as part of this package. + * **v2.0:** Complete internal rewrite of Makeself. The command-line parsing was vastly improved, the overall maintenance of the package was greatly improved by separating the stub from makeself.sh. Also Makeself was ported and tested to a variety of Unix platforms. + * **v2.0.1:** First public release of the new 2.0 branch. Prior versions are officially obsoleted. This release introduced the `--copy` argument that was introduced in response to a need for the [UT2K3][12] Linux installer. + * **v2.1.0:** Big change : Makeself can now support multiple embedded tarballs, each stored separately with their own checksums. An existing archive can be updated with the `--append` flag. Checksums are also better managed, and the `--nochown` option for archives appeared. + * **v2.1.1:** Fixes related to the Unix compression (compress command). Some Linux distributions made the insane choice to make it unavailable, even though gzip is capable of uncompressing these files, plus some more bugfixes in the extraction and checksum code. + * **v2.1.2:** Some bug fixes. Use head -n to avoid problems with POSIX conformance. + * **v2.1.3:** Bug fixes with the command line when spawning terminals. Added `--tar`, `--noexec` for archives. Added `--nomd5` and `--nocrc` to avoid creating checksums in archives. The embedded script is now run through "eval". The `--info` output now includes the command used to create the archive. A man page was contributed by Bartosz Fenski. + * **v2.1.4:** Fixed `--info` output. Generate random directory name when extracting files to . to avoid problems. Better handling of errors with wrong permissions for the directory containing the files. Avoid some race conditions, Unset the $CDPATH variable to avoid problems if it is set. Better handling of dot files in the archive directory. + * **v2.1.5:** Made the md5sum detection consistent with the header code. Check for the presence of the archive directory. Added `--encrypt` for symmetric encryption through gpg (Eric Windisch). Added support for the digest command on Solaris 10 for MD5 checksums. Check for available disk space before extracting to the target directory (Andreas Schweitzer). Allow extraction to run asynchronously (patch by Peter Hatch). Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo). + * **v2.1.6:** Replaced one dot per file progress with a realtime progress percentage and a spinning cursor. Added `--noprogress` to prevent showing the progress during the decompression. Added `--target` dir to allow extracting directly to a target directory. (Guy Baconniere) + * **v2.2.0:** First major new release in years! Includes many bugfixes and user contributions. Please look at the [project page on Github][10] for all the details. + * **v2.3.0:** Support for archive encryption via GPG or OpenSSL. Added LZO and LZ4 compression support. Options to set the packaging date and stop the umask from being overriden. Optionally ignore check for available disk space when extracting. New option to check for root permissions before extracting. + * **v2.3.1:** Various compatibility updates. Added unit tests for Travis CI in the GitHub repo. New `--tar-extra`, `--untar-extra`, `--gpg-extra`, `--gpg-asymmetric-encrypt-sign` options. + * **v2.4.0:** Added optional support for SHA256 archive integrity checksums. + * **v2.4.2:** New --cleanup and --cleanup-args arguments for cleanup scripts. Added threading support for supported compressors. Now supports zstd compression. + * **v2.4.3:** Make explicit POSIX tar archives for increased compatibility. + * **v2.4.4:** Fixed various compatibility issues (no longer use POSIX tar archives), Github Actions to check on Solaris and FreeBSD. + * **v2.4.5:** Added `--tar-format` option to set the tar archive format (default is ustar) + +## Links + + * Check out the ["Loki Setup"][11] installer, used to install many Linux games and other applications, and of which I am the co-author. Since the demise of Loki, I am now the official maintainer of the project, and it is now being hosted here on GitHub. + * Bjarni R. Einarsson also wrote the **setup.sh** installer script, inspired by Makeself. [Check it out !][14] + +## Contact + +This script was written by [Stéphane Peter][15] (megastep at megastep.org). Any enhancements and suggestions are welcome. + +Contributions were included from John C. Quillan, Bjarni R. Einarsson, +Francois Petitjean, Ryan C. Gordon, and many contributors on GitHub. If you think I forgot +your name, don't hesitate to contact me. + +This project is now hosted on GitHub. Feel free to submit patches and bug reports on the [project page][10]. + +* * * + +[Stephane Peter][2] + + [1]: http://makeself.io/ + [2]: mailto:megastep@megastep.org + [3]: http://www.idsoftware.com/ + [4]: http://www.lokigames.com/products/myth2/updates.php3 + [5]: http://www.nvidia.com/ + [6]: http://earth.google.com/ + [7]: http://www.virtualbox.org/ + [8]: http://www.gnu.org/copyleft/gpl.html + [9]: https://github.com/megastep/makeself/releases/download/release-2.4.5/makeself-2.4.5.run + [10]: https://github.com/megastep/makeself + [11]: https://github.com/megastep/loki_setup/ + [12]: http://www.unrealtournament2003.com/ + [13]: http://www.icculus.org/ + [14]: http://bre.klaki.net/programs/setup.sh/ + [15]: https://stephanepeter.com/ diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION new file mode 100644 index 000000000..59aa62c1f --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION @@ -0,0 +1 @@ +2.4.5 diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh new file mode 100644 index 000000000..b5692d490 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh @@ -0,0 +1,9 @@ +#!/bin/sh +# +# Create a distributable archive of the current version of Makeself + +VER=`cat VERSION` +mkdir -p /tmp/makeself-$VER release +cp -pPR makeself* test README.md COPYING VERSION .gitmodules /tmp/makeself-$VER/ +./makeself.sh --notemp /tmp/makeself-$VER release/makeself-$VER.run "Makeself v$VER" echo "Makeself has extracted itself" + diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh new file mode 100644 index 000000000..940903148 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh @@ -0,0 +1,660 @@ +cat << EOF > "$archname" +#!/bin/bash +# This script was generated using Makeself $MS_VERSION +# The license covering this archive and its contents, if any, is wholly independent of the Makeself license (GPL) +# 2022.3.19-Modified the MS_Help function and some options +# Huawei Technologies Co., Ltd. + +ORIG_UMASK=\`umask\` + +CRCsum="$CRCsum" +MD5="$MD5sum" +SHA="$SHAsum" +SIGNATURE="$Signature" +TMPROOT=\${TMPDIR:="\$HOME"} +if ! test -d "\$TMPROOT"; then + TMPROOT="\$PWD" +fi +export TMPDIR="\$TMPROOT" +USER_PWD="\$PWD" +if ! test -d "\$USER_PWD"; then + exit 1 +fi +export USER_PWD +ARCHIVE_DIR=\`dirname "\$0"\` +export ARCHIVE_DIR + +name_of_file="\$0 " +pwd_of_file="\$PWD" +label="$LABEL" +script="$SCRIPT" +scriptargs="$SCRIPTARGS" +cleanup_script="${CLEANUP_SCRIPT}" +licensetxt="$LICENSE" +helpheader='$HELPHEADER' +targetdir="$archdirname" +filesizes="$filesizes" +totalsize="$totalsize" +keep="$KEEP" +nooverwrite="$NOOVERWRITE" +quiet="n" +accept="n" +nodiskspace="n" +export_conf="$EXPORT_CONF" +decrypt_cmd="$DECRYPT_CMD" +skip="$SKIP" + +print_cmd_arg="" +if type printf > /dev/null; then + print_cmd="printf" +elif test -x /usr/ucb/echo; then + print_cmd="/usr/ucb/echo" +else + print_cmd="echo" +fi + +if test -d /usr/xpg4/bin; then + PATH=/usr/xpg4/bin:\$PATH + export PATH +fi + +if test -d /usr/sfw/bin; then + PATH=\$PATH:/usr/sfw/bin + export PATH +fi + +unset CDPATH + +MS_Printf() +{ + \$print_cmd \$print_cmd_arg "\$1" +} + +MS_PrintLicense() +{ + PAGER=\${PAGER:=more} + if test x"\$licensetxt" != x; then + PAGER_PATH=\`exec <&- 2>&-; which \$PAGER || command -v \$PAGER || type \$PAGER\` + if test -x "\$PAGER_PATH"; then + echo "\$licensetxt" | \$PAGER + else + echo "\$licensetxt" + fi + if test x"\$accept" != xy; then + while true + do + MS_Printf "Please type y to accept, n otherwise: " + read yn + if test x"\$yn" = xn; then + keep=n + eval \$finish; exit 1 + break; + elif test x"\$yn" = xy; then + break; + fi + done + fi + fi +} + +MS_diskspace() +{ + ( + df -kP "\$1" | tail -1 | awk '{ if (\$4 ~ /%/) {print \$3} else {print \$4} }' + ) +} + +MS_dd() +{ + blocks=\`expr \$3 / 1024\` + bytes=\`expr \$3 % 1024\` + # Test for ibs, obs and conv feature + if dd if=/dev/zero of=/dev/null count=1 ibs=512 obs=512 conv=sync 2> /dev/null; then + dd if="\$1" ibs=\$2 skip=1 obs=1024 conv=sync 2> /dev/null | \\ + { test \$blocks -gt 0 && dd ibs=1024 obs=1024 count=\$blocks ; \\ + test \$bytes -gt 0 && dd ibs=1 obs=1024 count=\$bytes ; } 2> /dev/null + else + dd if="\$1" bs=\$2 skip=1 2> /dev/null + fi +} + +MS_dd_Progress() +{ + if test x"\$noprogress" = xy; then + MS_dd "\$@" + return \$? + fi + file="\$1" + offset=\$2 + length=\$3 + pos=0 + bsize=4194304 + while test \$bsize -gt \$length; do + bsize=\`expr \$bsize / 4\` + done + blocks=\`expr \$length / \$bsize\` + bytes=\`expr \$length % \$bsize\` + ( + dd ibs=\$offset skip=1 2>/dev/null + pos=\`expr \$pos \+ \$bsize\` + MS_Printf " 0%% " 1>&2 + if test \$blocks -gt 0; then + while test \$pos -le \$length; do + dd bs=\$bsize count=1 2>/dev/null + pcent=\`expr \$length / 100\` + pcent=\`expr \$pos / \$pcent\` + if test \$pcent -lt 100; then + MS_Printf "\b\b\b\b\b\b\b" 1>&2 + if test \$pcent -lt 10; then + MS_Printf " \$pcent%% " 1>&2 + else + MS_Printf " \$pcent%% " 1>&2 + fi + fi + pos=\`expr \$pos \+ \$bsize\` + done + fi + if test \$bytes -gt 0; then + dd bs=\$bytes count=1 2>/dev/null + fi + MS_Printf "\b\b\b\b\b\b\b" 1>&2 + MS_Printf " 100%% " 1>&2 + ) < "\$file" +} + +MS_Help() +{ + cat << EOH >&2 +Usage: \$0 [options] +Options: + --help | -h Print this message + --info Print embedded info : title, default target directory, embedded script ... + --list Print the list of files in the archive + --check Checks integrity and version dependency of the archive + --quiet Quiet install mode, skip human-computer interactions + --nox11 Do not spawn an xterm + --noexec Do not run embedded script + --extract= Extract directly to a target directory (absolute or relative) + Usually used with --noexec to just extract files without running + --tar arg1 [arg2 ...] Access the contents of the archive through the tar command +\${helpheader} +EOH +} + +MS_Verify_Sig() +{ + GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` + MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` + test -x "\$GPG_PATH" || GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` + test -x "\$MKTEMP_PATH" || MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` + offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` + temp_sig=\`mktemp -t XXXXX\` + echo \$SIGNATURE | base64 --decode > "\$temp_sig" + gpg_output=\`MS_dd "\$1" \$offset \$totalsize | LC_ALL=C "\$GPG_PATH" --verify "\$temp_sig" - 2>&1\` + gpg_res=\$? + rm -f "\$temp_sig" + if test \$gpg_res -eq 0 && test \`echo \$gpg_output | grep -c Good\` -eq 1; then + if test \`echo \$gpg_output | grep -c \$sig_key\` -eq 1; then + test x"\$quiet" = xn && echo "GPG signature is good" >&2 + else + echo "GPG Signature key does not match" >&2 + exit 2 + fi + else + test x"\$quiet" = xn && echo "GPG signature failed to verify" >&2 + exit 2 + fi +} + +MS_Check() +{ + OLD_PATH="\$PATH" + PATH=\${GUESS_MD5_PATH:-"\$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} + MD5_ARG="" + MD5_PATH=\`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum\` + test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which md5 || command -v md5 || type md5\` + test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which digest || command -v digest || type digest\` + PATH="\$OLD_PATH" + + SHA_PATH=\`exec <&- 2>&-; which shasum || command -v shasum || type shasum\` + test -x "\$SHA_PATH" || SHA_PATH=\`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum\` + + if test x"\$quiet" = xn; then + MS_Printf "Verifying archive integrity..." + fi + offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` + fsize=\`cat "\$1" | wc -c | tr -d " "\` + if test \$totalsize -ne \`expr \$fsize - \$offset\`; then + echo " Unexpected archive size." >&2 + exit 2 + fi + verb=\$2 + i=1 + for s in \$filesizes + do + crc=\`echo \$CRCsum | cut -d" " -f\$i\` + if test -x "\$SHA_PATH"; then + if test x"\`basename \$SHA_PATH\`" = xshasum; then + SHA_ARG="-a 256" + fi + sha=\`echo \$SHA | cut -d" " -f\$i\` + if test x"\$sha" = x0000000000000000000000000000000000000000000000000000000000000000; then + test x"\$verb" = xy && echo " \$1 does not contain an embedded SHA256 checksum." >&2 + else + shasum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$SHA_PATH \$SHA_ARG" | cut -b-64\`; + if test x"\$shasum" != x"\$sha"; then + echo "Error in SHA256 checksums: \$shasum is different from \$sha" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " SHA256 checksums are OK." >&2 + fi + crc="0000000000"; + fi + fi + if test -x "\$MD5_PATH"; then + if test x"\`basename \$MD5_PATH\`" = xdigest; then + MD5_ARG="-a md5" + fi + md5=\`echo \$MD5 | cut -d" " -f\$i\` + if test x"\$md5" = x00000000000000000000000000000000; then + test x"\$verb" = xy && echo " \$1 does not contain an embedded MD5 checksum." >&2 + else + md5sum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$MD5_PATH \$MD5_ARG" | cut -b-32\`; + if test x"\$md5sum" != x"\$md5"; then + echo "Error in MD5 checksums: \$md5sum is different from \$md5" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " MD5 checksums are OK." >&2 + fi + crc="0000000000"; verb=n + fi + fi + if test x"\$crc" = x0000000000; then + test x"\$verb" = xy && echo " \$1 does not contain a CRC checksum." >&2 + else + sum1=\`MS_dd_Progress "\$1" \$offset \$s | CMD_ENV=xpg4 cksum | awk '{print \$1}'\` + if test x"\$sum1" != x"\$crc"; then + echo "Error in checksums: \$sum1 is different from \$crc" >&2 + exit 2 + elif test x"\$quiet" = xn; then + MS_Printf " CRC checksums are OK." >&2 + fi + fi + i=\`expr \$i + 1\` + offset=\`expr \$offset + \$s\` + done + if test x"\$quiet" = xn; then + echo " All good." + fi +} + +MS_Decompress() +{ + if test x"\$decrypt_cmd" != x""; then + { eval "\$decrypt_cmd" || echo " ... Decryption failed." >&2; } | eval "$GUNZIP_CMD" + else + eval "$GUNZIP_CMD" + fi + + if test \$? -ne 0; then + echo " ... Decompression failed." >&2 + fi +} + +UnTAR() +{ + if test x"\$quiet" = xn; then + tar \$1vf - $UNTAR_EXTRA 2>&1 || { echo " ... Extraction failed." >&2; kill -15 \$$; } + else + tar \$1f - $UNTAR_EXTRA 2>&1 || { echo Extraction failed. >&2; kill -15 \$$; } + fi +} + +MS_exec_cleanup() { + if test x"\$cleanup" = xy && test x"\$cleanup_script" != x""; then + cleanup=n + cd "\$tmpdir" + eval "\"\$cleanup_script\" \$scriptargs \$cleanupargs" + fi +} + +MS_cleanup() +{ + echo 'Signal caught, cleaning up' >&2 + MS_exec_cleanup + cd "\$TMPROOT" + rm -rf "\$tmpdir" + eval \$finish; exit 15 +} + +Script_Args_Check() +{ + script_supported_args=\$(echo \${helpheader} | grep -o -E "\-\-[^ ]+" | awk -F"=" {'print \$1'}) + arg_to_test=\$(echo \$1|awk -F"=" {'print \$1'}) + + for arg in \${script_supported_args}; + do + if test x"\$arg_to_test" = x"\$arg" ;then + return + fi + done + + MS_Help + exit 1 +} + +finish=true +xterm_loop= +noprogress=$NOPROGRESS +nox11=$NOX11 +copy=$COPY +ownership=$OWNERSHIP +verbose=n +cleanup=y +cleanupargs= +sig_key= + +initargs="\$@" + +while [ -n "\$*" ] +do + case "\$1" in + -h | --help) + MS_Help + exit 0 + ;; + -q | --quiet) + quiet=y + noprogress=y + shift + ;; + --info) + echo Identification: "\$label" + echo Target directory: "\$targetdir" + echo Uncompressed size: $USIZE KB + echo Compression: $COMPRESS + if test x"$ENCRYPT" != x""; then + echo Encryption: $ENCRYPT + fi + echo Date of packaging: $DATE + echo Built with Makeself version $MS_VERSION + echo Build command was: "$MS_COMMAND" + if test x"\$script" != x; then + echo Script run after extraction: + echo " " \$script \$scriptargs + fi + if test x"$copy" = xcopy; then + echo "Archive will copy itself to a temporary location" + fi + if test x"$NEED_ROOT" = xy; then + echo "Root permissions required for extraction" + fi + if test x"$KEEP" = xy; then + echo "directory \$targetdir is permanent" + else + echo "\$targetdir will be removed after extraction" + fi + exit 0 + ;; + --list) + echo Target directory: \$targetdir + offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + for s in \$filesizes + do + MS_dd "\$0" \$offset \$s | MS_Decompress | UnTAR t + offset=\`expr \$offset + \$s\` + done + exit 0 + ;; + --tar) + offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + arg1="\$2" + shift 2 || { MS_Help; exit 1; } + for s in \$filesizes + do + MS_dd "\$0" \$offset \$s | MS_Decompress | tar "\$arg1" - "\$@" + offset=\`expr \$offset + \$s\` + done + exit 0 + ;; + --check) + MS_Check "\$0" y + scriptargs="\$scriptargs \$1" + shift + ;; + --noexec) + script="" + cleanup_script="" + shift + ;; + --extract=*) + keep=y + targetdir=\`echo \$1 | cut -d"=" -f2 \` + if ! shift; then MS_Help; exit 1; fi + ;; + --nox11) + nox11=y + shift + ;; + --xwin) + if test "$NOWAIT" = n; then + finish="echo Press Return to close this window...; read junk" + fi + xterm_loop=1 + shift + ;; + --phase2) + copy=phase2 + shift + ;; + --repack | --repack-path=*) + Script_Args_Check \$1 + scriptargs="\$scriptargs '\$1'" + shift + if [[ ! "\$1" =~ ^-.* ]]; then + scriptargs="\$scriptargs '\$1'" + shift + fi + ;; + *) + Script_Args_Check \$1 + scriptargs="\$scriptargs '\$1'" + shift + ;; + esac +done + +quiet_para="" +if test x"\$quiet" = xy; then + quiet_para="--quiet " +fi +scriptargs="--\$name_of_file""--\"\$pwd_of_file\""" \$quiet_para""\$scriptargs" + +if test x"\$quiet" = xy -a x"\$verbose" = xy; then + echo Cannot be verbose and quiet at the same time. >&2 + exit 1 +fi + +if test x"$NEED_ROOT" = xy -a \`id -u\` -ne 0; then + echo "Administrative privileges required for this archive (use su or sudo)" >&2 + exit 1 +fi + +if test x"\$copy" \!= xphase2; then + MS_PrintLicense +fi + +case "\$copy" in +copy) + tmpdir="\$TMPROOT"/makeself.\$RANDOM.\`date +"%y%m%d%H%M%S"\`.\$\$ + mkdir "\$tmpdir" || { + echo "Could not create temporary directory \$tmpdir" >&2 + exit 1 + } + SCRIPT_COPY="\$tmpdir/makeself" + echo "Copying to a temporary location..." >&2 + cp "\$0" "\$SCRIPT_COPY" + chmod +x "\$SCRIPT_COPY" + cd "\$TMPROOT" + exec "\$SCRIPT_COPY" --phase2 -- \$initargs + ;; +phase2) + finish="\$finish ; rm -rf \`dirname \$0\`" + ;; +esac + +if test x"\$nox11" = xn; then + if tty -s; then # Do we have a terminal? + : + else + if test x"\$DISPLAY" != x -a x"\$xterm_loop" = x; then # No, but do we have X? + if xset q > /dev/null 2>&1; then # Check for valid DISPLAY variable + GUESS_XTERMS="xterm gnome-terminal rxvt dtterm eterm Eterm xfce4-terminal lxterminal kvt konsole aterm terminology" + for a in \$GUESS_XTERMS; do + if type \$a >/dev/null 2>&1; then + XTERM=\$a + break + fi + done + chmod a+x \$0 || echo Please add execution rights on \$0 + if test \`echo "\$0" | cut -c1\` = "/"; then # Spawn a terminal! + exec \$XTERM -e "\$0 --xwin \$initargs" + else + exec \$XTERM -e "./\$0 --xwin \$initargs" + fi + fi + fi + fi +fi + +if test x"\$targetdir" = x.; then + tmpdir="." +else + if test x"\$keep" = xy; then + if test x"\$nooverwrite" = xy && test -d "\$targetdir"; then + echo "Target directory \$targetdir already exists, aborting." >&2 + exit 1 + fi + if test x"\$quiet" = xn; then + echo "Creating directory \$targetdir" >&2 + fi + tmpdir="\$targetdir" + dashp="-p" + else + tmpdir="\$TMPROOT/selfgz\$\$\$RANDOM" + dashp="" + fi + mkdir \$dashp "\$tmpdir" || { + echo 'Cannot create target directory' \$tmpdir >&2 + echo 'You should try option --extract=' >&2 + eval \$finish + exit 1 + } +fi + +location="\`pwd\`" +if test x"\$SETUP_NOCHECK" != x1; then + MS_Check "\$0" +fi +offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` + +if test x"\$verbose" = xy; then + MS_Printf "About to extract $USIZE KB in \$tmpdir ... Proceed ? [Y/n] " + read yn + if test x"\$yn" = xn; then + eval \$finish; exit 1 + fi +fi + +if test x"\$quiet" = xn; then + # Decrypting with openssl will ask for password, + # the prompt needs to start on new line + if test x"$ENCRYPT" = x"openssl"; then + echo "Decrypting and uncompressing \$label..." + else + MS_Printf "Uncompressing \$label" + fi +fi +res=3 +if test x"\$keep" = xn; then + trap MS_cleanup 1 2 3 15 +fi + +if test x"\$nodiskspace" = xn; then + leftspace=\`MS_diskspace "\$tmpdir"\` + if test -n "\$leftspace"; then + if test "\$leftspace" -lt $USIZE; then + echo + echo "Not enough space left in "\`dirname \$tmpdir\`" (\$leftspace KB) to decompress \$0 ($USIZE KB)" >&2 + if test x"\$keep" = xn; then + echo "Consider setting TMPDIR to a directory with more free space." + fi + eval \$finish; exit 1 + fi + fi +fi + +for s in \$filesizes +do + if MS_dd_Progress "\$0" \$offset \$s | MS_Decompress | ( cd "\$tmpdir"; umask \$ORIG_UMASK ; UnTAR xp ) 1>/dev/null; then + if test x"\$ownership" = xy; then + (cd "\$tmpdir"; chown -R \`id -u\` .; chgrp -R \`id -g\` .) + fi + else + echo >&2 + echo "Unable to decompress \$0" >&2 + eval \$finish; exit 1 + fi + offset=\`expr \$offset + \$s\` +done +if test x"\$quiet" = xn; then + echo +fi + +cd "\$tmpdir" +res=0 +if test x"\$script" != x; then + if test x"\$export_conf" = x"y"; then + MS_BUNDLE="\$0" + MS_LABEL="\$label" + MS_SCRIPT="\$script" + MS_SCRIPTARGS="\$scriptargs" + MS_ARCHDIRNAME="\$archdirname" + MS_KEEP="\$KEEP" + MS_NOOVERWRITE="\$NOOVERWRITE" + MS_COMPRESS="\$COMPRESS" + MS_CLEANUP="\$cleanup" + export MS_BUNDLE MS_LABEL MS_SCRIPT MS_SCRIPTARGS + export MS_ARCHDIRNAME MS_KEEP MS_NOOVERWRITE MS_COMPRESS + fi + + if test x"\$verbose" = x"y"; then + yn="x" + while test x"\$yn" != x -a x"\$yn" != xy -a x"\$yn" != xY -a x"\$yn" != xn -a x"\$yn" != xN + do + MS_Printf "OK to execute: \$script \$scriptargs \$* ? [Y/n] " + read yn + if test x"\$yn" = x -o x"\$yn" = xy -o x"\$yn" = xY; then + eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$?; + elif test x"\$yn" = xn -o x"\$yn" = xN; then + echo "Unable to decompress \$script ,because of aborting! ";res=\$? + else + echo "Input value is unacceptable,please try again." + fi + done + else + eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$? + fi + if test "\$res" -ne 0; then + test x"\$verbose" = xy && echo "The program '\$script' returned an error code (\$res)" >&2 + fi +fi + +MS_exec_cleanup + +if test x"\$keep" = xn; then + cd "\$TMPROOT" + rm -rf "\$tmpdir" +fi +eval \$finish; exit \$res +EOF diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 new file mode 100644 index 000000000..81bf6e4ff --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 @@ -0,0 +1,110 @@ +.TH "MAKESELF" "1" "2.4.5" +.SH "NAME" +makeself \- An utility to generate self-extractable archives. +.SH "SYNTAX" +.B makeself [\fIoptions\fP] archive_dir file_name label +.B [\fIstartup_script\fP] [\fIargs\fP] +.SH "DESCRIPTION" +This program is a free (GPL) utility designed to create self-extractable +archives from a directory. +.SH "OPTIONS" +The following options are supported. +.TP 15 +.B -v, --version +Prints out the makeself version number and exits. +.TP +.B -h, --help +Print out help information. +.TP +.B --tar-quietly +Suppress verbose output from the tar command +.TP +.B --quiet +Do not print any messages other than errors +.TP +.B --gzip +Compress using gzip (default if detected). +.TP +.B --bzip2 +Compress using bzip2. +.TP +.B --pbzip2 +Compress using pbzip2. +.TP +.B --xz +Compress using xz. +.TP +.B --lzo +Compress using lzop. +.TP +.B --lz4 +Compress using lz4. +.TP +.B --compress +Compress using the UNIX 'compress' command. +.TP +.B --nocomp +Do not compress the data. +.TP +.B --complevel lvl +Specify the compression level for gzip,bzip2,pbzui2,xz,lzo or lz4 +.TP +.B --notemp +The archive will create archive_dir in the current directory and +uncompress in ./archive_dir. +.TP +.B --copy +Upon extraction, the archive will first copy itself to a temporary directory. +.TP +.B --append +Append more files to an existing makeself archive. The label and startup scripts will then be ignored. +.TP +.B --current +Files will be extracted to the current directory. Both --current and --target dir imply --notemp. +.TP +.B --target dir +Extract directly to a target directory. Directory path can be either absolute or relative. +.TP +.B --header file +Specify location of the header script. +.TP +.B --cleanup file +Specify a cleanup script that executes on interrupt and when finished successfully. +.TP +.B --follow +Follow the symlinks in the archive. +.TP +.B --noprogress +Do not show the progress during the decompression. +.TP +.B --nox11 +Disable automatic spawn of an xterm if running in X11. +.TP +.B --nowait +Do not wait for user input after executing embedded program from an xterm. +.TP +.B --nomd5 +Do not create a MD5 checksum for the archive. +.TP +.B --nocrc +Do not create a CRC32 checksum for the archive. +.TP +.B --lsm file +LSM file describing the package. +.B --packaging-date date +Use provided string as the packaging date instead of the current date. +.SH "EXAMPLES" +Here is an example, assuming the user has a package image stored in a /home/joe/mysoft, +and he wants to generate a self-extracting package named mysoft.sh, which will launch +the "setup" script initially stored in /home/joe/mysoft: +.TP +makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup +.TP +Here is also how I created the makeself.run archive which contains the Makeself distribution: +.TP +makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" +.SH "AUTHORS" +Makeself has been written by Stéphane Peter . +.BR +This man page was originally written by Bartosz Fenski for the +Debian GNU/Linux distribution (but it may be used by others). diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm new file mode 100644 index 000000000..3c4cea8c1 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm @@ -0,0 +1,16 @@ +Begin3 +Title: makeself.sh +Version: 2.4.5 +Description: makeself.sh is a shell script that generates a self-extractable + tar.gz archive from a directory. The resulting file appears as a shell + script, and can be launched as is. The archive will then uncompress + itself to a temporary directory and an arbitrary command will be + executed (for example an installation script). This is pretty similar + to archives generated with WinZip Self-Extractor in the Windows world. +Keywords: Installation archive tar winzip +Author: Stephane Peter (megastep@megastep.org) +Maintained-by: Stephane Peter (megastep@megastep.org) +Original-site: https://makeself.io/ +Platform: Unix +Copying-policy: GPL +End diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh new file mode 100755 index 000000000..c8ea56597 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh @@ -0,0 +1,822 @@ +#!/bin/sh +# +# Makeself version 2.4.x +# by Stephane Peter +# +# Utility to create self-extracting tar.gz archives. +# The resulting archive is a file holding the tar.gz archive with +# a small Shell script stub that uncompresses the archive to a temporary +# directory and then executes a given script from withing that directory. +# +# Makeself home page: https://makeself.io/ +# +# Version 2.0 is a rewrite of version 1.0 to make the code easier to read and maintain. +# +# Version history : +# - 1.0 : Initial public release +# - 1.1 : The archive can be passed parameters that will be passed on to +# the embedded script, thanks to John C. Quillan +# - 1.2 : Package distribution, bzip2 compression, more command line options, +# support for non-temporary archives. Ideas thanks to Francois Petitjean +# - 1.3 : More patches from Bjarni R. Einarsson and Francois Petitjean: +# Support for no compression (--nocomp), script is no longer mandatory, +# automatic launch in an xterm, optional verbose output, and -target +# archive option to indicate where to extract the files. +# - 1.4 : Improved UNIX compatibility (Francois Petitjean) +# Automatic integrity checking, support of LSM files (Francois Petitjean) +# - 1.5 : Many bugfixes. Optionally disable xterm spawning. +# - 1.5.1 : More bugfixes, added archive options -list and -check. +# - 1.5.2 : Cosmetic changes to inform the user of what's going on with big +# archives (Quake III demo) +# - 1.5.3 : Check for validity of the DISPLAY variable before launching an xterm. +# More verbosity in xterms and check for embedded command's return value. +# Bugfix for Debian 2.0 systems that have a different "print" command. +# - 1.5.4 : Many bugfixes. Print out a message if the extraction failed. +# - 1.5.5 : More bugfixes. Added support for SETUP_NOCHECK environment variable to +# bypass checksum verification of archives. +# - 1.6.0 : Compute MD5 checksums with the md5sum command (patch from Ryan Gordon) +# - 2.0 : Brand new rewrite, cleaner architecture, separated header and UNIX ports. +# - 2.0.1 : Added --copy +# - 2.1.0 : Allow multiple tarballs to be stored in one archive, and incremental updates. +# Added --nochown for archives +# Stopped doing redundant checksums when not necesary +# - 2.1.1 : Work around insane behavior from certain Linux distros with no 'uncompress' command +# Cleaned up the code to handle error codes from compress. Simplified the extraction code. +# - 2.1.2 : Some bug fixes. Use head -n to avoid problems. +# - 2.1.3 : Bug fixes with command line when spawning terminals. +# Added --tar for archives, allowing to give arbitrary arguments to tar on the contents of the archive. +# Added --noexec to prevent execution of embedded scripts. +# Added --nomd5 and --nocrc to avoid creating checksums in archives. +# Added command used to create the archive in --info output. +# Run the embedded script through eval. +# - 2.1.4 : Fixed --info output. +# Generate random directory name when extracting files to . to avoid problems. (Jason Trent) +# Better handling of errors with wrong permissions for the directory containing the files. (Jason Trent) +# Avoid some race conditions (Ludwig Nussel) +# Unset the $CDPATH variable to avoid problems if it is set. (Debian) +# Better handling of dot files in the archive directory. +# - 2.1.5 : Made the md5sum detection consistent with the header code. +# Check for the presence of the archive directory +# Added --encrypt for symmetric encryption through gpg (Eric Windisch) +# Added support for the digest command on Solaris 10 for MD5 checksums +# Check for available disk space before extracting to the target directory (Andreas Schweitzer) +# Allow extraction to run asynchronously (patch by Peter Hatch) +# Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo) +# - 2.1.6 : Replaced one dot per file progress with a realtime progress percentage and a spining cursor (Guy Baconniere) +# Added --noprogress to prevent showing the progress during the decompression (Guy Baconniere) +# Added --target dir to allow extracting directly to a target directory (Guy Baconniere) +# - 2.2.0 : Many bugfixes, updates and contributions from users. Check out the project page on Github for the details. +# - 2.3.0 : Option to specify packaging date to enable byte-for-byte reproducibility. (Marc Pawlowsky) +# - 2.4.0 : Optional support for SHA256 checksums in archives. +# - 2.4.2 : Add support for threads for several compressors. (M. Limber) +# Added zstd support. +# - 2.4.3 : Make explicit POSIX tar archives for increased compatibility. +# - 2.4.5 : Added --tar-format to override ustar tar archive format +# +# (C) 1998-2021 by Stephane Peter +# +# This software is released under the terms of the GNU GPL version 2 and above +# Please read the license at http://www.gnu.org/copyleft/gpl.html +# Self-extracting archives created with this script are explictly NOT released under the term of the GPL +# + +MS_VERSION=2.4.5 +MS_COMMAND="$0" +unset CDPATH + +for f in ${1+"$@"}; do + MS_COMMAND="$MS_COMMAND \\\\ + \\\"$f\\\"" +done + +# For Solaris systems +if test -d /usr/xpg4/bin; then + PATH=/usr/xpg4/bin:$PATH + export PATH +fi + +# Procedures + +MS_Usage() +{ + echo "Usage: $0 [args] archive_dir file_name label startup_script [script_args]" + echo "args can be one or more of the following :" + echo " --version | -v : Print out Makeself version number and exit" + echo " --help | -h : Print out this help message" + echo " --tar-quietly : Suppress verbose output from the tar command" + echo " --quiet | -q : Do not print any messages other than errors." + echo " --gzip : Compress using gzip (default if detected)" + echo " --pigz : Compress with pigz" + echo " --zstd : Compress with zstd" + echo " --bzip2 : Compress using bzip2 instead of gzip" + echo " --pbzip2 : Compress using pbzip2 instead of gzip" + echo " --xz : Compress using xz instead of gzip" + echo " --lzo : Compress using lzop instead of gzip" + echo " --lz4 : Compress using lz4 instead of gzip" + echo " --compress : Compress using the UNIX 'compress' command" + echo " --complevel lvl : Compression level for gzip pigz zstd xz lzo lz4 bzip2 and pbzip2 (default 9)" + echo " --threads thds : Number of threads to be used by compressors that support parallelization." + echo " Omit to use compressor's default. Most useful (and required) for opting" + echo " into xz's threading, usually with '--threads=0' for all available cores." + echo " pbzip2 and pigz are parallel by default, and setting this value allows" + echo " limiting the number of threads they use." + echo " --base64 : Instead of compressing, encode the data using base64" + echo " --gpg-encrypt : Instead of compressing, encrypt the data using GPG" + echo " --gpg-asymmetric-encrypt-sign" + echo " : Instead of compressing, asymmetrically encrypt and sign the data using GPG" + echo " --gpg-extra opt : Append more options to the gpg command line" + echo " --ssl-encrypt : Instead of compressing, encrypt the data using OpenSSL" + echo " --ssl-passwd pass : Use the given password to encrypt the data using OpenSSL" + echo " --ssl-pass-src src : Use the given src as the source of password to encrypt the data" + echo " using OpenSSL. See \"PASS PHRASE ARGUMENTS\" in man openssl." + echo " If this option is not supplied, the user will be asked to enter" + echo " encryption password on the current terminal." + echo " --ssl-no-md : Do not use \"-md\" option not supported by older OpenSSL." + echo " --nochown : Do not give the target folder to the current user (default)" + echo " --chown : Give the target folder to the current user recursively" + echo " --nocomp : Do not compress the data" + echo " --notemp : The archive will create archive_dir in the" + echo " current directory and uncompress in ./archive_dir" + echo " --needroot : Check that the root user is extracting the archive before proceeding" + echo " --copy : Upon extraction, the archive will first copy itself to" + echo " a temporary directory" + echo " --append : Append more files to an existing Makeself archive" + echo " The label and startup scripts will then be ignored" + echo " --target dir : Extract directly to a target directory" + echo " directory path can be either absolute or relative" + echo " --nooverwrite : Do not extract the archive if the specified target directory exists" + echo " --current : Files will be extracted to the current directory" + echo " Both --current and --target imply --notemp" + echo " --tar-format opt : Specify a tar archive format (default is ustar)" + echo " --tar-extra opt : Append more options to the tar command line" + echo " --untar-extra opt : Append more options to the during the extraction of the tar archive" + echo " --nomd5 : Don't calculate an MD5 for archive" + echo " --nocrc : Don't calculate a CRC for archive" + echo " --sha256 : Compute a SHA256 checksum for the archive" + echo " --header file : Specify location of the header script" + echo " --cleanup file : Specify a cleanup script that executes on interrupt and when finished successfully." + echo " --follow : Follow the symlinks in the archive" + echo " --noprogress : Do not show the progress during the decompression" + echo " --nox11 : Disable automatic spawn of a xterm" + echo " --nowait : Do not wait for user input after executing embedded" + echo " program from an xterm" + echo " --sign passphrase : Signature private key to sign the package with" + echo " --lsm file : LSM file describing the package" + echo " --license file : Append a license file" + echo " --help-header file : Add a header to the archive's --help output" + echo " --packaging-date date" + echo " : Use provided string as the packaging date" + echo " instead of the current date." + echo + echo " --keep-umask : Keep the umask set to shell default, rather than overriding when executing self-extracting archive." + echo " --export-conf : Export configuration variables to startup_script" + echo + echo "Do not forget to give a fully qualified startup script name" + echo "(i.e. with a ./ prefix if inside the archive)." + exit 1 +} + +# Default settings +if type gzip >/dev/null 2>&1; then + COMPRESS=gzip +elif type compress >/dev/null 2>&1; then + COMPRESS=compress +else + echo "ERROR: missing commands: gzip, compress" >&2 + MS_Usage +fi +ENCRYPT=n +PASSWD="" +PASSWD_SRC="" +OPENSSL_NO_MD=n +COMPRESS_LEVEL=9 +DEFAULT_THREADS=123456 # Sentinel value +THREADS=$DEFAULT_THREADS +KEEP=n +CURRENT=n +NOX11=n +NOWAIT=n +APPEND=n +TAR_QUIETLY=n +KEEP_UMASK=n +QUIET=n +NOPROGRESS=n +COPY=none +NEED_ROOT=n +TAR_ARGS=rvf +TAR_FORMAT=ustar +TAR_EXTRA="" +GPG_EXTRA="" +DU_ARGS=-ks +HEADER=`dirname "$0"`/makeself-header.sh +SIGNATURE="" +TARGETDIR="" +NOOVERWRITE=n +DATE=`LC_ALL=C date` +EXPORT_CONF=n +SHA256=n +OWNERSHIP=n +SIGN=n +GPG_PASSPHRASE="" + +# LSM file stuff +LSM_CMD="echo No LSM. >> \"\$archname\"" + +while true +do + case "$1" in + --version | -v) + echo Makeself version $MS_VERSION + exit 0 + ;; + --pbzip2) + COMPRESS=pbzip2 + shift + ;; + --bzip2) + COMPRESS=bzip2 + shift + ;; + --gzip) + COMPRESS=gzip + shift + ;; + --pigz) + COMPRESS=pigz + shift + ;; + --zstd) + COMPRESS=zstd + shift + ;; + --xz) + COMPRESS=xz + shift + ;; + --lzo) + COMPRESS=lzo + shift + ;; + --lz4) + COMPRESS=lz4 + shift + ;; + --compress) + COMPRESS=compress + shift + ;; + --base64) + COMPRESS=base64 + shift + ;; + --gpg-encrypt) + COMPRESS=gpg + shift + ;; + --gpg-asymmetric-encrypt-sign) + COMPRESS=gpg-asymmetric + shift + ;; + --gpg-extra) + GPG_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-encrypt) + ENCRYPT=openssl + shift + ;; + --ssl-passwd) + PASSWD=$2 + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-pass-src) + PASSWD_SRC=$2 + shift 2 || { MS_Usage; exit 1; } + ;; + --ssl-no-md) + OPENSSL_NO_MD=y + shift + ;; + --nocomp) + COMPRESS=none + shift + ;; + --complevel) + COMPRESS_LEVEL="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --threads) + THREADS="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --nochown) + OWNERSHIP=n + shift + ;; + --chown) + OWNERSHIP=y + shift + ;; + --notemp) + KEEP=y + shift + ;; + --copy) + COPY=copy + shift + ;; + --current) + CURRENT=y + KEEP=y + shift + ;; + --tar-format) + TAR_FORMAT="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --tar-extra) + TAR_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --untar-extra) + UNTAR_EXTRA="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --target) + TARGETDIR="$2" + KEEP=y + shift 2 || { MS_Usage; exit 1; } + ;; + --sign) + SIGN=y + GPG_PASSPHRASE="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --nooverwrite) + NOOVERWRITE=y + shift + ;; + --needroot) + NEED_ROOT=y + shift + ;; + --header) + HEADER="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --cleanup) + CLEANUP_SCRIPT="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --license) + # We need to escape all characters having a special meaning in double quotes + LICENSE=$(sed 's/\\/\\\\/g; s/"/\\\"/g; s/`/\\\`/g; s/\$/\\\$/g' "$2") + shift 2 || { MS_Usage; exit 1; } + ;; + --follow) + TAR_ARGS=rvhf + DU_ARGS=-ksL + shift + ;; + --noprogress) + NOPROGRESS=y + shift + ;; + --nox11) + NOX11=y + shift + ;; + --nowait) + NOWAIT=y + shift + ;; + --nomd5) + NOMD5=y + shift + ;; + --sha256) + SHA256=y + shift + ;; + --nocrc) + NOCRC=y + shift + ;; + --append) + APPEND=y + shift + ;; + --lsm) + LSM_CMD="cat \"$2\" >> \"\$archname\"" + shift 2 || { MS_Usage; exit 1; } + ;; + --packaging-date) + DATE="$2" + shift 2 || { MS_Usage; exit 1; } + ;; + --help-header) + HELPHEADER=`sed -e "s/'/'\\\\\''/g" $2` + shift 2 || { MS_Usage; exit 1; } + [ -n "$HELPHEADER" ] && HELPHEADER="$HELPHEADER +" + ;; + --tar-quietly) + TAR_QUIETLY=y + shift + ;; + --keep-umask) + KEEP_UMASK=y + shift + ;; + --export-conf) + EXPORT_CONF=y + shift + ;; + -q | --quiet) + QUIET=y + shift + ;; + -h | --help) + MS_Usage + ;; + -*) + echo Unrecognized flag : "$1" + MS_Usage + ;; + *) + break + ;; + esac +done + +if test $# -lt 1; then + MS_Usage +else + if test -d "$1"; then + archdir="$1" + else + echo "Directory $1 does not exist." >&2 + exit 1 + fi +fi +archname="$2" + +if test "$QUIET" = "y" || test "$TAR_QUIETLY" = "y"; then + if test "$TAR_ARGS" = "rvf"; then + TAR_ARGS="rf" + elif test "$TAR_ARGS" = "rvhf"; then + TAR_ARGS="rhf" + fi +fi + +if test "$APPEND" = y; then + if test $# -lt 2; then + MS_Usage + fi + + # Gather the info from the original archive + OLDENV=`sh "$archname" --dumpconf` + if test $? -ne 0; then + echo "Unable to update archive: $archname" >&2 + exit 1 + else + eval "$OLDENV" + OLDSKIP=`expr $SKIP + 1` + fi +else + if test "$KEEP" = n -a $# = 3; then + echo "ERROR: Making a temporary archive with no embedded command does not make sense!" >&2 + echo >&2 + MS_Usage + fi + # We don't want to create an absolute directory unless a target directory is defined + if test "$CURRENT" = y; then + archdirname="." + elif test x"$TARGETDIR" != x; then + archdirname="$TARGETDIR" + else + archdirname=`basename "$1"` + fi + + if test $# -lt 3; then + MS_Usage + fi + + LABEL="$3" + SCRIPT="$4" + test "x$SCRIPT" = x || shift 1 + shift 3 + SCRIPTARGS="$*" +fi + +if test "$KEEP" = n -a "$CURRENT" = y; then + echo "ERROR: It is A VERY DANGEROUS IDEA to try to combine --notemp and --current." >&2 + exit 1 +fi + +case $COMPRESS in +gzip) + GZIP_CMD="gzip -c$COMPRESS_LEVEL" + GUNZIP_CMD="gzip -cd" + ;; +pigz) + GZIP_CMD="pigz -$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD --processes $THREADS" + fi + GUNZIP_CMD="gzip -cd" + ;; +zstd) + GZIP_CMD="zstd -$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD --threads=$THREADS" + fi + GUNZIP_CMD="zstd -cd" + ;; +pbzip2) + GZIP_CMD="pbzip2 -c$COMPRESS_LEVEL" + if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated + GZIP_CMD="$GZIP_CMD -p$THREADS" + fi + GUNZIP_CMD="bzip2 -d" + ;; +bzip2) + GZIP_CMD="bzip2 -$COMPRESS_LEVEL" + GUNZIP_CMD="bzip2 -d" + ;; +xz) + GZIP_CMD="xz -c$COMPRESS_LEVEL" + # Must opt-in by specifying a value since not all versions of xz support threads + if test $THREADS -ne $DEFAULT_THREADS; then + GZIP_CMD="$GZIP_CMD --threads=$THREADS" + fi + GUNZIP_CMD="xz -d" + ;; +lzo) + GZIP_CMD="lzop -c$COMPRESS_LEVEL" + GUNZIP_CMD="lzop -d" + ;; +lz4) + GZIP_CMD="lz4 -c$COMPRESS_LEVEL" + GUNZIP_CMD="lz4 -d" + ;; +base64) + GZIP_CMD="base64" + GUNZIP_CMD="base64 --decode -i -" + ;; +gpg) + GZIP_CMD="gpg $GPG_EXTRA -ac -z$COMPRESS_LEVEL" + GUNZIP_CMD="gpg -d" + ENCRYPT="gpg" + ;; +gpg-asymmetric) + GZIP_CMD="gpg $GPG_EXTRA -z$COMPRESS_LEVEL -es" + GUNZIP_CMD="gpg --yes -d" + ENCRYPT="gpg" + ;; +compress) + GZIP_CMD="compress -fc" + GUNZIP_CMD="(type compress >/dev/null 2>&1 && compress -fcd || gzip -cd)" + ;; +none) + GZIP_CMD="cat" + GUNZIP_CMD="cat" + ;; +esac + +if test x"$ENCRYPT" = x"openssl"; then + if test x"$APPEND" = x"y"; then + echo "Appending to existing archive is not compatible with OpenSSL encryption." >&2 + fi + + ENCRYPT_CMD="openssl enc -aes-256-cbc -salt" + DECRYPT_CMD="openssl enc -aes-256-cbc -d" + + if test x"$OPENSSL_NO_MD" != x"y"; then + ENCRYPT_CMD="$ENCRYPT_CMD -md sha256" + DECRYPT_CMD="$DECRYPT_CMD -md sha256" + fi + + if test -n "$PASSWD_SRC"; then + ENCRYPT_CMD="$ENCRYPT_CMD -pass $PASSWD_SRC" + elif test -n "$PASSWD"; then + ENCRYPT_CMD="$ENCRYPT_CMD -pass pass:$PASSWD" + fi +fi + +tmpfile="${TMPDIR:-/tmp}/mkself$$" + +if test -f "$HEADER"; then + oldarchname="$archname" + archname="$tmpfile" + # Generate a fake header to count its lines + SKIP=0 + . "$HEADER" + SKIP=`cat "$tmpfile" |wc -l` + # Get rid of any spaces + SKIP=`expr $SKIP` + rm -f "$tmpfile" + if test "$QUIET" = "n"; then + echo "Header is $SKIP lines long" >&2 + fi + archname="$oldarchname" +else + echo "Unable to open header file: $HEADER" >&2 + exit 1 +fi + +if test "$QUIET" = "n"; then + echo +fi + +if test "$APPEND" = n; then + if test -f "$archname"; then + echo "WARNING: Overwriting existing file: $archname" >&2 + fi +fi + +USIZE=`du $DU_ARGS "$archdir" | awk '{print $1}'` + +if test "." = "$archdirname"; then + if test "$KEEP" = n; then + archdirname="makeself-$$-`date +%Y%m%d%H%M%S`" + fi +fi + +test -d "$archdir" || { echo "Error: $archdir does not exist."; rm -f "$tmpfile"; exit 1; } +if test "$QUIET" = "n"; then + echo "About to compress $USIZE KB of data..." + echo "Adding files to archive named \"$archname\"..." +fi + +# See if we have GNU tar +TAR=`exec <&- 2>&-; which gtar || command -v gtar || type gtar` +test -x "$TAR" || TAR=tar + +tmparch="${TMPDIR:-/tmp}/mkself$$.tar" +( + if test "$APPEND" = "y"; then + tail -n "+$OLDSKIP" "$archname" | eval "$GUNZIP_CMD" > "$tmparch" + fi + cd "$archdir" + # "Determining if a directory is empty" + # https://www.etalabs.net/sh_tricks.html + find . \ + \( \ + ! -type d \ + -o \ + \( -links 2 -exec sh -c ' + is_empty () ( + cd "$1" + set -- .[!.]* ; test -f "$1" && return 1 + set -- ..?* ; test -f "$1" && return 1 + set -- * ; test -f "$1" && return 1 + return 0 + ) + is_empty "$0"' {} \; \ + \) \ + \) -print \ + | LC_ALL=C sort \ + | sed 's/./\\&/g' \ + | xargs $TAR $TAR_EXTRA --format $TAR_FORMAT -$TAR_ARGS "$tmparch" +) || { + echo "ERROR: failed to create temporary archive: $tmparch" + rm -f "$tmparch" "$tmpfile" + exit 1 +} + +USIZE=`du $DU_ARGS "$tmparch" | awk '{print $1}'` + +eval "$GZIP_CMD" <"$tmparch" >"$tmpfile" || { + echo "ERROR: failed to create temporary file: $tmpfile" + rm -f "$tmparch" "$tmpfile" + exit 1 +} +rm -f "$tmparch" + +if test x"$ENCRYPT" = x"openssl"; then + echo "About to encrypt archive \"$archname\"..." + { eval "$ENCRYPT_CMD -in $tmpfile -out ${tmpfile}.enc" && mv -f ${tmpfile}.enc $tmpfile; } || \ + { echo Aborting: could not encrypt temporary file: "$tmpfile".; rm -f "$tmpfile"; exit 1; } +fi + +fsize=`cat "$tmpfile" | wc -c | tr -d " "` + +# Compute the checksums + +shasum=0000000000000000000000000000000000000000000000000000000000000000 +md5sum=00000000000000000000000000000000 +crcsum=0000000000 + +if test "$NOCRC" = y; then + if test "$QUIET" = "n"; then + echo "skipping crc at user request" + fi +else + crcsum=`CMD_ENV=xpg4 cksum < "$tmpfile" | sed -e 's/ /Z/' -e 's/ /Z/' | cut -dZ -f1` + if test "$QUIET" = "n"; then + echo "CRC: $crcsum" + fi +fi + +if test "$SHA256" = y; then + SHA_PATH=`exec <&- 2>&-; which shasum || command -v shasum || type shasum` + if test -x "$SHA_PATH"; then + shasum=`eval "$SHA_PATH -a 256" < "$tmpfile" | cut -b-64` + else + SHA_PATH=`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum` + shasum=`eval "$SHA_PATH" < "$tmpfile" | cut -b-64` + fi + if test "$QUIET" = "n"; then + if test -x "$SHA_PATH"; then + echo "SHA256: $shasum" + else + echo "SHA256: none, SHA command not found" + fi + fi +fi +if test "$NOMD5" = y; then + if test "$QUIET" = "n"; then + echo "Skipping md5sum at user request" + fi +else + # Try to locate a MD5 binary + OLD_PATH=$PATH + PATH=${GUESS_MD5_PATH:-"$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} + MD5_ARG="" + MD5_PATH=`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum` + test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which md5 || command -v md5 || type md5` + test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which digest || command -v digest || type digest` + PATH=$OLD_PATH + if test -x "$MD5_PATH"; then + if test `basename ${MD5_PATH}`x = digestx; then + MD5_ARG="-a md5" + fi + md5sum=`eval "$MD5_PATH $MD5_ARG" < "$tmpfile" | cut -b-32` + if test "$QUIET" = "n"; then + echo "MD5: $md5sum" + fi + else + if test "$QUIET" = "n"; then + echo "MD5: none, MD5 command not found" + fi + fi +fi +if test "$SIGN" = y; then + GPG_PATH=`exec <&- 2>&-; which gpg || command -v gpg || type gpg` + if test -x "$GPG_PATH"; then + SIGNATURE=`$GPG_PATH --pinentry-mode=loopback --batch --yes --passphrase "$GPG_PASSPHRASE" --output - --detach-sig $tmpfile | base64 | tr -d \\\\n` + if test "$QUIET" = "n"; then + echo "Signature: $SIGNATURE" + fi + else + echo "Missing gpg command" >&2 + fi +fi + +totalsize=0 +for size in $fsize; +do + totalsize=`expr $totalsize + $size` +done + +if test "$APPEND" = y; then + mv "$archname" "$archname".bak || exit + + # Prepare entry for new archive + filesizes="$fsize" + CRCsum="$crcsum" + MD5sum="$md5sum" + SHAsum="$shasum" + Signature="$SIGNATURE" + # Generate the header + . "$HEADER" + # Append the new data + cat "$tmpfile" >> "$archname" + + chmod +x "$archname" + rm -f "$archname".bak + if test "$QUIET" = "n"; then + echo "Self-extractable archive \"$archname\" successfully updated." + fi +else + filesizes="$fsize" + CRCsum="$crcsum" + MD5sum="$md5sum" + SHAsum="$shasum" + Signature="$SIGNATURE" + + # Generate the header + . "$HEADER" + + # Append the compressed tar data after the stub + if test "$QUIET" = "n"; then + echo + fi + cat "$tmpfile" >> "$archname" + chmod +x "$archname" + if test "$QUIET" = "n"; then + echo Self-extractable archive \"$archname\" successfully created. + fi +fi +rm -f "$tmpfile" diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh new file mode 100644 index 000000000..31ee16511 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh @@ -0,0 +1,8 @@ +#!/bin/sh +# Run every available test - Bash needed +cd test +for test in *test; +do + echo "Running test $test ..." + bash $test || { echo "*** ERROR: Test '$test' failed!"; exit 1; } +done diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh new file mode 100755 index 000000000..a977bd51d --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh @@ -0,0 +1,31 @@ +#!/bin/bash + +project_path=$1 +build_path=$2 +vendor_name=customize +echo $@ +if [[ ! -d "$project_path" ]]; then + echo "[ERROR] No projcet path is provided" + exit 1 +fi + +if [[ ! -d "$build_path" ]]; then + echo "[ERROR] No build path is provided" + exit 1 +fi + +if [[ ! -d "$ASCEND_OPP_PATH" ]]; then + echo "[ERROR] No opp install path is provided" + exit 1 +fi +custom_exist_info_json=$ASCEND_OPP_PATH/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json +custom_new_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json +temp_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/temp_cust_aicpu_kernel.json + +if [[ -f "$custom_exist_info_json" ]] && [[ -f "$custom_new_info_json" ]]; then + cp -f $custom_exist_info_json $temp_info_json + chmod +w $temp_info_json + python3 ${project_path}/cmake/util/insert_op_info.py ${custom_new_info_json} ${temp_info_json} + cp -f $temp_info_json $custom_new_info_json + rm -f $temp_info_json +fi diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py new file mode 100755 index 000000000..c38e79edc --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py @@ -0,0 +1,260 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import sys +import os + + +OP_ALL = '__ALLOP__' +SOC_ALL = '__ALLSOC__' +SOC_TO_SHORT_SOC_MAP = { + "ascend910a": "ascend910", + "ascend910proa": "ascend910", + "ascend910b": "ascend910", + "ascend910prob": "ascend910", + "ascend910premiuma": "ascend910", + "ascend910b1": "ascend910b", + "ascend910b2": "ascend910b", + "ascend910b3": "ascend910b", + "ascend910b4": "ascend910b", + "ascend310p1": "ascend310p", + "ascend310p3": "ascend310p", + "ascend310p3vir01": "ascend310p", + "ascend310p3vir02": "ascend310p", + "ascend310p3vir04": "ascend310p", + "ascend310p3vir08": "ascend310p", + "ascend310b1": "ascend310b", + "bs9sx1aa": "bs9sx1a" +} + + +class OpDesc: + def __init__(self: any, op_type: str): + self.op_type = op_type + self.attr_list = [] + self.attr_val = {} + self.input_name = [] + self.input_type = [] + self.input_dtype = [] + self.input_fmt = [] + self.output_name = [] + self.output_type = [] + self.output_dtype = [] + self.output_fmt = [] + self.op_fmt_sel = False + self.op_chk_support = False + self.op_intf = '' + self.kern_name = '' + self.op_file = '' + self.op_replay_flag = False + self.op_replay_batch = False + self.input_idx = -1 + self.output_idx = -1 + self.max_block_dim = 32 + self.max_shape_size = 268435456 + self.dynamic_shape = False + self.op_range_limit = '' + self.custom_compile_options = {} + self.custom_all_compile_options = {} + + @staticmethod + def _parse_digit(conf: str) -> int: + return int(conf.split('=')[1]) + + @staticmethod + def _parse_flag(conf: str) -> bool: + if 'true' == conf.split('=')[1]: + return True + return False + + @staticmethod + def _parse_str(conf: str) -> str: + return conf.split('=')[1] + + @staticmethod + def _parse_list(conf: str) -> list: + return conf.split('=')[1].split(',') + + def parse_input(self: any, conf: str): + if conf.startswith('input{}.name'.format(int(self.input_idx) + 1)): + self.input_idx += 1 + self.input_name.append(self._parse_str(conf)) + elif conf.startswith('input{}.paramType'.format(int(self.input_idx))): + self.input_type.append(self._parse_str(conf)) + elif conf.startswith('input{}.dtype'.format(int(self.input_idx))): + self.input_dtype.append(self._parse_str(conf)) + elif conf.startswith('input{}.format'.format(int(self.input_idx))): + self.input_fmt.append(self._parse_str(conf)) + else: + return + + def parse_output(self: any, conf: str): + if conf.startswith('output{}.name'.format(int(self.output_idx) + 1)): + self.output_idx += 1 + self.output_name.append(self._parse_str(conf)) + elif conf.startswith('output{}.paramType'.format(int(self.output_idx))): + self.output_type.append(self._parse_str(conf)) + elif conf.startswith('output{}.dtype'.format(int(self.output_idx))): + self.output_dtype.append(self._parse_str(conf)) + elif conf.startswith('output{}.format'.format(int(self.output_idx))): + self.output_fmt.append(self._parse_str(conf)) + else: + return + + def parse_op_format(self: any, conf: str): + self.op_fmt_sel = self._parse_flag(conf) + + def parse_check_support(self: any, conf: str): + self.op_chk_support = self._parse_flag(conf) + + def parse_range_limit(self: any, conf: str): + self.op_range_limit = self._parse_str(conf) + + def parse_kern_name(self: any, conf: str): + self.kern_name = self._parse_str(conf) + + def parse_op_intf(self: any, conf: str): + self.op_intf = self._parse_str(conf) + + def parse_op_file(self: any, conf: str): + self.op_file = self._parse_str(conf) + + def parse_dynamic_shape(self: any, conf: str): + self.dynamic_shape = self._parse_flag(conf) + + def parse_attr_list(self: any, conf: str): + self.attr_list = self._parse_list(conf) + + def parse_attr_val(self: any, conf: str): + for attr in self.attr_list: + if self.attr_val.get(attr) is None: + self.attr_val[attr] = {} + if conf.startswith('attr_{}.type'.format(attr)): + self.attr_val.get(attr)['type'] = self._parse_str(conf) + elif conf.startswith('attr_{}.paramType'.format(attr)): + self.attr_val.get(attr)['paramType'] = self._parse_str(conf) + elif conf.startswith('attr_{}.defaultValue'.format(attr)): + self.attr_val.get(attr)['defaultValue'] = self._parse_str(conf) + + def parse_replay_val(self: any, batch_list: list, iterator_list: list): + if self.op_type in batch_list: + self.op_replay_flag = True + self.op_replay_batch = True + elif self.op_type in iterator_list: + self.op_replay_flag = True + self.op_replay_batch = False + + +def _is_op_type_in_opdesc(op_descs: list, op_type: str): + for op in op_descs: + if op_type == op.op_type: + return True + return False + + +def _set_all_options_to_opdescs(op_descs, soc_ver_compile_options): + for op in op_descs: + op.custom_all_compile_options = soc_ver_compile_options + + +def _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options): + for op in op_descs: + if op.op_type != op_type: + continue + op.custom_compile_options = soc_ver_compile_options + + +def _trans_soc_ver_to_short(soc_ver: str): + low_soc_ver = soc_ver.lower() + if low_soc_ver not in SOC_TO_SHORT_SOC_MAP: + print(f'WARNING: caution: {soc_ver} will trans into ascend910, if not your intention,' + f'use ascend910b1~4 instead') + return SOC_TO_SHORT_SOC_MAP[low_soc_ver] + + +def _get_op_custom_options(op_descs: list, auto_gen_dir: str): + if auto_gen_dir is None: + return {} + file = os.path.join(auto_gen_dir, "custom_compile_options.ini") + if not os.path.exists(file): + print(f'WARNING: cannot find {auto_gen_dir}/custom_compile_options.ini') + return {} + with open (file, 'r') as fd: + lines = fd.readlines() + for line in lines: + param_list = str.split(line.rstrip('\n'), ',') + if len(param_list) != 3: + raise Exception(f'ERROR: custom compile option {param_list} len is not 3') + op_type = param_list[0] + if op_type.upper() == 'ALL': + op_type = OP_ALL + if op_type != OP_ALL and _is_op_type_in_opdesc(op_descs, op_type) == False: + print(f'WARNING: op: {op_type} are not exists in this project') + continue + soc_ver_compile_options = {} + soc_ver = param_list[1] + options_str = param_list[2] + options = str.split(options_str, ';') + if soc_ver == '': + soc_ver_compile_options[SOC_ALL] = options + else: + soc_ver_list = str.split(soc_ver, ';') + for ver in soc_ver_list: + short_ver = _trans_soc_ver_to_short(ver) + soc_ver_compile_options[short_ver] = options + if op_type == OP_ALL: + _set_all_options_to_opdescs(op_descs, soc_ver_compile_options) + else: + _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options) + + +def get_op_desc(file: str, batch_list: list, iterator_list: list, builder: any, + op_type: list, auto_gen_dir: str = None) -> list: + op_descs = [] + op_match = False + with open (file, 'r') as fd: + lines = fd.readlines() + for line in lines: + line = line.strip() + if line.startswith('['): + name = line[1:-1] + if op_type is None or name in op_type: + op_match = True + op_desc = builder(name) + op_desc.parse_replay_val(batch_list, iterator_list) + op_descs.append(op_desc) + else: + op_match = False + if op_type is not None and len(op_descs) == len(op_type): + return op_descs + continue + if not op_match: + continue + if line.startswith('input'): + op_desc.parse_input(line) + elif line.startswith('output'): + op_desc.parse_output(line) + elif line.startswith('dynamicFormat.flag'): + op_desc.parse_op_format(line) + elif line.startswith('needCheckSupport.flag'): + op_desc.parse_check_support(line) + elif line.startswith('rangeLimit.value'): + op_desc.parse_range_limit(line) + elif line.startswith('opInterface.value'): + op_desc.parse_op_intf(line) + elif line.startswith('kernel.name'): + op_desc.parse_kern_name(line) + elif line.startswith('opFile.value'): + op_desc.parse_op_file(line) + elif line.startswith('dynamicShapeSupport.flag'): + op_desc.parse_dynamic_shape(line) + elif line.startswith('attr.list'): + op_desc.parse_attr_list(line) + elif line.startswith('attr_'): + op_desc.parse_attr_val(line) + _get_op_custom_options(op_descs, auto_gen_dir) + return op_descs diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py new file mode 100755 index 000000000..721465fee --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py @@ -0,0 +1,338 @@ +# Copyright 2020-2021 Huawei Technologies Co., Ltd +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +""" +parser ini to json +""" + +import json +import os +import stat +import sys + + +ATTR_TYPE_LIST = ["int", "float", "bool", "str", "listInt", "listFloat", "listBool", "listStr", "listListInt", + "type", "listType", "tensor", "listTensor"] +ATTR_PARAMTYPE_LIST = ["optional", "required"] +BOOL_FLAG_KEY = ["dynamicFormat", "dynamicShapeSupport", "dynamicRankSupport", "precision_reduce", "heavyOp", + "needCheckSupport"] +BOOL_LIST = ["true", "false"] +DTYPE_LIST = ["float16", "float", "float32", "int8", "int16", "int32", "uint8", "uint16", "uint32", "bool", + "int64", "uint64", "qint8", "qint16", "qint32", "quint8", "quint16", "double", "complex64", + "complex128", "string", "resource", "dual", "dual_sub_int8", "dual_sub_uint8", "string_ref", + "int4", "bfloat16", "uint1"] +FORMAT_LIST = ["NCHW", "NHWC", "ND", "NC1HWC0", "FRACTAL_Z", "NC1C0HWPAD", "NHWC1C0", "FSR_NCHW", "FRACTAL_DECONV", + "C1HWNC0", "FRACTAL_DECONV_TRANSPOSE", "FRACTAL_DECONV_SP_STRIDE_TRANS", "NC1HWC0_C04", + "FRACTAL_Z_C04", "CHWN", "FRACTAL_DECONV_SP_STRIDE8_TRANS", "HWCN", "NC1KHKWHWC0", "BN_WEIGHT", + "FILTER_HWCK", "HASHTABLE_LOOKUP_LOOKUPS", "HASHTABLE_LOOKUP_KEYS", "HASHTABLE_LOOKUP_VALUE", + "HASHTABLE_LOOKUP_OUTPUT", "HASHTABLE_LOOKUP_HITS", "C1HWNCoC0", "MD", "NDHWC", "FRACTAL_ZZ", + "FRACTAL_NZ", "NCDHW", "DHWCN", "NDC1HWC0", "FRACTAL_Z_3D", "CN", "NC", "DHWNC", + "FRACTAL_Z_3D_TRANSPOSE", "FRACTAL_ZN_LSTM", "FRACTAL_ZN_RNN", "FRACTAL_Z_G", "NULL"] + + +def parse_ini_files(ini_files): + """ + parse ini files to json + Parameters: + ---------------- + ini_files:input file list + return:ops_info + ---------------- + """ + tbe_ops_info = {} + for ini_file in ini_files: + check_file_size(ini_file) + parse_ini_to_obj(ini_file, tbe_ops_info) + return tbe_ops_info + + +def check_file_size(input_file): + try: + file_size = os.path.getsize(input_file) + except OSError as os_error: + print('[ERROR] Failed to open "%s". %s' % (input_file, str(os_error))) + raise OSError from os_error + if file_size > 10*1024*1024: + print('[WARN] The size of %s exceeds 10MB, it may take more time to run, please wait.' % input_file) + + +def parse_ini_to_obj(ini_file, tbe_ops_info): + """ + parse ini file to json obj + Parameters: + ---------------- + ini_file:ini file path + tbe_ops_info:ops_info + ---------------- + """ + with open(ini_file) as ini_file: + lines = ini_file.readlines() + op_dict = {} + op_name = "" + find_op_type = False + for line in lines: + line = line.rstrip() + if line == "": + continue + if line.startswith("["): + if line.endswith("]"): + op_name = line[1:-1] + op_dict = {} + tbe_ops_info[op_name] = op_dict + find_op_type = True + elif "=" in line: + key1 = line[:line.index("=")] + key2 = line[line.index("=")+1:] + key1_0, key1_1 = key1.split(".") + if key1_0 not in op_dict: + op_dict[key1_0] = {} + if key1_1 in op_dict.get(key1_0): + raise RuntimeError("Op:" + op_name + " " + key1_0 + " " + + key1_1 + " is repeated!") + dic_key = op_dict.get(key1_0) + dic_key[key1_1] = key2 + else: + continue + if not find_op_type: + raise RuntimeError("Not find OpType in .ini file.") + + +def check_output_exist(op_dict, is_valid): + """ + Function Description: + Check output is exist + Parameter: op_dict + Parameter: is_valid + """ + if "output0" in op_dict: + output0_dict = op_dict.get("output0") + if output0_dict.get("name", None) is None: + is_valid = False + print("output0.name is required in .ini file!") + else: + is_valid = False + print("output0 is required in .ini file!") + return is_valid + + +def check_attr_dict(attr_dict, is_valid, attr): + """ + Function Description: + Check attr_dict + Parameter: attr_dict + Parameter: is_valid + Parameter: attr + """ + attr_type = attr_dict.get("type") + value = attr_dict.get("value") + param_type = attr_dict.get("paramType") + if attr_type is None or value is None: + is_valid = False + print("If attr.list is exist, {0}.type and {0}.value is required".format(attr)) + if param_type and param_type not in ATTR_PARAMTYPE_LIST: + is_valid = False + print("{0}.paramType only support {1}.".format(attr, ATTR_PARAMTYPE_LIST)) + if attr_type and attr_type not in ATTR_TYPE_LIST: + is_valid = False + print("{0}.type only support {1}.".format(attr, ATTR_TYPE_LIST)) + return is_valid + + +def check_attr(op_dict, is_valid): + """ + Function Description: + Check attr + Parameter: op_dict + Parameter: is_valid + """ + if "attr" in op_dict: + attr_dict = op_dict.get("attr") + attr_list_str = attr_dict.get("list", None) + if attr_list_str is None: + is_valid = False + print("attr.list is required in .ini file!") + else: + attr_list = attr_list_str.split(",") + for attr_name in attr_list: + attr = "attr_" + attr_name.strip() + attr_dict = op_dict.get(attr) + if attr_dict: + is_valid = check_attr_dict(attr_dict, is_valid, attr) + else: + is_valid = False + print("%s is required in .ini file, when attr.list is %s!" % (attr, attr_list_str)) + return is_valid + + +def check_bool_flag(op_dict, is_valid): + """ + Function Description: + check_bool_flag + Parameter: op_dict + Parameter: is_valid + """ + for key in BOOL_FLAG_KEY: + if key in op_dict: + op_bool_key = op_dict.get(key) + if op_bool_key.get("flag").strip() not in BOOL_LIST: + is_valid = False + print("{0}.flag only support {1}.".format(key, BOOL_LIST)) + return is_valid + + +def check_type_format(op_info, is_valid, op_info_key): + """ + Function Description: + Check type and format + Parameter: op_info + Parameter: is_valid + Parameter: op_info_key + """ + op_info_dtype_str = op_info.get("dtype") + op_info_dtype_num = 0 + op_info_format_num = 0 + if op_info_dtype_str: + op_info_dtype = op_info_dtype_str.split(",") + op_info_dtype_num = len(op_info_dtype) + for dtype in op_info_dtype: + if dtype.strip() not in DTYPE_LIST: + is_valid = False + print("{0}.dtype not support {1}.".format(op_info_key, dtype)) + op_info_format_str = op_info.get("format") + if op_info_format_str: + op_info_format = op_info_format_str.split(",") + op_info_format_num = len(op_info_format) + for op_format in op_info_format: + if op_format.strip() not in FORMAT_LIST: + is_valid = False + print("{0}.format not support {1}.".format(op_info_key, op_format)) + if op_info_dtype_num > 0 and op_info_format_num > 0: + if op_info_dtype_num != op_info_format_num: + is_valid = False + print("The number of {0}.dtype not match the number of {0}.format.".format(op_info_key)) + return is_valid + + +def check_op_info(tbe_ops): + """ + Function Description: + Check info. + Parameter: tbe_ops + Return Value: is_valid + """ + print("\n\n==============check valid for ops info start==============") + required_op_input_info_keys = ["paramType", "name"] + required_op_output_info_keys = ["paramType", "name"] + param_type_valid_value = ["dynamic", "optional", "required"] + is_valid = True + for op_key in tbe_ops: + op_dict = tbe_ops[op_key] + for op_info_key in op_dict: + if op_info_key.startswith("input"): + op_input_info = op_dict[op_info_key] + missing_keys = [] + for required_op_input_info_key in required_op_input_info_keys: + if required_op_input_info_key not in op_input_info: + missing_keys.append(required_op_input_info_key) + if len(missing_keys) > 0: + print("op: " + op_key + " " + op_info_key + " missing: " + + ",".join(missing_keys)) + is_valid = False + else: + if not op_input_info["paramType"] in param_type_valid_value: + print("op: " + op_key + " " + op_info_key + \ + " paramType not valid, valid key:[dynamic, " + "optional, required]") + is_valid = False + is_valid = check_type_format(op_input_info, is_valid, op_info_key) + if op_info_key.startswith("output"): + op_input_info = op_dict[op_info_key] + missing_keys = [] + for required_op_input_info_key in required_op_output_info_keys: + if required_op_input_info_key not in op_input_info: + missing_keys.append(required_op_input_info_key) + if len(missing_keys) > 0: + print("op: " + op_key + " " + op_info_key + " missing: " + + ",".join(missing_keys)) + is_valid = False + else: + if not op_input_info["paramType"] in param_type_valid_value: + print("op: " + op_key + " " + op_info_key + + " paramType not valid, valid key:[dynamic, " + "optional, required]") + is_valid = False + is_valid = check_type_format(op_input_info, is_valid, op_info_key) + is_valid = check_attr(op_dict, is_valid) + is_valid = check_bool_flag(op_dict, is_valid) + print("==============check valid for ops info end================\n\n") + return is_valid + + +def write_json_file(tbe_ops_info, json_file_path): + """ + Save info to json file + Parameters: + ---------------- + tbe_ops_info: ops_info + json_file_path: json file path + ---------------- + """ + json_file_real_path = os.path.realpath(json_file_path) + wr_flag = os.O_WRONLY | os.O_CREAT + wr_mode = stat.S_IWUSR | stat.S_IRUSR + with os.fdopen(os.open(json_file_real_path, wr_flag, wr_mode), 'w') as file_path: + # Only the owner and group have rights + os.chmod(json_file_real_path, stat.S_IWGRP + stat.S_IWUSR + stat.S_IRGRP + + stat.S_IRUSR) + json.dump(tbe_ops_info, file_path, sort_keys=True, indent=4, + separators=(',', ':')) + print("Compile op info cfg successfully.") + + +def parse_ini_to_json(ini_file_paths, outfile_path): + """ + parse ini files to json file + Parameters: + ---------------- + ini_file_paths: list of ini file path + outfile_path: output file path + ---------------- + """ + tbe_ops_info = parse_ini_files(ini_file_paths) + if not check_op_info(tbe_ops_info): + print("Compile op info cfg failed.") + return False + write_json_file(tbe_ops_info, outfile_path) + return True + + +if __name__ == '__main__': + args = sys.argv + + OUTPUT_FILE_PATH = "tbe_ops_info.json" + ini_file_path_list = [] + + for arg in args: + if arg.endswith("ini"): + ini_file_path_list.append(arg) + OUTPUT_FILE_PATH = arg.replace(".ini", ".json") + if arg.endswith("json"): + OUTPUT_FILE_PATH = arg + + if len(ini_file_path_list) == 0: + ini_file_path_list.append("tbe_ops_info.ini") + + if not parse_ini_to_json(ini_file_path_list, OUTPUT_FILE_PATH): + sys.exit(1) + sys.exit(0) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/preset_parse.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/preset_parse.py new file mode 100755 index 000000000..8f1124b1d --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/preset_parse.py @@ -0,0 +1,23 @@ +import json +import sys +import os + + +def get_config_opts(file): + src_dir = os.path.abspath(os.path.dirname(file)) + opts = '' + with open(file, 'r') as fd: + config = json.load(fd) + for conf in config: + if conf == 'configurePresets': + for node in config[conf]: + macros = node.get('cacheVariables') + if macros is not None: + for key in macros: + opts += '-D{}={} '.format(key, macros[key]['value']) + opts = opts.replace('${sourceDir}', src_dir) + print(opts) + + +if __name__ == "__main__": + get_config_opts(sys.argv[1]) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py new file mode 100755 index 000000000..1baa364ef --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py @@ -0,0 +1,105 @@ +#!/usr/bin/env python +# -*- coding: UTF-8 -*- +""" +Created on Feb 28 20:56:45 2020 +Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. +""" + +import os +import stat +import collections +import kernel_entry as keb +from tiling_data_def_build import gen_tiling +import code_channel_infer +import const_var + +PYF_PATH = os.path.dirname(__file__) + +ReplayCodeGenParams = collections.namedtuple('ReplayCodeGenParams',\ +['op_type', 'impl', 'tiling_file', 'kernel', 'entry', 'argn', 'op_replay_batch', 'max_block_dim', 'max_shape_size']) + + +class ReplayCodeGen: + def __init__(self, replayCodeGenParams): + self.op_type = replayCodeGenParams.op_type + self.impl = replayCodeGenParams.impl + self.tiling_file = replayCodeGenParams.tiling_file + self.tiling_data_file = '' + self.kernel = replayCodeGenParams.kernel + self.entry = replayCodeGenParams.entry + self.argn = replayCodeGenParams.argn + self.batch = False + self.outdir = '' + self.data_type = 'uint8_t' + self.blknum = 32 + self.op_replay_batch = replayCodeGenParams.op_replay_batch + self.max_block_dim = replayCodeGenParams.max_block_dim + self.max_shape_size = replayCodeGenParams.max_shape_size + + def set_batch(self, is_batch): + self.batch = is_batch + + def set_outdir(self, outdir): + self.outdir = outdir + + def gen_replay(self, ops_product: str): + kerentry = os.path.join(self.outdir, self.kernel + '_entry.cce') + kerimpl = os.path.join(self.outdir, self.kernel + '_impl.cpp') + replayimpl = os.path.join(self.outdir, self.kernel + '_replay.cpp') + if self.batch: + reptmp = os.path.join(PYF_PATH, 'batch_replay_impl.temp') + else: + reptmp = os.path.join(PYF_PATH, 'replay_impl.temp') + kertmp = os.path.join(PYF_PATH, 'kernel_impl.temp') + self._gen_kentry(kerentry) + self._gen_kimpl_code(kerimpl, kertmp) + self._gen_tiling_data_header() + self._gen_replay_code(replayimpl, reptmp, ops_product) + + def _gen_tiling_data_header(self): + self.tiling_data_file = os.path.join(self.outdir, self.kernel + '_tiling_data.h') + gen_tiling(self.tiling_file, self.tiling_data_file) + + def _gen_kimpl_code(self, src, tmpfile): + with open(tmpfile, 'r') as fd: + temp = fd.read() + temp = temp.replace('__CCE_FILE__', self.impl) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(temp) + + def _gen_replay_code(self, src, tmpfile, ops_product: str): + with open(tmpfile, 'r') as fd: + temp = fd.read() + temp = temp.replace('__ARG_NUM__', str(self.argn)) + argdef = [] + kargs = [] + for i in range(0, self.argn): + argdef.append('{} *'.format(self.data_type)) + kargs.append('({} *)GetArg({})'.format(self.data_type, i)) + temp = temp.replace('__ARGS_DEF__', ', '.join(argdef)) + temp = temp.replace('__KERNEL_ARGS__', ', '.join(kargs)) + temp = temp.replace('__KERNEL_FUN__', self.entry) + core_type_infer = 'core_type' + code_channel = code_channel_infer.infer_code_channel(code_channel_infer.InfoCodeChanelParams(self.impl,\ + self.tiling_data_file, self.kernel, self.outdir, ops_product, None)) + if code_channel == code_channel_infer.CODE_VEC: + core_type_infer = '0' + elif code_channel == code_channel_infer.CODE_CUBE: + core_type_infer = '1' + temp = temp.replace('__CORE_TYPE__', core_type_infer) + # regist function + temp = temp.replace('__OPS_PRODUCT__', ops_product) + temp = temp.replace('__OPTYPE__', self.op_type) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(temp) + + def _gen_kentry(self, src): + kf = '' + pre_alloc_str = 'A' * 256 + if self.batch: + kf += keb.batch_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str), self.argn, self.data_type) + else: + kf += keb.mc_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str),\ + self.argn, self.data_type, self.blknum) + with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(kf) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp b/atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp new file mode 100755 index 000000000..1d30dd865 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp @@ -0,0 +1,120 @@ +#include +#include +#include +#include +#include +#include +#include "replay_def.h" +#include "code_gen.h" +#include "replay_fun.h" +#include "register/op_check.h" +#define __ASCENDC_REPLAY_CODE__ +using namespace std; +using namespace optiling; +using namespace AscendCReplay; + +extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); +extern "C" int elf_append(char *elf, uint32_t elfSize, char *jit, int kernum, int blknum[], char *atext[], + int alen[], int atlen, const char* kernelname[]); + +#define KERNEL_N 1 +#define ARG_N (__ARG_NUM__) +#define MAX_L (1024 * 1024 * 100) +#define MAX_E (1024 * 1024) + +int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) +{ + // gen type 1 : direct call codes 0: load .o file + if (param.gentype < 0 || param.gentype > 1) { + printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); + return 0; + } else if (param.gentype == 1 && param.objptr == nullptr) { + printf("Error: call replay with direct call mode, but code obj addr is null\n"); + return 0; + } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { + printf("Error: call replay with object file mode, but object file path is null\n"); + return 0; + } + // core_type 0:MIX 1:CUBE 2:VEC + if (core_type < 0 || core_type > 2) { + printf("Error: call replay core type is %d !\n", core_type); + return 0; + } + g_coreType = __CORE_TYPE__; + g_taskRation = param.task_ration; + g_tilingKey = param.tiling_key; + + unsigned char *buf, *jit; + char *kernel[KERNEL_N * 32]; + int len[KERNEL_N * 32]; + int blknum[KERNEL_N]; + int max; + block_num = param.block_dim; + g_ubBase = block_num; + uint8_t *code = (uint8_t *)malloc(MAX_L); + uint8_t *pos = code; + struct timespec tp1, tp2; + + clock_gettime(CLOCK_MONOTONIC, &tp1); + if (block_num > 32) { + printf("Error: block_num > 32\n"); + return 0; + } + //__OP_FOPEN__ + for (int i = 0; i < KERNEL_N; i++) { + for (int j = 0; j < ARG_N; j++) + AddArg(j, ARG_STEP * (j + 1)); + for (block_idx = 0; block_idx < block_num; block_idx++) { + //__OP_SET_KERNEL__ + int code_idx = i * block_num + block_idx; +#ifdef FP_CEILING + SetCtrlFloatEnable(); +#else + SetCtrlFloatDisable(); +#endif + CodeInit(pos, false); + __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); + CodeEnd(); + kernel[code_idx] = (char *)pos; + len[code_idx] = CodeLen(); + pos += len[code_idx]; + printf("kernel %d core %ld code generated len %d\n", i, block_idx, len[code_idx]); + } + blknum[i] = block_num; + } + //__OP_FCLOSE__ + clock_gettime(CLOCK_MONOTONIC, &tp2); + buf = (unsigned char *)malloc(MAX_E); + int fd = open(param.entry_file, O_RDONLY); + if (fd < 0) { + printf("[error]: cannot find entry.o : %s\n", param.entry_file); + return 0; + } + uint32_t bufSize = read(fd, buf, MAX_E); + if (bufSize <= 0) { + printf("[error]: entry.o : %s is too small ! \n", param.entry_file); + } + close(fd); + jit = (unsigned char *)malloc(MAX_L); + printf("total code generated %ld\n", pos - code); + int sz = elf_append((char *)buf, bufSize, (char *)jit, KERNEL_N, blknum, kernel, len, pos - code, ¶m.kernel_name); + if (tp1.tv_sec != tp2.tv_sec) { + printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); + } else { + printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); + } + printf("new elf size %d\n", sz); + if (param.gentype == 0) { + fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); + (void)write(fd, jit, sz); + close(fd); + free(jit); + } else if (param.gentype == 1) { + *param.objptr = (char*)jit; + } + free(buf); + free(code); + return sz; +} + +REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py new file mode 100755 index 000000000..a96304261 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py @@ -0,0 +1,84 @@ +#!/usr/bin/env python +# coding=utf-8 +""" +Function: +The replay funtion entry +Copyright Information: +Huawei Technologies Co., Ltd. All Rights Reserved © 2020 +""" + +import sys +import os +import stat +import re +import const_var + + +def gen_tiling(tiling_header_file: str, tiling_file_out: str): + if not os.path.exists(tiling_header_file): + print("warning: no userdef tiling header file: ", tiling_header_file) + return + print("generate tiling def header file: ", tiling_file_out) + tmp_name = os.path.splitext(os.path.basename(tiling_header_file))[0].upper() + tiling_source = '#ifndef __{}_H__\n'.format(tmp_name) + tiling_source += '#define __{}_H__\n\n'.format(tmp_name) + tiling_source += '#include \n' + tiling_source += '#include \n\n' + tiling_source += '#include "kernel_tiling/kernel_tiling.h"\n\n' + end_source = "" + pattern = re.compile(r'[(](.*)[)]', re.S) + with open(tiling_header_file, 'r') as fd: + lines = fd.readlines() + for line in lines: + line = line.strip() + if (line.startswith('BEGIN_TILING_DATA_DEF')): + tiling_source += '#pragma pack(1)\n' + tiling_source += 'struct ' + struct_def = re.findall(pattern, line)[0] + tiling_source += struct_def + ' {\n' + elif (line.startswith('TILING_DATA_FIELD_DEF_ARR')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {}[{}] = {{}};\n'.format(fds[0].strip(), fds[2].strip(), fds[1].strip()) + elif (line.startswith('TILING_DATA_FIELD_DEF_STRUCT')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {};\n'.format(fds[0].strip(), fds[1].strip()) + elif (line.startswith('TILING_DATA_FIELD_DEF')): + field_params = re.findall(pattern, line)[0] + fds = field_params.split(',') + tiling_source += ' {} {} = 0;\n'.format(fds[0].strip(), fds[1].strip()) + elif (line.startswith('END_TILING_DATA_DEF')): + tiling_source += '};\n' + tiling_source += '#pragma pack()\n\n' + tiling_source += '#ifdef __NPU_TILING__\n' + tiling_source += \ + 'inline [aicore] void Init{stru}(const __gm__ uint8_t* tiling, {stru}* const_data)\n'\ + .format(stru=struct_def) + tiling_source += '{\n' + tiling_source += ' const __gm__ uint32_t *src = (const __gm__ uint32_t *)tiling;\n' + tiling_source += ' uint32_t *dst = (uint32_t *)const_data;\n' + tiling_source += ' for (auto i = 0; i < sizeof({}) / 4; i++) *(dst + i) = *(src + i);\n'\ + .format(struct_def) + tiling_source += '}\n' + tiling_source += '#else\n' + tiling_source += 'inline void Init{stru}(uint8_t* tiling, {stru}* const_data)\n'.format(stru=struct_def) + tiling_source += '{\n' + tiling_source += ' memcpy(const_data, tiling, sizeof({}));\n'.format(struct_def) + tiling_source += '}\n' + tiling_source += '#endif\n\n' + end_source = ''' +#define GET_TILING_DATA(tiling_data, tiling_arg) \\ +{stru} tiling_data; \\ +Init{stru}(tiling_arg, &tiling_data)\n +'''.format(stru=struct_def) + tiling_source += end_source + tiling_source += '#endif' + with os.fdopen(os.open(tiling_file_out, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: + ofd.write(tiling_source) + + +if __name__ == '__main__': + if len(sys.argv) <= 2: + raise RuntimeError('arguments must greater than 2') + gen_tiling(sys.argv[1], sys.argv[2]) diff --git a/atb_operator/AtbAddSample/AddAscendC/framework/CMakeLists.txt b/atb_operator/AtbAddSample/AddAscendC/framework/CMakeLists.txt new file mode 100644 index 000000000..b6be9b492 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/framework/CMakeLists.txt @@ -0,0 +1,11 @@ +if(NOT EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/mindspore") + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/caffe_plugin") + add_subdirectory(caffe_plugin) + endif() + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/tf_plugin") + add_subdirectory(tf_plugin) + endif() + if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/onnx_plugin") + add_subdirectory(onnx_plugin) + endif() +endif() diff --git a/atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt b/atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt new file mode 100644 index 000000000..a6aba5c20 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt @@ -0,0 +1,14 @@ + +aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} plugin_srcs) +add_library(cust_tf_parsers SHARED ${plugin_srcs}) +target_compile_definitions(cust_tf_parsers PRIVATE google=ascend_private) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_tf_parsers PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_tf_parsers PRIVATE intf_pub graph) +install(TARGETS cust_tf_parsers + LIBRARY DESTINATION packages/vendors/${vendor_name}/framework/tensorflow +) diff --git a/atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc b/atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc new file mode 100644 index 000000000..2cd837ce5 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc @@ -0,0 +1,23 @@ +/* Copyright (C) 2020-2021. Huawei Technologies Co., Ltd. All +rights reserved. + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the Apache License Version 2.0. + * You may not use this file except in compliance with the License. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * Apache License for more details at + * http://www.apache.org/licenses/LICENSE-2.0 + */ + +#include "register/register.h" + +namespace domi { +// register op info to GE +REGISTER_CUSTOM_OP("AddCustom") + .FrameworkType(TENSORFLOW) // type: CAFFE, TENSORFLOW + .OriginOpType("Add") // name in tf module + .ParseParamsByOperatorFn(AutoMappingByOpFn); +} // namespace domi diff --git a/atb_operator/AtbAddSample/AddAscendC/op_host/CMakeLists.txt b/atb_operator/AtbAddSample/AddAscendC/op_host/CMakeLists.txt new file mode 100644 index 000000000..40dd51cfa --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/op_host/CMakeLists.txt @@ -0,0 +1,82 @@ + +aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} ops_srcs) + +opbuild(OPS_SRC ${ops_srcs} + OUT_DIR ${ASCEND_AUTOGEN_PATH} +) + +add_library(cust_op_proto SHARED ${ops_srcs} ${ASCEND_AUTOGEN_PATH}/op_proto.cc) +target_compile_definitions(cust_op_proto PRIVATE OP_PROTO_LIB) +target_compile_options(cust_op_proto PRIVATE + -fvisibility=hidden +) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_op_proto PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_op_proto PRIVATE + intf_pub + exe_graph + register + tiling_api + -Wl,--whole-archive + rt2_registry + -Wl,--no-whole-archive +) +set_target_properties(cust_op_proto PROPERTIES OUTPUT_NAME + cust_opsproto_rt2.0 +) +add_library(cust_optiling SHARED ${ops_srcs}) +target_compile_definitions(cust_optiling PRIVATE OP_TILING_LIB) +target_compile_options(cust_optiling PRIVATE + -fvisibility=hidden +) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_optiling PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_optiling PRIVATE + intf_pub + exe_graph + register + tiling_api + -Wl,--whole-archive + rt2_registry + -Wl,--no-whole-archive +) +set_target_properties(cust_optiling PROPERTIES OUTPUT_NAME + cust_opmaster_rt2.0 +) + +file(GLOB aclnn_src ${ASCEND_AUTOGEN_PATH}/aclnn_*.cpp) +file(GLOB aclnn_inc ${ASCEND_AUTOGEN_PATH}/aclnn_*.h) +add_library(cust_opapi SHARED ${aclnn_src}) +if(ENABLE_CROSS_COMPILE) + target_link_directories(cust_opapi PRIVATE + ${CMAKE_COMPILE_COMPILER_LIBRARY} + ${CMAKE_COMPILE_RUNTIME_LIBRARY} + ) +endif() +target_link_libraries(cust_opapi PRIVATE intf_pub ascendcl nnopbase) + +add_custom_target(optiling_compat ALL + COMMAND ln -sf lib/linux/${CMAKE_SYSTEM_PROCESSOR}/$ + ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so +) + +install(TARGETS cust_op_proto + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_proto/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) +install(FILES ${ASCEND_AUTOGEN_PATH}/op_proto.h + DESTINATION packages/vendors/${vendor_name}/op_proto/inc) +install(TARGETS cust_optiling + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) +install(FILES ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so + DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling) +install(TARGETS cust_opapi + LIBRARY DESTINATION packages/vendors/${vendor_name}/op_api/lib) +install(FILES ${aclnn_inc} + DESTINATION packages/vendors/${vendor_name}/op_api/include) diff --git a/atb_operator/AtbAddSample/AddAscendC/op_host/add_custom.cpp b/atb_operator/AtbAddSample/AddAscendC/op_host/add_custom.cpp new file mode 100644 index 000000000..b61a67f11 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/op_host/add_custom.cpp @@ -0,0 +1,117 @@ +#include "add_custom_tiling.h" +#include "register/op_def_registry.h" +#include "graph/utils/type_utils.h" +#include "tiling/platform/platform_ascendc.h" + +namespace optiling { +const uint32_t BLOCK_SIZE = 32; +const uint32_t BUFFER_NUM = 2; +static ge::graphStatus TilingFunc(gert::TilingContext* context) +{ + TilingData tiling; + uint64_t ubSize; + auto ascendcPlatform = platform_ascendc::PlatformAscendC(context->GetPlatformInfo()); + ascendcPlatform.GetCoreMemSize(platform_ascendc::CoreMemType::UB, ubSize); + auto coreNum = ascendcPlatform.GetCoreNum(); + + // Based on the input length and the number of inputs, the number of bytes of the input data type is obtained + uint32_t inputNum = context->GetInputShape(0)->GetStorageShape().GetShapeSize(); + uint32_t typeLength = 0; + ge::TypeUtils::GetDataTypeLength(context->GetInputDesc(0)->GetDataType(), typeLength); + uint32_t inputLength = inputNum * typeLength; + uint32_t inputBytes = inputLength / inputNum; + + // There are a total of 3 shared UB spaces in the input and output. If it's int8, there are 2 more TBUFs + uint32_t ubDataNumber = (inputBytes == 1) ? 5 : 3; + // The number of 32B data blocks that can be used for each data. DOUBLE BUFFER is already counted here + uint32_t tileBlockNum = (ubSize / BLOCK_SIZE / BUFFER_NUM) / ubDataNumber; + uint32_t tileDataNum = (tileBlockNum * BLOCK_SIZE) / inputBytes; + + // Input data for 32B alignment + uint32_t inputLengthAlgin32 = (((inputLength + BLOCK_SIZE - 1) / BLOCK_SIZE) * BLOCK_SIZE); + // There is at least 32B of data on each core, satisfying several settings for several cores. The maximum number of audits is the actual number of audits + coreNum = (coreNum < inputLengthAlgin32 / BLOCK_SIZE) ? coreNum : inputLengthAlgin32 / BLOCK_SIZE; + coreNum = (coreNum >= 1) ? coreNum : 1; + uint32_t everyCoreInputBlockNum = inputLengthAlgin32 / BLOCK_SIZE / coreNum; + uint32_t tailBlockNum = (inputLengthAlgin32 / BLOCK_SIZE) % coreNum; + + // Small chunks are calculated and sliced several times using the number of data on each core + uint32_t smallCoreDataNum = everyCoreInputBlockNum * BLOCK_SIZE / inputBytes; + uint32_t smallTileNum = everyCoreInputBlockNum / tileBlockNum; + uint32_t finalSmallTileNum = (everyCoreInputBlockNum % tileBlockNum) == 0 ? smallTileNum : smallTileNum + 1; + // Tail block calculation for small chunks of data + uint32_t smallTailDataNum = smallCoreDataNum - (tileDataNum * smallTileNum); + smallTailDataNum = smallTailDataNum == 0 ? tileDataNum : smallTailDataNum; + + // The total length of a large block of data is 32B larger than that of a small block of data + everyCoreInputBlockNum += 1; + uint32_t bigCoreDataNum = everyCoreInputBlockNum * BLOCK_SIZE / inputBytes; + uint32_t bigTileNum = everyCoreInputBlockNum / tileBlockNum; + uint32_t finalBigTileNum = (everyCoreInputBlockNum % tileBlockNum) == 0 ? bigTileNum : bigTileNum + 1; + uint32_t bigTailDataNum = bigCoreDataNum - tileDataNum * bigTileNum; + bigTailDataNum = bigTailDataNum == 0 ? tileDataNum : bigTailDataNum; + + tiling.set_smallCoreDataNum(smallCoreDataNum); + tiling.set_bigCoreDataNum(bigCoreDataNum); + tiling.set_tileDataNum(tileDataNum); + tiling.set_smallTailDataNum(smallTailDataNum); + tiling.set_bigTailDataNum(bigTailDataNum); + tiling.set_finalSmallTileNum(finalSmallTileNum); + tiling.set_finalBigTileNum(finalBigTileNum); + tiling.set_tailBlockNum(tailBlockNum); + + context->SetBlockDim(coreNum); + tiling.SaveToBuffer(context->GetRawTilingData()->GetData(), context->GetRawTilingData()->GetCapacity()); + context->GetRawTilingData()->SetDataSize(tiling.GetDataSize()); + size_t *currentWorkspace = context->GetWorkspaceSizes(1); + currentWorkspace[0] = 0; + return ge::GRAPH_SUCCESS; +} +} + +namespace ge { +static ge::graphStatus InferShape(gert::InferShapeContext* context) +{ + const gert::Shape* x1_shape = context->GetInputShape(0); + gert::Shape* y_shape = context->GetOutputShape(0); + *y_shape = *x1_shape; + return GRAPH_SUCCESS; +} +static graphStatus InferDataType(gert::InferDataTypeContext* context) +{ + const auto inputDataType = context->GetInputDataType(0); + context->SetOutputDataType(0, inputDataType); + return ge::GRAPH_SUCCESS; +} +} + +namespace ops { +class AddCustom : public OpDef { +public: + explicit AddCustom(const char* name) : OpDef(name) + { + this->Input("x") + .ParamType(REQUIRED) + .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) + .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) + .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); + this->Input("y") + .ParamType(REQUIRED) + .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) + .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) + .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); + this->Output("z") + .ParamType(REQUIRED) + .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) + .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) + .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); + + this->SetInferShape(ge::InferShape).SetInferDataType(ge::InferDataType); + this->AICore() + .SetTiling(optiling::TilingFunc) + .AddConfig("ascend310b") + .AddConfig("ascend910b"); + } +}; +OP_ADD(AddCustom); +} diff --git a/atb_operator/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h b/atb_operator/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h new file mode 100644 index 000000000..28fd00a37 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h @@ -0,0 +1,22 @@ +/* + * Copyright (c) Huawei Technologies Co., Ltd. 2022-2023. All rights reserved. + */ +#ifndef ADD_CUSTOM_TILING_H +#define ADD_CUSTOM_TILING_H +#include "register/tilingdata_base.h" + +namespace optiling { +BEGIN_TILING_DATA_DEF(TilingData) + TILING_DATA_FIELD_DEF(uint32_t, smallCoreDataNum); + TILING_DATA_FIELD_DEF(uint32_t, bigCoreDataNum); + TILING_DATA_FIELD_DEF(uint32_t, finalBigTileNum); + TILING_DATA_FIELD_DEF(uint32_t, finalSmallTileNum); + TILING_DATA_FIELD_DEF(uint32_t, tileDataNum); + TILING_DATA_FIELD_DEF(uint32_t, smallTailDataNum); + TILING_DATA_FIELD_DEF(uint32_t, bigTailDataNum); + TILING_DATA_FIELD_DEF(uint32_t, tailBlockNum); +END_TILING_DATA_DEF; + +REGISTER_TILING_DATA_CLASS(AddCustom, TilingData) +} +#endif // ADD_CUSTOM_TILING_H \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt b/atb_operator/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt new file mode 100644 index 000000000..0d31a444c --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt @@ -0,0 +1,61 @@ +# set custom compile options +if ("${CMAKE_BUILD_TYPE}x" STREQUAL "Debugx") + add_ops_compile_options(ALL OPTIONS -g -O0) +endif() + +foreach(compute_unit ${ASCEND_COMPUTE_UNIT}) + + # generate aic-${compute_unit}-ops-info.json + add_ops_info_target(TARGET ops_info_gen_${compute_unit} + OUTPUT ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core/${compute_unit}/aic-${compute_unit}-ops-info.json + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/config/${compute_unit} + ) + + # generate ascendc impl py once + if (NOT TARGET ascendc_impl_gen) + add_ops_impl_target(TARGET ascendc_impl_gen + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl + ) + endif() + + # dynamic shape binary compile + if (${ENABLE_BINARY_PACKAGE}) + add_bin_compile_target(TARGET ascendc_bin_${compute_unit} + OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini + IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} + ADP_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/dynamic + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/binary/${compute_unit} + INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/kernel + COMPUTE_UNIT ${compute_unit} + ) + add_dependencies(ascendc_bin_${compute_unit} ascendc_impl_gen) + endif() + +endforeach() + +# generate npu_supported_ops.json +add_npu_support_target(TARGET npu_supported_ops + OPS_INFO_DIR ${ASCEND_AUTOGEN_PATH} + OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core + INSTALL_DIR packages/vendors/${vendor_name}/framework/${ASCEND_FRAMEWORK_TYPE} +) + +if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) + add_subdirectory(testcases) +endif() + +# install kernel file +if (${ENABLE_SOURCE_PACKAGE}) + file(GLOB KERNEL_FILES + ${CMAKE_CURRENT_SOURCE_DIR}/*.cpp + ${CMAKE_CURRENT_SOURCE_DIR}/*.h + ${CMAKE_CURRENT_SOURCE_DIR}/*.py + ) + install(FILES ${KERNEL_FILES} + DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl/dynamic + ) +endif() diff --git a/atb_operator/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp b/atb_operator/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp new file mode 100644 index 000000000..37c80a825 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp @@ -0,0 +1,134 @@ +/* + * Copyright (c) Huawei Technologies Co., Ltd. 2022-2023. All rights reserved. + * + * Function : z = x + y + * This sample is a very basic sample that implements vector add on Ascend plaform. + */ +#include "kernel_operator.h" +// tensor num for each queue +constexpr int32_t BUFFER_NUM = 2; + +template class KernelAdd { + using T = TYPE_X; +public: + __aicore__ inline KernelAdd() {} + __aicore__ inline void Init(GM_ADDR x, GM_ADDR y, GM_ADDR z, uint32_t smallCoreDataNum, + uint32_t bigCoreDataNum, uint32_t finalBigTileNum, + uint32_t finalSmallTileNum, uint32_t tileDataNum, + uint32_t smallTailDataNum, uint32_t bigTailDataNum, + uint32_t tailBlockNum) + { + ASSERT(AscendC::GetBlockNum() != 0 && "block dim can not be zero!"); + uint32_t coreNum = AscendC::GetBlockIdx(); + uint32_t globalBufferIndex = bigCoreDataNum * AscendC::GetBlockIdx(); + this->tileDataNum = tileDataNum; + if (coreNum < tailBlockNum) { + this->coreDataNum = bigCoreDataNum; + this->tileNum = finalBigTileNum; + this->tailDataNum = bigTailDataNum; + } + else { + this->coreDataNum = smallCoreDataNum; + this->tileNum = finalSmallTileNum; + this->tailDataNum = smallTailDataNum; + globalBufferIndex -= (bigCoreDataNum - smallCoreDataNum) * (AscendC::GetBlockIdx() - tailBlockNum); + } + xGm.SetGlobalBuffer((__gm__ TYPE_X*)x + globalBufferIndex, this->coreDataNum); + yGm.SetGlobalBuffer((__gm__ TYPE_Y*)y + globalBufferIndex, this->coreDataNum); + zGm.SetGlobalBuffer((__gm__ TYPE_Z*)z + globalBufferIndex, this->coreDataNum); + pipe.InitBuffer(inQueueX, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_X)); + pipe.InitBuffer(inQueueY, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_Y)); + pipe.InitBuffer(outQueueZ, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_Z)); + pipe.InitBuffer(tmp1, this->tileDataNum * sizeof(half)); + pipe.InitBuffer(tmp2, this->tileDataNum * sizeof(half)); + } + __aicore__ inline void Process() + { + int32_t loopCount = this->tileNum; + this->processDataNum = this->tileDataNum; + for (int32_t i = 0; i < loopCount; i++) { + if (i == this->tileNum - 1) { + this->processDataNum = this->tailDataNum; + } + CopyIn(i); + Compute(i); + CopyOut(i); + } + } + +private: + __aicore__ inline void CopyIn(int32_t progress) + { + AscendC::LocalTensor xLocal = inQueueX.AllocTensor(); + AscendC::LocalTensor yLocal = inQueueY.AllocTensor(); + AscendC::DataCopy(xLocal, xGm[progress * this->tileDataNum], this->processDataNum); + AscendC::DataCopy(yLocal, yGm[progress * this->tileDataNum], this->processDataNum); + inQueueX.EnQue(xLocal); + inQueueY.EnQue(yLocal); + } + __aicore__ inline void Compute(int32_t progress) + { + AscendC::LocalTensor xLocal = inQueueX.DeQue(); + AscendC::LocalTensor yLocal = inQueueY.DeQue(); + AscendC::LocalTensor zLocal = outQueueZ.AllocTensor(); + if constexpr (std::is_same_v) { + auto p1 = tmp1.Get(); + auto p2 = tmp2.Get(); + AscendC::Cast(p1, xLocal, AscendC::RoundMode::CAST_NONE, this->processDataNum); + AscendC::Cast(p2, yLocal, AscendC::RoundMode::CAST_NONE, this->processDataNum); + AscendC::Add(p2, p1, p2, this->processDataNum); + AscendC::Cast(p1.ReinterpretCast(), p2, AscendC::RoundMode::CAST_RINT, this->processDataNum); + AscendC::ShiftLeft(p1.ReinterpretCast(), p1.ReinterpretCast(), int16_t(8), this->processDataNum); + AscendC::ShiftRight(p1.ReinterpretCast(), p1.ReinterpretCast(), int16_t(8), this->processDataNum); + AscendC::Cast(p2, p1.ReinterpretCast(), AscendC::RoundMode::CAST_NONE, this->processDataNum); + AscendC::Cast(zLocal, p2, AscendC::RoundMode::CAST_NONE, this->processDataNum); + } + else { + AscendC::Add(zLocal, xLocal, yLocal, this->processDataNum); + } + outQueueZ.EnQue(zLocal); + inQueueX.FreeTensor(xLocal); + inQueueY.FreeTensor(yLocal); + } + __aicore__ inline void CopyOut(int32_t progress) + { + AscendC::LocalTensor zLocal = outQueueZ.DeQue(); + AscendC::DataCopy(zGm[progress * this->tileDataNum], zLocal, this->processDataNum); + outQueueZ.FreeTensor(zLocal); + } + +private: + AscendC::TPipe pipe; + AscendC::TQue inQueueX, inQueueY; + AscendC::TQue outQueueZ; + AscendC::TBuf tmp1, tmp2; + AscendC::GlobalTensor xGm; + AscendC::GlobalTensor yGm; + AscendC::GlobalTensor zGm; + uint32_t coreDataNum; + uint32_t tileNum; + uint32_t tileDataNum; + uint32_t tailDataNum; + uint32_t processDataNum; +}; + +extern "C" __global__ __aicore__ void add_custom(GM_ADDR x, GM_ADDR y, GM_ADDR z, GM_ADDR workspace, GM_ADDR tiling) +{ + GET_TILING_DATA(tiling_data, tiling); + KernelAdd op; + op.Init(x, y, z, tiling_data.smallCoreDataNum, + tiling_data.bigCoreDataNum, tiling_data.finalBigTileNum, + tiling_data.finalSmallTileNum, tiling_data.tileDataNum, + tiling_data.smallTailDataNum, tiling_data.bigTailDataNum, + tiling_data.tailBlockNum); + op.Process(); +} + +#ifndef ASCENDC_CPU_DEBUG +// call of kernel function +void add_custom_do(uint32_t blockDim, void* l2ctrl, void* stream, uint8_t* x, uint8_t* y, uint8_t* z, + uint8_t* workspace, uint8_t* tiling) +{ + add_custom<<>>(x, y, z, workspace, tiling); +} +#endif diff --git a/atb_operator/AtbAddSample/AddAscendC/scripts/install.sh b/atb_operator/AtbAddSample/AddAscendC/scripts/install.sh new file mode 100644 index 000000000..9d45a8052 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/scripts/install.sh @@ -0,0 +1,318 @@ +#!/bin/bash +vendor_name=customize +targetdir=/usr/local/Ascend/opp +target_custom=0 + +sourcedir=$PWD/packages +vendordir=vendors/$vendor_name + +QUIET="y" + +while true +do + case $1 in + --quiet) + QUIET="y" + shift + ;; + --install-path=*) + INSTALL_PATH=$(echo $1 | cut -d"=" -f2-) + INSTALL_PATH=${INSTALL_PATH%*/} + shift + ;; + --*) + shift + ;; + *) + break + ;; + esac +done + +log() { + cur_date=`date +"%Y-%m-%d %H:%M:%S"` + echo "[runtime] [$cur_date] "$1 +} + +if [ -n "${INSTALL_PATH}" ]; then + if [[ ! "${INSTALL_PATH}" = /* ]]; then + log "[ERROR] use absolute path for --install-path argument" + exit 1 + fi + if [ ! -d ${INSTALL_PATH} ]; then + mkdir ${INSTALL_PATH} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${INSTALL_PATH} failed" + exit 1 + fi + fi + targetdir=${INSTALL_PATH} +elif [ -n "${ASCEND_CUSTOM_OPP_PATH}" ]; then + if [ ! -d ${ASCEND_CUSTOM_OPP_PATH} ]; then + mkdir -p ${ASCEND_CUSTOM_OPP_PATH} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${ASCEND_CUSTOM_OPP_PATH} failed" + fi + fi + targetdir=${ASCEND_CUSTOM_OPP_PATH} +else + if [ "x${ASCEND_OPP_PATH}" == "x" ]; then + log "[ERROR] env ASCEND_OPP_PATH no exist" + exit 1 + fi + targetdir="${ASCEND_OPP_PATH}" +fi + +if [ ! -d $targetdir ];then + log "[ERROR] $targetdir no exist" + exit 1 +fi + +upgrade() +{ + if [ ! -d ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 files" + return 0 + fi + + if [ ! -d ${targetdir}/$vendordir/$1 ];then + log "[INFO] create ${targetdir}/$vendordir/$1." + mkdir -p ${targetdir}/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/$1 failed" + return 1 + fi + else + has_same_file=-1 + for file_a in ${sourcedir}/$vendordir/$1/*; do + file_b=${file_a##*/}; + if [ "ls ${targetdir}/$vendordir/$1" = "" ]; then + log "[INFO] ${targetdir}/$vendordir/$1 is empty !!" + return 1 + fi + grep -q $file_b <<<`ls ${targetdir}/$vendordir/$1`; + if [[ $? -eq 0 ]]; then + echo -n "${file_b} " + has_same_file=0 + fi + done + if [ 0 -eq $has_same_file ]; then + if test $QUIET = "n"; then + echo "[INFO]: has old version in ${targetdir}/$vendordir/$1, \ + you want to Overlay Installation , please enter:[o]; \ + or replace directory installation , please enter: [r]; \ + or not install , please enter:[n]." + + while true + do + read orn + if [ "$orn" = n ]; then + return 0 + elif [ "$orn" = m ]; then + break; + elif [ "$0rn" = r ]; then + [ -n "${targetdir}/$vendordir/$1/" ] && rm -rf "${targetdir}/$vendordir/$1"/* + break; + else + echo "[ERROR] input error, please input again!" + fi + done + fi + fi + log "[INFO] replace or merge old ops $1 files .g....." + fi + + log "copy new ops $1 files ......" + if [ -d ${targetdir}/$vendordir/$1/ ]; then + chmod -R +w "$targetdir/$vendordir/$1/" >/dev/null 2>&1 + fi + cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 files failed" + return 1 + fi + + return 0 +} +upgrade_proto() +{ + if [ ! -f ${sourcedir}/$vendordir/custom.proto ]; then + log "[INFO] no need to upgrade custom.proto files" + return 0 + fi + if [ ! -d ${targetdir}/$vendordir/framework/caffe ];then + log "[INFO] create ${targetdir}/$vendordir/framework/caffe." + mkdir -p ${targetdir}/$vendordir/framework/caffe + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/framework/caffe failed" + return 1 + fi + else + if [ -f ${targetdir}/$vendordir/framework/caffe/custom.proto ]; then + # 有老版本,判断是否要覆盖式安装 + if test $QUIET = "n"; then + echo "[INFO] ${targetdir}/$vendordir/framework/caffe has old version"\ + "custom.proto file. Do you want to replace? [y/n] " + + while true + do + read yn + if [ "$yn" = n ]; then + return 0 + elif [ "$yn" = y ]; then + break; + else + echo "[ERROR] input error, please input again!" + fi + done + fi + fi + log "[INFO] replace old caffe.proto files ......" + fi + chmod -R +w "$targetdir/$vendordir/framework/caffe/" >/dev/null 2>&1 + cp -rf ${sourcedir}/$vendordir/custom.proto ${targetdir}/$vendordir/framework/caffe/ + if [ $? -ne 0 ];then + log "[ERROR] copy new custom.proto failed" + return 1 + fi + log "[INFO] copy custom.proto success" + + return 0 +} + +upgrade_file() +{ + if [ ! -e ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 file" + return 0 + fi + + log "copy new $1 files ......" + cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 file failed" + return 1 + fi + + return 0 +} + +delete_optiling_file() +{ + if [ ! -d ${targetdir}/vendors ];then + log "[INFO] $1 not exist, no need to uninstall" + return 0 + fi + sys_info=$(uname -m) + if [ ! -d ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/lib/linux/${sys_info} ];then + rm -rf ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/liboptiling.so + fi + return 0 +} + +log "[INFO] copy uninstall sh success" + +if [ ! -d ${targetdir}/vendors ];then + log "[INFO] create ${targetdir}/vendors." + mkdir -p ${targetdir}/vendors + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/vendors failed" + return 1 + fi +fi +chmod u+w ${targetdir}/vendors + +echo "[ops_custom]upgrade framework" +upgrade framework +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op proto" +upgrade op_proto +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade version.info" +upgrade_file version.info +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op impl" +delete_optiling_file op_impl +upgrade op_impl +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op api" +upgrade op_api +if [ $? -ne 0 ];then + exit 1 +fi + +upgrade_proto +if [ $? -ne 0 ];then + exit 1 +fi + +# set the set_env.bash +if [ -n "${INSTALL_PATH}" ] && [ -d ${INSTALL_PATH} ]; then + _ASCEND_CUSTOM_OPP_PATH=${targetdir}/${vendordir} + bin_path="${_ASCEND_CUSTOM_OPP_PATH}/bin" + set_env_variable="#!/bin/bash\nexport ASCEND_CUSTOM_OPP_PATH=${_ASCEND_CUSTOM_OPP_PATH}:\${ASCEND_CUSTOM_OPP_PATH}" + if [ ! -d ${bin_path} ]; then + mkdir -p ${bin_path} >> /dev/null 2>&1 + if [ $? -ne 0 ]; then + log "[ERROR] create ${bin_path} failed" + exit 1 + fi + fi + echo -e ${set_env_variable} > ${bin_path}/set_env.bash + if [ $? -ne 0 ]; then + log "[ERROR] write ASCEND_CUSTOM_OPP_PATH to set_env.bash failed" + exit 1 + else + log "[INFO] using requirements: when custom module install finished or before you run the custom module, \ + execute the command [ source ${bin_path}/set_env.bash ] to set the environment path" + fi +else + config_file=${targetdir}/vendors/config.ini + if [ ! -f ${config_file} ]; then + touch ${config_file} + chmod 640 ${config_file} + echo "load_priority=$vendor_name" > ${config_file} + if [ $? -ne 0 ];then + echo "echo load_priority failed" + exit 1 + fi + else + found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" + found_vendor=$(echo $found_vendors | sed "s/\<$vendor_name\>//g" | tr ',' ' ') + vendor=$(echo $found_vendor | tr -s ' ' ',') + if [ "$vendor" != "" ]; then + sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" + fi + fi +fi + +chmod u-w ${targetdir}/vendors + +if [ -d ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/ ]; then + chmod -R 440 ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/* >/dev/null 2>&1 +fi +if [ -f ${targetdir}/ascend_install.info ]; then + chmod -R 440 ${targetdir}/ascend_install.info +fi +if [ -f ${targetdir}/scene.info ]; then + chmod -R 440 ${targetdir}/scene.info +fi +if [ -f ${targetdir}/version.info ]; then + chmod -R 440 ${targetdir}/version.info +fi + +echo "SUCCESS" +exit 0 + diff --git a/atb_operator/AtbAddSample/AddAscendC/scripts/upgrade.sh b/atb_operator/AtbAddSample/AddAscendC/scripts/upgrade.sh new file mode 100644 index 000000000..34801efc4 --- /dev/null +++ b/atb_operator/AtbAddSample/AddAscendC/scripts/upgrade.sh @@ -0,0 +1,151 @@ +#!/bin/bash +vendor_name=customize +targetdir=/usr/local/Ascend/opp +target_custom=0 + +sourcedir=$PWD/packages +vendordir=vendors/$vendor_name + +log() { + cur_date=`date +"%Y-%m-%d %H:%M:%S"` + echo "[runtime] [$cur_date] "$1 +} + +if [[ "x${ASCEND_OPP_PATH}" == "x" ]];then + log "[ERROR] env ASCEND_OPP_PATH no exist" + exit 1 +fi + +targetdir=${ASCEND_OPP_PATH} + +if [ ! -d $targetdir ];then + log "[ERROR] $targetdir no exist" + exit 1 +fi + +upgrade() +{ + if [ ! -d ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 files" + return 0 + fi + + if [ ! -d ${targetdir}/$vendordir/$1 ];then + log "[INFO] create ${targetdir}/$vendordir/$1." + mkdir -p ${targetdir}/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] create ${targetdir}/$vendordir/$1 failed" + return 1 + fi + else + vendor_installed_dir=$(ls "$targetdir/vendors" 2> /dev/null) + for i in $vendor_installed_dir;do + vendor_installed_file=$(ls "$vendor_installed_dir/$vendor_name/$i" 2> /dev/null) + if [ "$i" = "$vendor_name" ] && [ "$vendor_installed_file" != "" ]; then + echo "[INFO]: $vendor_name custom opp package has been installed on the path $vendor_installed_dir, \ + you want to Overlay Installation , please enter:[o]; \ + or replace directory installation , please enter: [r]; \ + or not install , please enter:[n]." + fi + while true + do + read mrn + if [ "$mrn" = m ]; then + break + elif [ "$mrn" = r ]; then + [ -n "$vendor_installed_file"] && rm -rf "$vendor_installed_file" + break + elif [ "$mrn" = n ]; then + return 0 + else + echo "[WARNING]: Input error, please input m or r or n to choose!" + fi + done + done + log "[INFO] replace old ops $1 files ......" + fi + + log "copy new ops $1 files ......" + cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 files failed" + return 1 + fi + + return 0 +} + +upgrade_file() +{ + if [ ! -e ${sourcedir}/$vendordir/$1 ]; then + log "[INFO] no need to upgrade ops $1 file" + return 0 + fi + + log "copy new $1 files ......" + cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 + if [ $? -ne 0 ];then + log "[ERROR] copy new $1 file failed" + return 1 + fi + + return 0 +} + +log "[INFO] copy uninstall sh success" + +echo "[ops_custom]upgrade framework" +upgrade framework +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op proto" +upgrade op_proto +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op impl" +upgrade op_impl +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade op api" +upgrade op_api +if [ $? -ne 0 ];then + exit 1 +fi + +echo "[ops_custom]upgrade version.info" +upgrade_file version.info +if [ $? -ne 0 ];then + exit 1 +fi + +config_file=${targetdir}/vendors/config.ini +found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" +found_vendor=$(echo $found_vendors | sed "s/\<$vendor_name\>//g" | tr ',' ' ') +vendor=$(echo $found_vendor | tr -s ' ' ',') +if [ "$vendor" != "" ]; then + sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" +fi + +changemode() +{ + if [ -d ${targetdir} ];then + chmod -R 550 ${targetdir}>/dev/null 2>&1 + fi + + return 0 +} +echo "[ops_custom]changemode..." +#changemode +if [ $? -ne 0 ];then + exit 1 +fi + +echo "SUCCESS" +exit 0 + diff --git a/atb_operator/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt b/atb_operator/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt new file mode 100644 index 000000000..c89578fcd --- /dev/null +++ b/atb_operator/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt @@ -0,0 +1,21 @@ +# Copyright 2024 Tencent Inc. All rights reserved. +# +# ============================================================================== +cmake_minimum_required(VERSION 3.14) +set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") +file(GLOB_RECURSE atb_plugin_operations_SRCS "./*.cpp") +set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") +message(STATUS "atb_plugin_operations_SRCS: ${atb_plugin_operations_SRCS}") +include_directories( + $ENV{ATB_HOME_PATH}/include + $ENV{ASCEND_HOME_PATH}/include + ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 + ${CMAKE_CURRENT_SOURCE_DIR} + ) +link_directories( + $ENV{ATB_HOME_PATH}/lib + $ENV{ASCEND_HOME_PATH}/lib64 + ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 + ) +add_library(atb_plugin_operations STATIC ${atb_plugin_operations_SRCS}) +target_link_libraries(atb_plugin_operations PRIVATE ${ACL_SHARED_LIBS} ${ACL_SHARED_LIBS} ascendcl) \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp b/atb_operator/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp new file mode 100644 index 000000000..468b84aff --- /dev/null +++ b/atb_operator/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp @@ -0,0 +1,142 @@ +#include "aclnn_add_operation.h" +#include "aclnn_add_custom.h" + +AddOperation::AddOperation(const std::string &name, AddAttrParam param){ + attrParam = param; + opName_ = name; +} + +atb::SVector GetCopyTensorStride(atb::Dims &tensorDims) +{ + atb::SVector tmpStrides(tensorDims.dimNum, 1); + if (tensorDims.dimNum > 8) { // 8: tensor最大维度数量 + printf("tensor's dimNum is larger than 8, GetCopyTensorStride failed."); + return tmpStrides; + } + for (int64_t i = static_cast(tensorDims.dimNum) - 2; i >= 0; i--) { + tmpStrides[i] = (tensorDims.dims[i + 1] * tmpStrides[i + 1]); + } + return tmpStrides; +} + +std::shared_ptr AddOperation::CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx) +{ + auto aclnnTensor = std::make_shared(); + aclnnTensor->tensorIdx = static_cast(tensorIdx); + aclnnTensor->needUpdateTensorDataPtr = true; + aclnnTensor->atbTensor = atbTensor; + aclnnTensor->strides = GetCopyTensorStride(atbTensor.desc.shape); + + // 创建Aclnn tensor + aclnnTensor->tensor = aclCreateTensor(atbTensor.desc.shape.dims, + atbTensor.desc.shape.dimNum, + atbTensor.desc.dtype, + aclnnTensor->strides.data(), + 0, + atbTensor.desc.format, + atbTensor.desc.shape.dims, + atbTensor.desc.shape.dimNum, + atbTensor.deviceData); + return aclnnTensor; +} + +atb::Status AddOperation::UpdateAclnnVariantPack(const atb::VariantPack &variantPack) +{ + // 更新inTensor的device地址 + for (size_t i = 0; i < aclInTensors_.size(); ++i) { + int ret = -1; + if (!aclInTensors_[i]->needUpdateTensorDataPtr) { + continue; + } + aclInTensors_[i]->atbTensor = variantPack.inTensors.at(i); + ret = aclSetInputTensorAddr(aclExecutor_, + aclInTensors_[i]->tensorIdx, + aclInTensors_[i]->tensor, + aclInTensors_[i]->atbTensor.deviceData); + + if (ret != 0) { + printf("set input fail"); + return atb::ERROR_CANN_ERROR; + } + } + + // 更新outTensor的device地址 + for (size_t i = 0; i < aclOutTensors_.size(); ++i) { + int ret = -1; + if (!aclOutTensors_[i]->needUpdateTensorDataPtr) { + continue; + } + aclOutTensors_[i]->atbTensor = variantPack.outTensors.at(i); + ret = aclSetOutputTensorAddr(aclExecutor_, + aclOutTensors_[i]->tensorIdx, + aclOutTensors_[i]->tensor, + aclOutTensors_[i]->atbTensor.deviceData); + + if (ret != 0) { + printf("set output fail"); + return atb::ERROR_CANN_ERROR; + } + } + return atb::NO_ERROR; +} + +atb::Status AddOperation::Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) { + + aclInTensors_.resize(GetInputNum()); + for (size_t i = 0; i < aclInTensors_.size(); ++i) { + auto aclnnTensor = CreateAclnnTensor(variantPack.inTensors.at(i), i); + if (aclnnTensor->tensor == nullptr) { + printf("creat input tensor %ld fail", i); + return atb::ERROR_INTERNAL_ERROR; + } + aclInTensors_[i] = aclnnTensor; + } + + aclOutTensors_.resize(GetOutputNum()); + for (size_t i = 0; i < aclOutTensors_.size(); ++i) { + auto aclnnTensor = CreateAclnnTensor(variantPack.outTensors.at(i), i); + if (aclnnTensor->tensor == nullptr) { + printf("creat output tensor %ld fail", i); + return atb::ERROR_INTERNAL_ERROR; + } + aclOutTensors_[i] = aclnnTensor; + } + + + auto ret = aclnnAddCustomGetWorkspaceSize(aclInTensors_.at(0)->tensor, + aclInTensors_.at(1)->tensor, + aclOutTensors_.at(0)->tensor, + &workspaceSize_, + &aclExecutor_); + + workspaceSize = workspaceSize_; + return ret; + +} + +atb::Status AddOperation::Execute(const atb::VariantPack &variantPack, uint8_t *workspace, uint64_t workspaceSize, atb::Context *context) { + + + + aclrtStream stream = context->GetExecuteStream(); + if (!stream) { + printf("get stream fail"); + return atb::ERROR_INVALID_PARAM; + } + // 更新数据传入的地址 + int ret = UpdateAclnnVariantPack(variantPack); + if (ret != 0) { + printf("UpdateAclnnVariantPack fail"); + return atb::ERROR_CANN_ERROR; + } + ret = aclnnAddCustom(workspace, workspaceSize_, aclExecutor_, stream); + + return ret; +} + +atb::Status AddOperation::InferShape( + const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const +{ + outTensorDesc.at(0) = inTensorDesc.at(0); + return atb::NO_ERROR; +} diff --git a/atb_operator/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h b/atb_operator/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h new file mode 100644 index 000000000..8670088d3 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h @@ -0,0 +1,56 @@ +#include +#include +#include +#include +#include +#include "atb/infer_op_params.h" + + +struct AddAttrParam +{ + // add没属性,此处空 +}; + +struct AclnnTensor +{ +public: + atb::Tensor atbTensor; // + aclTensor *tensor = nullptr; + int tensorIdx = -1; // aclTensor在aclExecutor中的index + bool needUpdateTensorDataPtr = false; + atb::SVector strides = {}; +}; + +class AddOperation: public atb::Operation{ +public: + AddOperation(const std::string &name, AddAttrParam param); + atb::Status Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) override; + atb::Status Execute(const atb::VariantPack &variantPack, uint8_t *workspace, + uint64_t workspaceSize, atb::Context *context) override; + atb::Status InferShape( + const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const; + std::shared_ptr CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx); + atb::Status UpdateAclnnVariantPack(const atb::VariantPack &variantPack); + + uint32_t GetInputNum() const + { + return 2; // 算子入参个数 + } + + uint32_t GetOutputNum() const + { + return 1; // 算子出参个数 + } + std::string GetName() const + { + return opName_; + } + + aclOpExecutor *aclExecutor_ = nullptr; + AddAttrParam attrParam; + std::string opName_; + uint64_t workspaceSize_; + + atb::SVector> aclInTensors_; + atb::SVector> aclOutTensors_; +}; \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AddOperationATBPlugin/build.sh b/atb_operator/AtbAddSample/AddOperationATBPlugin/build.sh new file mode 100755 index 000000000..ba28cc143 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOperationATBPlugin/build.sh @@ -0,0 +1,33 @@ +#!/bin/bash + +# 定义构建目录 +BUILD_DIR="build" + +# 创建构建目录 +mkdir -p "$BUILD_DIR" +cd "$BUILD_DIR" + +# 运行 CMake 配置和编译 +cmake .. +make + +# 查找生成的 .a 文件 +A_FILE=$(find . -name "*.a" -type f) + +# 检查是否找到了 .a 文件 +if [ -z "$A_FILE" ]; then + echo "未找到 .a 文件,编译可能失败。" + exit 1 +fi + +# 复制头文件到 /usr/include +HEADER_FILES=$(find .. -name "*.h" -type f) +for header in $HEADER_FILES; do + cp "$header" /usr/include/ +done + +# 复制 .a 文件到 /usr/local/lib +cp "$A_FILE" /usr/local/lib/ + +echo "构建完成,头文件和 .a 文件已复制到目标目录。" + diff --git a/atb_operator/AtbAddSample/AddOperationTest/CMakeLists.txt b/atb_operator/AtbAddSample/AddOperationTest/CMakeLists.txt new file mode 100644 index 000000000..b9c2cec46 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOperationTest/CMakeLists.txt @@ -0,0 +1,40 @@ +cmake_minimum_required(VERSION 3.16) +project("test_model") +option(USE_CXX11_ABI "USE_CXX11_ABI" OFF) +set(CMAKE_BUILD_TYPE Debug) +if(USE_CXX11_ABI) + set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=1") +else() + set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") +endif() + +set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -std=c++17") + +set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") + +include_directories( + $ENV{ATB_HOME_PATH}/include + $ENV{ASCEND_HOME_PATH}/include + ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 + ${CMAKE_CURRENT_SOURCE_DIR} + ./src/ + ) + +message(".h path = ${CUST_PKG_PATH}/include") +link_directories( + $ENV{ATB_HOME_PATH}/lib + $ENV{ASCEND_HOME_PATH}/lib64 + ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 + ) + + +link_libraries(atb_plugin_operations) +add_executable(test_model ./src/main.cpp) + +# 列出所有的头文件目录 +target_link_libraries(test_model PRIVATE atb ascendcl cust_opapi opapi nnopbase pthread) # 添加自定义算子so及适配so + + + + + diff --git a/atb_operator/AtbAddSample/AddOperationTest/script/gen_data.py b/atb_operator/AtbAddSample/AddOperationTest/script/gen_data.py new file mode 100644 index 000000000..83ecafbce --- /dev/null +++ b/atb_operator/AtbAddSample/AddOperationTest/script/gen_data.py @@ -0,0 +1,20 @@ +import torch +import os +def gen_golden_data_simple(): + dtype = "float" + input1 = torch.randn(133, 4095, dtype=torch.float) + input2 = torch.randn(133, 4095, dtype=torch.float) + + + golden = input1 + input2 + os.system("mkdir -p input") + os.system("mkdir -p output") + input1.numpy().tofile('./script/input/input0.bin') + input2.numpy().tofile('./script/input/input1.bin') + golden.numpy().tofile("./script/output/golden0.bin") + + with open("./output/meta", "w") as fp: + print(dtype, file=fp) + +if __name__ == "__main__": + gen_golden_data_simple() diff --git a/atb_operator/AtbAddSample/AddOperationTest/script/run.sh b/atb_operator/AtbAddSample/AddOperationTest/script/run.sh new file mode 100644 index 000000000..450b717f2 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOperationTest/script/run.sh @@ -0,0 +1,52 @@ +#!/bin/bash + + +# 步骤1: 运行gen_data.py生成输入bin文件和golden标杆输出数据 +echo "正在生成输入数据和golden标杆数据..." +mkdir -p script/input +mkdir -p script/output +python3 script/gen_data.py +if [ $? -ne 0 ]; then + echo "生成数据失败,脚本终止。" + exit 1 +fi + +# 步骤2: 创建构建目录并进入 +mkdir -p build +cd build +if [ $? -ne 0 ]; then + echo "无法进入构建目录,脚本终止。" + exit 1 +fi + +# 步骤3: 使用CMake配置项目 +echo "正在配置CMake项目..." +cmake .. +if [ $? -ne 0 ]; then + echo "CMake配置失败,脚本终止。" + exit 1 +fi + +# 步骤4: 编译代码 +echo "正在编译代码..." +make +if [ $? -ne 0 ]; then + echo "编译失败,脚本终止。" + exit 1 +fi + +mv test_model ../ +cd .. + +# 步骤5: 运行可执行文件生成实际输出文件 +echo "正在运行可执行文件生成实际输出..." +./test_model +if [ $? -ne 0 ]; then + echo "运行可执行文件失败,脚本终止。" + exit 1 +fi + +# 步骤6: 调用verify_result.py进行golden标杆数据和实际输出数据的比对 +echo "正在验证结果..." +python3 script/verify_result.py script/output/output_0.bin script/output/golden0.bin + diff --git a/atb_operator/AtbAddSample/AddOperationTest/script/verify_result.py b/atb_operator/AtbAddSample/AddOperationTest/script/verify_result.py new file mode 100644 index 000000000..629845e13 --- /dev/null +++ b/atb_operator/AtbAddSample/AddOperationTest/script/verify_result.py @@ -0,0 +1,28 @@ +import os +import sys +import numpy + +loss = 1e-3 # 容忍偏差,一般fp16要求绝对误差和相对误差均不超过千分之一 +minimum = 10e-10 + +def verify_result(real_result, golden): + with open("output/meta", "r") as fp: + dtype_str = fp.readline().strip() + dtype = eval(dtype_str) + real_result = numpy.fromfile(real_result, dtype=dtype) # 从bin文件读取实际运算结果 + golden = numpy.fromfile(golden, dtype=dtype) # 从bin文件读取预期运算结果 + print("=" * 50, real_result[:5], golden[:5], "=" * 50, sep='\n', end='\n', file=sys.stderr) + result = numpy.abs(real_result - golden) # 计算运算结果和预期结果偏差 + deno = numpy.maximum(numpy.abs(real_result), numpy.abs(golden)) # 获取最大值并组成新数组 + result_atol = numpy.less_equal(result, loss) # 计算绝对误差 + result_rtol = numpy.less_equal(result / numpy.add(deno, minimum), loss) # 计算相对误差 + if not result_rtol.all() and not result_atol.all(): + if numpy.sum(result_rtol == False) > real_result.size * loss and numpy.sum(result_atol == False) > real_result.size * loss: # 误差超出预期时返回打印错误,返回对比失败 + print("[ERROR] result error") + return False + print("test Operation success!") + return True + + +if __name__ == '__main__': + verify_result(sys.argv[1],sys.argv[2]) diff --git a/atb_operator/AtbAddSample/AddOperationTest/src/main.cpp b/atb_operator/AtbAddSample/AddOperationTest/src/main.cpp new file mode 100644 index 000000000..258c0cb3c --- /dev/null +++ b/atb_operator/AtbAddSample/AddOperationTest/src/main.cpp @@ -0,0 +1,217 @@ +/* + * Copyright (c) Huawei Technologies Co., Ltd. 2023. All rights reserved. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +#include "main.h" + +bool SetInputData(std::vector &inputData){ + char *xPath = "./script/input/input0.bin"; + char *yPath = "./script/input/input1.bin"; + InputData inputX; + InputData inputY; + inputX.data = ReadBinFile(xPath,inputX.size); + inputY.data = ReadBinFile(yPath,inputY.size); + inputData.push_back(inputX); + inputData.push_back(inputY); + return true; +} + +bool SetOperationInputDesc(atb::SVector &intensorDescs){ + atb::TensorDesc xDesc; + xDesc.dtype = ACL_FLOAT; + xDesc.format = ACL_FORMAT_ND; + xDesc.shape.dimNum = 2; + xDesc.shape.dims[0] = 133; + xDesc.shape.dims[1] = 4095; + + atb::TensorDesc yDesc; + yDesc.dtype = ACL_FLOAT; + yDesc.format = ACL_FORMAT_ND; + yDesc.shape.dimNum = 2; + yDesc.shape.dims[0] = 133; + yDesc.shape.dims[1] = 4095; + + intensorDescs.at(0) = xDesc; + intensorDescs.at(1) = yDesc; +} + + + +static void SetCurrentDevice() +{ + const int deviceId = 0; + std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; + int ret = aclrtSetDevice(deviceId); + if (ret != 0) { + std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; + return; + } + std::cout << "[INFO]: aclrtSetDevice success" << std::endl; +} + + +static void FreeTensor(atb::Tensor &tensor) +{ + if (tensor.deviceData) { + int ret = aclrtFree(tensor.deviceData); + if (ret != 0) { + std::cout << "[ERROR]: aclrtFree fail" << std::endl; + } + tensor.deviceData = nullptr; + tensor.dataSize = 0; + } + if (tensor.hostData) { + int ret = aclrtFreeHost(tensor.hostData); + if (ret != 0) { + std::cout << "[ERROR]: aclrtFreeHost fail, ret = " << ret << std::endl; + } + tensor.hostData = nullptr; + tensor.dataSize = 0; + } +} + +static void FreeTensors(atb::SVector &inTensors, atb::SVector &outTensors) +{ + for (size_t i = 0; i < inTensors.size(); ++i) { + FreeTensor(inTensors.at(i)); + } + for (size_t i = 0; i < outTensors.size(); ++i) { + FreeTensor(outTensors.at(i)); + } +} +bool SaveMemoryToBinFile(void* memoryAddress, size_t memorySize, size_t i) { + // 创建 output 目录(如果不存在) + std::filesystem::create_directories("output"); + + // 生成文件名 + std::string filename = "script/output/output_" + std::to_string(i) + ".bin"; + + // 打开文件以二进制写入模式 + std::ofstream file(filename, std::ios::binary); + if (!file) { + std::cerr << "无法打开文件: " << filename << std::endl; + return false; + } + + // 写入数据 + file.write(static_cast(memoryAddress), memorySize); + if (!file) { + std::cerr << "写入文件时出错: " << filename << std::endl; + file.close(); + return false; + } + + // 关闭文件 + file.close(); + std::cout << "数据已成功保存到: " << filename << std::endl; + return true; +} + +int main(int argc, const char *argv[]) +{ + const int deviceId = 0; + std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; + int ret = aclrtSetDevice(deviceId); + if (ret != 0) { + std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; + return 1; + } + std::cout << "[INFO]: aclrtSetDevice success" << std::endl; + atb::Context *context = nullptr; + ret = atb::CreateContext(&context); + void *stream = nullptr; + ret = aclrtCreateStream(&stream); + if (ret != 0) { + std::cout << "[ERROR]: AsdRtStreamCreate fail, ret:" << ret << std::endl; + return 1; + } + context->SetExecuteStream(stream); + + std::vector input; + SetInputData(input); + + AddAttrParam addAttrParam; + AddOperation *op = new AddOperation("Add",addAttrParam); + std::cout << "[INFO]: complete CreateOp!" << std::endl; + + if(input.size() != op->GetInputNum()) std::cout << "[ERROR]: Operation actual input num is not equal to GetInputNum()"; + + atb::SVector intensorDescs; + atb::SVector outtensorDescs; + intensorDescs.resize(op->GetInputNum()); + outtensorDescs.resize(op->GetOutputNum()); + SetOperationInputDesc(intensorDescs); + atb::Status st = op->InferShape(intensorDescs,outtensorDescs); + if (st != 0) { + std::cout << "[ERROR]: Operation InferShape fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation InferShape success" << std::endl; + atb::VariantPack variantPack; + variantPack.inTensors.resize(op->GetInputNum()); + variantPack.outTensors.resize(op->GetOutputNum()); + for(size_t i=0;iGetInputNum();i++){ + variantPack.inTensors.at(i).desc = intensorDescs.at(i); + variantPack.inTensors.at(i).hostData = input[i].data; + variantPack.inTensors.at(i).dataSize = input[i].size; + CheckAcl(aclrtMalloc(&variantPack.inTensors.at(i).deviceData, input[i].size, ACL_MEM_MALLOC_HUGE_FIRST)); + CheckAcl(aclrtMemcpy(variantPack.inTensors.at(i).deviceData, input[i].size, input[i].data, input[i].size, ACL_MEMCPY_HOST_TO_DEVICE)); + } + std::cout << "[INFO]: Operation Input prepare sucess" << std::endl; + for(size_t i=0;iGetOutputNum();i++){ + int64_t *dims = new int64_t[outtensorDescs.at(i).shape.dimNum]; + for(size_t j=0;jSetup(variantPack, workspaceSize, context); + if (st != 0) { + std::cout << "[ERROR]: Operation setup fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation setup success" << std::endl; + void *workspace = nullptr; + if (workspaceSize > 0) { + aclrtMalloc(&workspace, workspaceSize, ACL_MEM_MALLOC_HUGE_FIRST); + } + + std::cout << "[INFO]: Operation execute start" << std::endl; + st = op->Execute(variantPack, (uint8_t*)workspace, workspaceSize, context); + if (st != 0) { + std::cout << "[ERROR]: Operation execute fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation execute success" << std::endl; + for(size_t i = 0; i < op->GetOutputNum(); i++){ + CheckAcl(aclrtMemcpy(variantPack.outTensors.at(i).hostData, variantPack.outTensors.at(i).dataSize, variantPack.outTensors.at(0).deviceData, + variantPack.outTensors.at(i).dataSize, ACL_MEMCPY_DEVICE_TO_HOST)); + SaveMemoryToBinFile(variantPack.outTensors.at(i).hostData,variantPack.outTensors.at(i).dataSize,i); + } + + FreeTensors(variantPack.inTensors, variantPack.outTensors); + st = atb::DestroyContext(context); + CheckAcl(aclrtDestroyStream(stream)); + CheckAcl(aclrtResetDevice(0)); + CheckAcl(aclFinalize()); + return atb::ErrorType::NO_ERROR; +} diff --git a/atb_operator/AtbAddSample/AddOperationTest/src/main.h b/atb_operator/AtbAddSample/AddOperationTest/src/main.h new file mode 100644 index 000000000..44a94e9ec --- /dev/null +++ b/atb_operator/AtbAddSample/AddOperationTest/src/main.h @@ -0,0 +1,55 @@ +#include +#include +#include +#include +#include +#include +#include +#include "securec.h" +#include "atb/atb_infer.h" +#include "aclnn_add_operation.h" +#include + +struct InputData{ + void* data; + uint64_t size; +}; +aclError CheckAcl(aclError ret) +{ + if (ret != ACL_ERROR_NONE) { + std::cerr << __FILE__ << ":" << __LINE__ << " aclError:" << ret << std::endl; + } + return ret; +} +void* ReadBinFile(const char* filename, size_t& size) { + std::ifstream file(filename, std::ios::binary | std::ios::ate); + if (!file) { + std::cerr << "无法打开文件: " << filename << std::endl; + return nullptr; + } + + // 获取文件大小 + size = file.tellg(); + file.seekg(0, std::ios::beg); + + // 分配内存 + void* buffer; + int ret = aclrtMallocHost(&buffer,size); + if (!buffer) { + std::cerr << "内存分配失败" << std::endl; + file.close(); + return nullptr; + } + + // 读取文件内容到内存 + file.read(static_cast(buffer), size); + if (!file) { + std::cerr << "读取文件失败" << std::endl; + delete[] static_cast(buffer); + file.close(); + return nullptr; + } + + file.close(); + return buffer; +} \ No newline at end of file diff --git a/atb_operator/AtbAddSample/readme.md b/atb_operator/AtbAddSample/readme.md index e69de29bb..d205e5429 100644 --- a/atb_operator/AtbAddSample/readme.md +++ b/atb_operator/AtbAddSample/readme.md @@ -0,0 +1,81 @@ +### 概述 + +本样例基于AscendC自定义Add算子,开发了ATB插件并进行了插件调用测试. + +### 项目结构介绍 + +├── AddAscendC //Add AscendC自定义算子工程 + +├── AddOperationATBPlugin //AddOperation ATB插件代码 + +├── AddOperationTest //AddOperation 测试代码 + +### 样例运行 + +#### Add AscendC自定义算子部署 + +- 运行生成算子run包 + + ```shell + cd AddAscendC + bash build.sh + ``` + +- 算子run包部署 + + ```shell + cd AddAscendC/build_out + ./custom_opp_hce_aarch64.run + ``` + +#### AddOperation ATB插件部署 + +- 运行编译脚本完成部署(脚本会生成静态库.a文件,同时将头文件拷贝到/usr/include,.a文件拷贝到/usr/local/lib下) + + ``` + cd AddOperationATBPlugin + bash build.sh + ``` + +#### AddOperation测试 + +- 运行脚本完成算子测试 + + ```shell + cd AddOperationTest + bash run.sh + ``` + +### AddOperation算子介绍 + +#### 功能 + +实现两个输入张量相加 + +#### 定义 + +``` +struct AddParam { +}; +``` + +#### 参数列表 + +该算子参数为空 + +#### 输入 + +| **参数** | **维度** | **数据类型** | **格式** | 描述 | +| -------- | -------------------------- | --------------------- | -------- | ---------- | +| x | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | +| y | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | + +#### 输出 + +| **参数** | **维度** | **数据类型** | **格式** | 描述 | +| -------- | -------------------------- | --------------------- | -------- | ---------------------------------------- | +| output | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输出tensor。数据类型和shape与x保持一致。 | + +#### 规格约束 + +暂无 \ No newline at end of file -- Gitee From 4e864944fe13d2f8e9daea56212b178e8ec051c3 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Mon, 31 Mar 2025 06:22:40 +0000 Subject: [PATCH 04/26] =?UTF-8?q?=E5=88=A0=E9=99=A4=E6=96=87=E4=BB=B6=20at?= =?UTF-8?q?b=5Foperator/AtbAddSample/AddOp?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .../AtbAddSample/AddOp/CMakeLists.txt | 69 -- .../AtbAddSample/AddOp/CMakePresets.json | 63 -- atb_operator/AtbAddSample/AddOp/README.md | 199 ----- atb_operator/AtbAddSample/AddOp/build.sh | 37 - .../AtbAddSample/AddOp/cmake/config.cmake | 25 - .../AtbAddSample/AddOp/cmake/func.cmake | 228 ----- .../AtbAddSample/AddOp/cmake/intf.cmake | 26 - .../AtbAddSample/AddOp/cmake/makeself.cmake | 17 - .../AtbAddSample/AddOp/cmake/util/__init__.py | 8 - .../cmake/util/ascendc_bin_param_build.py | 129 --- .../AddOp/cmake/util/ascendc_impl_build.py | 446 ---------- .../AddOp/cmake/util/ascendc_ops_config.py | 114 --- .../AddOp/cmake/util/ascendc_replay_build.py | 65 -- .../AddOp/cmake/util/batch_replay_impl.temp | 117 --- .../AddOp/cmake/util/code_channel_infer.py | 115 --- .../AddOp/cmake/util/const_var.py | 33 - .../cmake/util/gen_impl_and_mrege_json.sh | 57 -- .../AddOp/cmake/util/gen_ops_filter.sh | 62 -- .../AddOp/cmake/util/gen_version_info.sh | 6 - .../AddOp/cmake/util/insert_op_info.py | 36 - .../cmake/util/insert_simplified_keys.py | 248 ------ .../AddOp/cmake/util/kernel_entry.py | 115 --- .../AddOp/cmake/util/kernel_impl.temp | 10 - .../AddOp/cmake/util/makeself/COPYING | 339 -------- .../AddOp/cmake/util/makeself/README.md | 246 ------ .../AddOp/cmake/util/makeself/VERSION | 1 - .../AddOp/cmake/util/makeself/make-release.sh | 9 - .../cmake/util/makeself/makeself-header.sh | 660 -------------- .../AddOp/cmake/util/makeself/makeself.1 | 110 --- .../AddOp/cmake/util/makeself/makeself.lsm | 16 - .../AddOp/cmake/util/makeself/makeself.sh | 822 ------------------ .../AddOp/cmake/util/makeself/run-tests.sh | 8 - .../AddOp/cmake/util/merge_aicpu_info_json.sh | 31 - .../AddOp/cmake/util/opdesc_parser.py | 260 ------ .../AddOp/cmake/util/parse_ini_to_json.py | 338 ------- .../AddOp/cmake/util/preset_parse.py | 23 - .../AddOp/cmake/util/replay_codegen.py | 105 --- .../AddOp/cmake/util/replay_impl.temp | 120 --- .../AddOp/cmake/util/tiling_data_def_build.py | 84 -- .../AddOp/framework/CMakeLists.txt | 11 - .../AddOp/framework/tf_plugin/CMakeLists.txt | 14 - .../tf_plugin/tensorflow_add_custom_plugin.cc | 23 - .../AtbAddSample/AddOp/op_host/CMakeLists.txt | 82 -- .../AtbAddSample/AddOp/op_host/add_custom.cpp | 117 --- .../AddOp/op_host/add_custom_tiling.h | 22 - .../AddOp/op_kernel/CMakeLists.txt | 61 -- .../AddOp/op_kernel/add_custom.cpp | 134 --- .../AtbAddSample/AddOp/scripts/install.sh | 318 ------- .../AtbAddSample/AddOp/scripts/upgrade.sh | 151 ---- 49 files changed, 6330 deletions(-) delete mode 100644 atb_operator/AtbAddSample/AddOp/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AddOp/CMakePresets.json delete mode 100644 atb_operator/AtbAddSample/AddOp/README.md delete mode 100644 atb_operator/AtbAddSample/AddOp/build.sh delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/config.cmake delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/func.cmake delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/intf.cmake delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/makeself.cmake delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/__init__.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_bin_param_build.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_impl_build.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_ops_config.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_replay_build.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/batch_replay_impl.temp delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/code_channel_infer.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/const_var.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/gen_impl_and_mrege_json.sh delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/gen_ops_filter.sh delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/gen_version_info.sh delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/insert_op_info.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/insert_simplified_keys.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/kernel_entry.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/kernel_impl.temp delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/COPYING delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/README.md delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/VERSION delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/make-release.sh delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself-header.sh delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.1 delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.lsm delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.sh delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/makeself/run-tests.sh delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/merge_aicpu_info_json.sh delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/opdesc_parser.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/parse_ini_to_json.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/preset_parse.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/replay_codegen.py delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/replay_impl.temp delete mode 100644 atb_operator/AtbAddSample/AddOp/cmake/util/tiling_data_def_build.py delete mode 100644 atb_operator/AtbAddSample/AddOp/framework/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AddOp/framework/tf_plugin/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AddOp/framework/tf_plugin/tensorflow_add_custom_plugin.cc delete mode 100644 atb_operator/AtbAddSample/AddOp/op_host/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AddOp/op_host/add_custom.cpp delete mode 100644 atb_operator/AtbAddSample/AddOp/op_host/add_custom_tiling.h delete mode 100644 atb_operator/AtbAddSample/AddOp/op_kernel/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AddOp/op_kernel/add_custom.cpp delete mode 100644 atb_operator/AtbAddSample/AddOp/scripts/install.sh delete mode 100644 atb_operator/AtbAddSample/AddOp/scripts/upgrade.sh diff --git a/atb_operator/AtbAddSample/AddOp/CMakeLists.txt b/atb_operator/AtbAddSample/AddOp/CMakeLists.txt deleted file mode 100644 index 584132d80..000000000 --- a/atb_operator/AtbAddSample/AddOp/CMakeLists.txt +++ /dev/null @@ -1,69 +0,0 @@ -cmake_minimum_required(VERSION 3.16.0) -project(opp) -if(ENABLE_CROSS_COMPILE) - if(${CMAKE_SYSTEM_PROCESSOR} STREQUAL x86_64) - set(CROSS_COMPILE_PLATFORM aarch64) - else() - set(CROSS_COMPILE_PLATFORM x86_64) - endif() - set(PLATFORM ${CMAKE_SYSTEM_PROCESSOR}) - set(CMAKE_COMPILE_COMPILER_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/linux/${CROSS_COMPILE_PLATFORM}/) - set(CMAKE_COMPILE_RUNTIME_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/${CROSS_COMPILE_PLATFORM}/) - set(CMAKE_SYSTEM_PROCESSOR ${CROSS_COMPILE_PLATFORM}) - set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) - set(CMAKE_CXX_COMPILER ${CMAKE_CROSS_PLATFORM_COMPILER}) -else() - set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) -endif() - -include(cmake/config.cmake) -include(cmake/func.cmake) -include(cmake/intf.cmake) - -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/framework) - add_subdirectory(framework) -endif() -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_host) - add_subdirectory(op_host) -endif() -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_kernel) - add_subdirectory(op_kernel) -endif() -if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) - add_subdirectory(testcases) -endif() - -# modify vendor_name in install.sh and upgrade.sh -add_custom_command(OUTPUT ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh - COMMAND mkdir -p ${CMAKE_BINARY_DIR}/scripts - COMMAND cp -r ${CMAKE_SOURCE_DIR}/scripts/* ${CMAKE_BINARY_DIR}/scripts/ - COMMAND sed -i "s/vendor_name=customize/vendor_name=${vendor_name}/g" ${CMAKE_BINARY_DIR}/scripts/* -) -add_custom_target(modify_vendor ALL DEPENDS ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh) -install(DIRECTORY ${CMAKE_BINARY_DIR}/scripts/ DESTINATION . FILE_PERMISSIONS OWNER_EXECUTE OWNER_READ GROUP_READ) - -install(FILES ${CMAKE_SOURCE_DIR}/custom.proto DESTINATION packages OPTIONAL) - -get_system_info(SYSTEM_INFO) - -# gen version.info -add_custom_target(gen_version_info ALL - COMMAND bash ${CMAKE_CURRENT_SOURCE_DIR}/cmake/util/gen_version_info.sh ${ASCEND_CANN_PACKAGE_PATH} ${CMAKE_CURRENT_BINARY_DIR} -) - -install(FILES ${CMAKE_CURRENT_BINARY_DIR}/version.info - DESTINATION packages/vendors/${vendor_name}/) - -# CPack config -set(CPACK_PACKAGE_NAME ${CMAKE_PROJECT_NAME}) -set(CPACK_PACKAGE_VERSION ${CMAKE_PROJECT_VERSION}) -set(CPACK_PACKAGE_DESCRIPTION "CPack opp project") -set(CPACK_PACKAGE_DESCRIPTION_SUMMARY "CPack opp project") -set(CPACK_PACKAGE_DIRECTORY ${CMAKE_INSTALL_PREFIX}) -set(CPACK_PACKAGE_FILE_NAME "custom_opp_${SYSTEM_INFO}.run") -set(CPACK_GENERATOR External) -set(CPACK_CMAKE_GENERATOR "Unix Makefiles") -set(CPACK_EXTERNAL_ENABLE_STAGING TRUE) -set(CPACK_EXTERNAL_PACKAGE_SCRIPT ${CMAKE_SOURCE_DIR}/cmake/makeself.cmake) -set(CPACK_EXTERNAL_BUILT_PACKAGES ${CPACK_PACKAGE_DIRECTORY}/_CPack_Packages/Linux/External/${CPACK_PACKAGE_FILE_NAME}/${CPACK_PACKAGE_FILE_NAME}) -include(CPack) diff --git a/atb_operator/AtbAddSample/AddOp/CMakePresets.json b/atb_operator/AtbAddSample/AddOp/CMakePresets.json deleted file mode 100644 index ddec9431b..000000000 --- a/atb_operator/AtbAddSample/AddOp/CMakePresets.json +++ /dev/null @@ -1,63 +0,0 @@ -{ - "version": 1, - "cmakeMinimumRequired": { - "major": 3, - "minor": 19, - "patch": 0 - }, - "configurePresets": [ - { - "name": "default", - "displayName": "Default Config", - "description": "Default build using Unix Makefiles generator", - "generator": "Unix Makefiles", - "binaryDir": "${sourceDir}/build_out", - "cacheVariables": { - "CMAKE_BUILD_TYPE": { - "type": "STRING", - "value": "Release" - }, - "ENABLE_SOURCE_PACKAGE": { - "type": "BOOL", - "value": "True" - }, - "ENABLE_BINARY_PACKAGE": { - "type": "BOOL", - "value": "True" - }, - "ASCEND_COMPUTE_UNIT": { - "type": "STRING", - "value": "ascend310b;ascend910b" - }, - "ENABLE_TEST": { - "type": "BOOL", - "value": "True" - }, - "vendor_name": { - "type": "STRING", - "value": "customize" - }, - "ASCEND_CANN_PACKAGE_PATH": { - "type": "PATH", - "value": "/usr/local/Ascend/ascend-toolkit/latest" - }, - "ASCEND_PYTHON_EXECUTABLE": { - "type": "STRING", - "value": "python3" - }, - "CMAKE_INSTALL_PREFIX": { - "type": "PATH", - "value": "${sourceDir}/build_out" - }, - "ENABLE_CROSS_COMPILE": { - "type": "BOOL", - "value": "False" - }, - "CMAKE_CROSS_PLATFORM_COMPILER": { - "type": "PATH", - "value": "/usr/bin/aarch64-linux-gnu-g++" - } - } - } - ] -} diff --git a/atb_operator/AtbAddSample/AddOp/README.md b/atb_operator/AtbAddSample/AddOp/README.md deleted file mode 100644 index f694535dd..000000000 --- a/atb_operator/AtbAddSample/AddOp/README.md +++ /dev/null @@ -1,199 +0,0 @@ -# 概述 -本样例基于AddOp算子工程,介绍了单算子工程及aclnn接口说明。 - -## 算子工程介绍 -### 算子描述 -Add算子实现了两个数据相加,返回相加结果的功能,该算子实现了非32B对齐shape下的算子功能,完成了算子的泛化实现。 -对应的数学表达式为: -``` -z = x + y -``` -### 算子规格描述 - - - - - - - - - - - -
算子类型(OpType)Add
算子输入nameshapedata typeformat
x-float32,float16,int32,int8ND
y-float32,float16,int32,int8ND
算子输出z-float32,float16,int32,int8ND
核函数名add_custom
- - -### 算子工程文件介绍 -其中,算子工程目录AddOp包含算子实现的模板文件、编译脚本等,如下所示: -``` -├── AddOp //Add自定义算子工程 -│ ├── cmake -│ ├── framework // 算子插件实现文件目录,单算子模型文件的生成不依赖算子适配插件,无需关注 -│ ├── op_host // host侧实现文件 -│ ├── op_kernel // kernel侧实现文件 -│ ├── scripts // 自定义算子工程打包相关脚本所在目录 -│ ├── build.sh // 编译入口脚本 -│ ├── CMakeLists.txt // 算子工程的CMakeLists.txt -│ └── CMakePresets.json // 编译配置项 -``` - -### 编译运行样例算子 -针对自定义算子工程,编译运行包含如下步骤: -- 编译自定义算子工程生成算子安装包; -- 安装自定义算子到算子库中; -- 调用执行自定义算子; - -详细操作如下所示。 -#### 1. 获取源码包 -编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 -#### 2. 编译算子工程 - 编译自定义算子工程,构建生成自定义算子包。 - - - 执行如下命令,切换到算子工程AddOp目录。 - - ```bash - cd ${git_clone_path}/samples/atb_operator/AtbAddSample/AddOp - ``` - - - 修改CMakePresets.json中ASCEND_CANN_PACKAGE_PATH为CANN软件包安装后的实际路径。 - - - ```json - { - …… - "configurePresets": [ - { - …… - "ASCEND_CANN_PACKAGE_PATH": { - "type": "PATH", - "value": "/usr/local/Ascend/ascend-toolkit/latest" //请替换为CANN软件包安装后的实际路径。eg:/home/HwHiAiUser/Ascend/ascend-toolkit/latest - }, - …… - } - ] - } - ``` - - 在算子工程AddOp目录下执行如下命令,进行算子工程编译。 - - ```bash - ./build.sh - ``` -编译成功后,会在当前目录下创建build_out目录,并在build_out目录下生成自定义算子安装包custom_opp_\_\.run,例如“custom_opp_ubuntu_x86_64.run”。 - -备注:如果要使用dump调试功能,需要移除op_host内和CMakeLists.txt内的Atlas 200/500 A2 推理产品的配置。 - -#### 3. 部署算子包 - -执行如下命令,在自定义算子安装包所在路径下,安装自定义算子包。 - ```bash - cd build_out - ./custom_opp__.run - ``` -命令执行成功后,自定义算子包中的相关文件将部署至当前环境的OPP算子库的vendors/customize目录中。 - -#### 4. 配置环境变量 - - 请根据当前环境上CANN开发套件包的[安装方式](https://hiascend.com/document/redirect/CannCommunityInstSoftware),选择对应配置环境变量的命令。 - - 默认路径,root用户安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=/usr/local/Ascend/ascend-toolkit/latest - ``` - - 默认路径,非root用户安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=$HOME/Ascend/ascend-toolkit/latest - ``` - - 指定路径install_path,安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=${install_path}/ascend-toolkit/latest - ``` - -## aclnn接口说明 - -自定义算子编译部署后,会自动生成单算子API,可以直接在应用程序中调用。算子API的形式一般定义为“两段式接口”,形如: - ```cpp - aclnnStatus aclnnAddOpGetWorkspaceSize(const aclTensor *x, const aclTensor *y, const alcTensor *out, uint64_t workspaceSize, aclOpExecutor **executor); - aclnnStatus aclnnAddOp(void *workspace, int64_t workspaceSize, aclOpExecutor **executor, aclrtStream stream); - ``` -其中aclnnAddOpGetWorkspaceSize为第一段接口,主要用于计算本次API调用计算过程中需要多少的workspace内存。获取到本次API计算需要的workspace大小之后,按照workspaceSize大小申请Device侧内存,然后调用第二段接口aclnnAddOp执行计算。具体参考[AscendCL单算子调用](https://hiascend.com/document/redirect/CannCommunityAscendCInVorkSingleOp)>单算子API执行 章节。 - -### 功能描述 -* 算子功能:完成加法计算。 -* 计算公式: -``` -z = x + y -``` - -#### 参数说明 -##### aclnnAddOpGetWorkspaceSize: - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
接口参数描述
aclnnAddOpGetWorkspaceSizex输入x的Tensor,支持flaot/half/int8/int32类型,ND排布格式
y输入y的Tensor,支持flaot/half/int8/int32类型,ND排布格式
out输出z的Tensor,支持flaot/half/int8/int32类型,ND排布格式
workspaceSize返回需要在Device侧申请的workspace大小
executor返回op执行器,包含了算子计算流程
- - -| 返回值 | 描述 | -| ---- | ---- | -| ACL_SUCCESS | 成功 | -| ACL_ERROR | 失败 | - -##### aclnnAddOp - - - - - - - - - - - - - - - - - - - - - - - -
接口参数描述
aclnnAddOpworkspace输入Device侧申请的workspace内存
workspaceSizeworkspace大小
executor输入op执行器,包含了算子计算流程
stream输入执行流
- -| 返回值 | 描述 | -| ---- | ---- | -| ACL_SUCCESS | 成功 | -| ACL_ERROR | 失败 | - - -## 更新说明 -| 时间 | 更新事项 | -|----|------| -| 2025/03/27 | 新增README | \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AddOp/build.sh b/atb_operator/AtbAddSample/AddOp/build.sh deleted file mode 100644 index 4be96d7d8..000000000 --- a/atb_operator/AtbAddSample/AddOp/build.sh +++ /dev/null @@ -1,37 +0,0 @@ -#!/bin/bash -script_path=$(realpath $(dirname $0)) - - -mkdir -p build_out -rm -rf build_out/* -cd build_out - -cmake_version=$(cmake --version | grep "cmake version" | awk '{print $3}') -if [ "$cmake_version" \< "3.19.0" ] ; then - opts=$(python3 $script_path/cmake/util/preset_parse.py $script_path/CMakePresets.json) - echo $opts - cmake .. $opts -else - cmake .. --preset=default -fi -target=package -if [ "$1"x != ""x ]; then target=$1; fi - -cmake --build . --target $target -j16 -if [ $? -ne 0 ]; then exit 1; fi - -if [ $target = "package" ]; then - if test -d ./op_kernel/binary ; then - ./cust*.run - if [ $? -ne 0 ]; then exit 1; fi - cmake --build . --target binary -j16 - if [ $? -ne 0 ]; then exit 1; fi - cmake --build . --target $target -j16 - fi -fi - -# for debug -# cd build_out -# make -# cpack -# verbose append -v diff --git a/atb_operator/AtbAddSample/AddOp/cmake/config.cmake b/atb_operator/AtbAddSample/AddOp/cmake/config.cmake deleted file mode 100644 index 886119daa..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/config.cmake +++ /dev/null @@ -1,25 +0,0 @@ - -set(CMAKE_CXX_FLAGS_DEBUG "") -set(CMAKE_CXX_FLAGS_RELEASE "") - -if (NOT DEFINED vendor_name) - set(vendor_name customize CACHE STRING "") -endif() -if (NOT DEFINED ASCEND_CANN_PACKAGE_PATH) - set(ASCEND_CANN_PACKAGE_PATH /usr/local/Ascend/latest CACHE PATH "") -endif() -if (NOT DEFINED ASCEND_PYTHON_EXECUTABLE) - set(ASCEND_PYTHON_EXECUTABLE python3 CACHE STRING "") -endif() -if (NOT DEFINED ASCEND_COMPUTE_UNIT) - message(FATAL_ERROR "ASCEND_COMPUTE_UNIT not set in CMakePreset.json ! -") -endif() -set(ASCEND_TENSOR_COMPILER_PATH ${ASCEND_CANN_PACKAGE_PATH}/compiler) -set(ASCEND_CCEC_COMPILER_PATH ${ASCEND_TENSOR_COMPILER_PATH}/ccec_compiler/bin) -set(ASCEND_AUTOGEN_PATH ${CMAKE_BINARY_DIR}/autogen) -set(ASCEND_FRAMEWORK_TYPE tensorflow) -file(MAKE_DIRECTORY ${ASCEND_AUTOGEN_PATH}) -set(CUSTOM_COMPILE_OPTIONS "custom_compile_options.ini") -execute_process(COMMAND rm -rf ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} - COMMAND touch ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS}) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/func.cmake b/atb_operator/AtbAddSample/AddOp/cmake/func.cmake deleted file mode 100644 index ad187e7d6..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/func.cmake +++ /dev/null @@ -1,228 +0,0 @@ - -function(get_system_info SYSTEM_INFO) - if (UNIX) - execute_process(COMMAND grep -i ^id= /etc/os-release OUTPUT_VARIABLE TEMP) - string(REGEX REPLACE "\n|id=|ID=|\"" "" SYSTEM_NAME ${TEMP}) - set(${SYSTEM_INFO} ${SYSTEM_NAME}_${CMAKE_SYSTEM_PROCESSOR} PARENT_SCOPE) - elseif (WIN32) - message(STATUS "System is Windows. Only for pre-build.") - else () - message(FATAL_ERROR "${CMAKE_SYSTEM_NAME} not support.") - endif () -endfunction() - -function(opbuild) - message(STATUS "Opbuild generating sources") - cmake_parse_arguments(OPBUILD "" "OUT_DIR;PROJECT_NAME;ACCESS_PREFIX" "OPS_SRC" ${ARGN}) - execute_process(COMMAND ${CMAKE_COMPILE} -g -fPIC -shared -std=c++11 ${OPBUILD_OPS_SRC} -D_GLIBCXX_USE_CXX11_ABI=0 - -I ${ASCEND_CANN_PACKAGE_PATH}/include -L ${ASCEND_CANN_PACKAGE_PATH}/lib64 -lexe_graph -lregister -ltiling_api - -o ${OPBUILD_OUT_DIR}/libascend_all_ops.so - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("build ops lib info: ${EXEC_INFO}") - message("build ops lib error: ${EXEC_ERROR}") - message(FATAL_ERROR "opbuild run failed!") - endif() - set(proj_env "") - set(prefix_env "") - if (NOT "${OPBUILD_PROJECT_NAME}x" STREQUAL "x") - set(proj_env "OPS_PROJECT_NAME=${OPBUILD_PROJECT_NAME}") - endif() - if (NOT "${OPBUILD_ACCESS_PREFIX}x" STREQUAL "x") - set(prefix_env "OPS_DIRECT_ACCESS_PREFIX=${OPBUILD_ACCESS_PREFIX}") - endif() - execute_process(COMMAND ${proj_env} ${prefix_env} ${ASCEND_CANN_PACKAGE_PATH}/toolkit/tools/opbuild/op_build - ${OPBUILD_OUT_DIR}/libascend_all_ops.so ${OPBUILD_OUT_DIR} - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("opbuild ops info: ${EXEC_INFO}") - message("opbuild ops error: ${EXEC_ERROR}") - endif() - message(STATUS "Opbuild generating sources - done") -endfunction() - -function(add_ops_info_target) - cmake_parse_arguments(OPINFO "" "TARGET;OPS_INFO;OUTPUT;INSTALL_DIR" "" ${ARGN}) - get_filename_component(opinfo_file_path "${OPINFO_OUTPUT}" DIRECTORY) - add_custom_command(OUTPUT ${OPINFO_OUTPUT} - COMMAND mkdir -p ${opinfo_file_path} - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/parse_ini_to_json.py - ${OPINFO_OPS_INFO} ${OPINFO_OUTPUT} - ) - add_custom_target(${OPINFO_TARGET} ALL - DEPENDS ${OPINFO_OUTPUT} - ) - install(FILES ${OPINFO_OUTPUT} - DESTINATION ${OPINFO_INSTALL_DIR} - ) -endfunction() - -function(add_ops_compile_options OP_TYPE) - cmake_parse_arguments(OP_COMPILE "" "OP_TYPE" "COMPUTE_UNIT;OPTIONS" ${ARGN}) - file(APPEND ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} - "${OP_TYPE},${OP_COMPILE_COMPUTE_UNIT},${OP_COMPILE_OPTIONS}\n") -endfunction() - -function(add_ops_impl_target) - cmake_parse_arguments(OPIMPL "" "TARGET;OPS_INFO;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) - add_custom_command(OUTPUT ${OPIMPL_OUT_DIR}/.impl_timestamp - COMMAND mkdir -m 700 -p ${OPIMPL_OUT_DIR}/dynamic - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py - ${OPIMPL_OPS_INFO} - \"${OPIMPL_OPS_BATCH}\" \"${OPIMPL_OPS_ITERATE}\" - ${OPIMPL_IMPL_DIR} - ${OPIMPL_OUT_DIR}/dynamic - ${ASCEND_AUTOGEN_PATH} - - COMMAND rm -rf ${OPIMPL_OUT_DIR}/.impl_timestamp - COMMAND touch ${OPIMPL_OUT_DIR}/.impl_timestamp - DEPENDS ${OPIMPL_OPS_INFO} - ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py - ) - add_custom_target(${OPIMPL_TARGET} ALL - DEPENDS ${OPIMPL_OUT_DIR}/.impl_timestamp) - if (${ENABLE_SOURCE_PACKAGE}) - install(DIRECTORY ${OPIMPL_OUT_DIR}/dynamic - DESTINATION ${OPIMPL_INSTALL_DIR} - ) - endif() -endfunction() - -function(add_ops_replay_targets) - cmake_parse_arguments(OPREPLAY "" "OPS_INFO;COMPUTE_UNIT;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) - # ccec compile options - set(ccec_base_opts -c -O2 --cce-aicore-only -mllvm -cce-aicore-function-stack-size=16000 - -mllvm -cce-aicore-record-overflow=false -std=c++17) - set(ccec_extopts_ascend310p --cce-aicore-arch=dav-m200 -mllvm -cce-aicore-fp-ceiling=2) - set(ccec_extopts_ascend910 --cce-aicore-arch=dav-c100) - set(ccec_extopts_ascend910b --cce-aicore-arch=dav-c220-cube) - file(MAKE_DIRECTORY ${OPREPLAY_OUT_DIR}) - execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_replay_build.py - ${OPREPLAY_OPS_INFO} - "${OPREPLAY_OPS_BATCH}" "${OPREPLAY_OPS_ITERATE}" - ${OPREPLAY_IMPL_DIR} - ${OPREPLAY_OUT_DIR} - ${OPREPLAY_COMPUTE_UNIT} - ) - file(GLOB replay_kernel_entries ${OPREPLAY_OUT_DIR}/*.cce) - if (NOT "${replay_kernel_entries}x" STREQUAL "x") - foreach(replay_kernel_file ${replay_kernel_entries}) - get_filename_component(replay_kernel_file_name "${replay_kernel_file}" NAME) - string(REPLACE "_entry.cce" "" op_kerne_name ${replay_kernel_file_name}) - file(GLOB replay_lib_src ${OPREPLAY_OUT_DIR}/${op_kerne_name}*.cpp) - set(OP_TILING_DATA_H_PATH ${OPREPLAY_OUT_DIR}/${op_kerne_name}_tiling_data.h) - add_library(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} SHARED ${replay_lib_src}) - if(EXISTS ${OP_TILING_DATA_H_PATH}) - target_compile_options(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE - -include ${OP_TILING_DATA_H_PATH} - ) - endif() - target_compile_definitions(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE - ${op_kerne_name}=${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} - ) - target_compile_options(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE - -D__ASCENDC_REPLAY__ - ) - target_link_libraries(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE intf_pub - tikreplaylib::${OPREPLAY_COMPUTE_UNIT} - register - ) - add_custom_command(OUTPUT ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o - COMMAND ccec ${ccec_base_opts} ${ccec_extopts_${OPREPLAY_COMPUTE_UNIT}} ${replay_kernel_file} - -o ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o - DEPENDS ${replay_kernel_file} - ) - add_custom_target(replay_kernel_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} ALL - DEPENDS ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o - ) - install(TARGETS replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_replay - ) - install(FILES ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o - DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_replay - ) - endforeach() - endif() -endfunction() - -function(add_npu_support_target) - cmake_parse_arguments(NPUSUP "" "TARGET;OPS_INFO_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) - get_filename_component(npu_sup_file_path "${NPUSUP_OUT_DIR}" DIRECTORY) - add_custom_command(OUTPUT ${NPUSUP_OUT_DIR}/npu_supported_ops.json - COMMAND mkdir -p ${NPUSUP_OUT_DIR} - COMMAND ${CMAKE_SOURCE_DIR}/cmake/util/gen_ops_filter.sh - ${NPUSUP_OPS_INFO_DIR} - ${NPUSUP_OUT_DIR} - ) - add_custom_target(npu_supported_ops ALL - DEPENDS ${NPUSUP_OUT_DIR}/npu_supported_ops.json - ) - install(FILES ${NPUSUP_OUT_DIR}/npu_supported_ops.json - DESTINATION ${NPUSUP_INSTALL_DIR} - ) -endfunction() - -function(add_bin_compile_target) - cmake_parse_arguments(BINCMP "" "TARGET;OPS_INFO;COMPUTE_UNIT;IMPL_DIR;ADP_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/src) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/gen) - execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_bin_param_build.py - ${BINCMP_OPS_INFO} ${BINCMP_OUT_DIR}/gen ${BINCMP_COMPUTE_UNIT} - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("ops binary compile scripts gen info: ${EXEC_INFO}") - message("ops binary compile scripts gen error: ${EXEC_ERROR}") - message(FATAL_ERROR "ops binary compile scripts gen failed!") - endif() - if (NOT TARGET binary) - add_custom_target(binary) - endif() - add_custom_target(${BINCMP_TARGET} - COMMAND cp -r ${BINCMP_IMPL_DIR}/*.* ${BINCMP_OUT_DIR}/src - ) - add_custom_target(${BINCMP_TARGET}_gen_ops_config - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/insert_simplified_keys.py -p ${BINCMP_OUT_DIR}/bin - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_ops_config.py -p ${BINCMP_OUT_DIR}/bin - -s ${BINCMP_COMPUTE_UNIT} - ) - add_dependencies(binary ${BINCMP_TARGET}_gen_ops_config) - file(GLOB bin_scripts ${BINCMP_OUT_DIR}/gen/*.sh) - foreach(bin_script ${bin_scripts}) - get_filename_component(bin_file ${bin_script} NAME_WE) - string(REPLACE "-" ";" bin_sep ${bin_file}) - list(GET bin_sep 0 op_type) - list(GET bin_sep 1 op_file) - list(GET bin_sep 2 op_index) - if (NOT TARGET ${BINCMP_TARGET}_${op_file}_copy) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file}) - add_custom_target(${BINCMP_TARGET}_${op_file}_copy - COMMAND cp ${BINCMP_ADP_DIR}/${op_file}.py ${BINCMP_OUT_DIR}/src/${op_type}.py - ) - install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} - DESTINATION ${BINCMP_INSTALL_DIR}/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) - install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json - DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT}/ OPTIONAL - ) - endif() - add_custom_target(${BINCMP_TARGET}_${op_file}_${op_index} - COMMAND export HI_PYTHON=${ASCEND_PYTHON_EXECUTABLE} && bash ${bin_script} ${BINCMP_OUT_DIR}/src/${op_type}.py ${BINCMP_OUT_DIR}/bin/${op_file} - WORKING_DIRECTORY ${BINCMP_OUT_DIR} - ) - add_dependencies(${BINCMP_TARGET}_${op_file}_${op_index} ${BINCMP_TARGET} ${BINCMP_TARGET}_${op_file}_copy) - add_dependencies(${BINCMP_TARGET}_gen_ops_config ${BINCMP_TARGET}_${op_file}_${op_index}) - endforeach() - install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json - DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) -endfunction() diff --git a/atb_operator/AtbAddSample/AddOp/cmake/intf.cmake b/atb_operator/AtbAddSample/AddOp/cmake/intf.cmake deleted file mode 100644 index 2f362c396..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/intf.cmake +++ /dev/null @@ -1,26 +0,0 @@ - -add_library(intf_pub INTERFACE) -target_compile_options(intf_pub INTERFACE - -fPIC - -fvisibility=hidden - -fvisibility-inlines-hidden - $<$:-O2> - $<$:-O0 -g> - $<$:-std=c++11> - $<$,$>:-ftrapv -fstack-check> - $<$:-pthread -Wfloat-equal -Wshadow -Wformat=2 -Wno-deprecated -Wextra> - $,-fstack-protector-strong,-fstack-protector-all> -) -target_compile_definitions(intf_pub INTERFACE - _GLIBCXX_USE_CXX11_ABI=0 - $<$:_FORTIFY_SOURCE=2> -) -target_include_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/include) -target_link_options(intf_pub INTERFACE - $<$,EXECUTABLE>:-pie> - $<$:-s> - -Wl,-z,relro - -Wl,-z,now - -Wl,-z,noexecstack -) -target_link_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/lib64) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/makeself.cmake b/atb_operator/AtbAddSample/AddOp/cmake/makeself.cmake deleted file mode 100644 index 48c565bfb..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/makeself.cmake +++ /dev/null @@ -1,17 +0,0 @@ -execute_process(COMMAND chmod +x ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh) -execute_process(COMMAND ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh - --header ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself-header.sh - --help-header ./help.info - --gzip --complevel 4 --nomd5 --sha256 - ./ ${CPACK_PACKAGE_FILE_NAME} "version:1.0" ./install.sh - WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} - RESULT_VARIABLE EXEC_RESULT - ERROR_VARIABLE EXEC_ERROR -) -if (NOT "${EXEC_RESULT}x" STREQUAL "0x") - message(FATAL_ERROR "CPack Command error: ${EXEC_RESULT}\n${EXEC_ERROR}") -endif() -execute_process(COMMAND cp ${CPACK_EXTERNAL_BUILT_PACKAGES} ${CPACK_PACKAGE_DIRECTORY}/ - COMMAND echo "Copy ${CPACK_EXTERNAL_BUILT_PACKAGES} to ${CPACK_PACKAGE_DIRECTORY}/" - WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} -) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/__init__.py b/atb_operator/AtbAddSample/AddOp/cmake/util/__init__.py deleted file mode 100644 index c4ddc893a..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/__init__.py +++ /dev/null @@ -1,8 +0,0 @@ -#!/usr/bin/env python -# -*- coding: utf-8 -*- - -import sys -import os - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) -sys.path.append(PYF_PATH) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_bin_param_build.py b/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_bin_param_build.py deleted file mode 100644 index decf34544..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_bin_param_build.py +++ /dev/null @@ -1,129 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import json -import hashlib -import const_var -import opdesc_parser - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - - -class BinParamBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - super().__init__(op_type) - self.soc = '' - self.out_path = '' - - def set_soc_version(self: any, soc: str): - self.soc = soc - - def set_out_path(self: any, out_path: str): - self.out_path = out_path - - def gen_input_json(self: any): - key_map = {} - count = len(self.input_dtype[0].split(',')) - for i in range(0, count): - inputs = [] - outputs = [] - attrs = [] - op_node = {} - for idx in range(0, len(self.input_name)): - idtypes = self.input_dtype[idx].split(',') - ifmts = self.input_fmt[idx].split(',') - itype = self.input_type[idx] - para = {} - para['name'] = self.input_name[idx] - para['index'] = idx - para['dtype'] = idtypes[i] - para['format'] = ifmts[i] - para['paramType'] = itype - para['shape'] = [-2] - if itype == 'dynamic': - inputs.append([para]) - else: - inputs.append(para) - for idx in range(0, len(self.output_name)): - odtypes = self.output_dtype[idx].split(',') - ofmts = self.output_fmt[idx].split(',') - otype = self.output_type[idx] - para = {} - para['name'] = self.output_name[idx] - para['index'] = idx - para['dtype'] = odtypes[i] - para['format'] = ofmts[i] - para['paramType'] = otype - para['shape'] = [-2] - if otype == 'dynamic': - outputs.append([para]) - else: - outputs.append(para) - for attr in self.attr_list: - att = {} - att['name'] = attr - atype = self.attr_val.get(attr).get('type').lower() - atype = atype.replace('list', 'list_') - att['dtype'] = atype - att['value'] = const_var.ATTR_DEF_VAL.get(atype) - attrs.append(att) - op_node['bin_filename'] = '' - op_node['inputs'] = inputs - op_node['outputs'] = outputs - if len(attrs) > 0: - op_node['attrs'] = attrs - param = {} - param['op_type'] = self.op_type - param['op_list'] = [op_node] - objstr = json.dumps(param, indent=' ') - md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() - while key_map.get(md5sum) is not None: - objstr += '1' - md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() - key_map[md5sum] = md5sum - bin_file = self.op_type + '_' + md5sum - op_node['bin_filename'] = bin_file - param_file = os.path.join(self.out_path, bin_file + '_param.json') - param_file = os.path.realpath(param_file) - with os.fdopen(os.open(param_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - json.dump(param, fd, indent=' ') - self._write_buld_cmd(param_file, bin_file, i) - - - def _write_buld_cmd(self: any, param_file: str, bin_file: str, index: int): - hard_soc = const_var.SOC_MAP_EXT.get(self.soc) - if not hard_soc: - hard_soc = soc.capitalize() - name_com = [self.op_type, self.op_file, str(index)] - compile_file = os.path.join(self.out_path, '-'.join(name_com) + '.sh') - compile_file = os.path.realpath(compile_file) - with os.fdopen(os.open(compile_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - fd.write('#!/bin/bash\n') - fd.write('echo "[{}] Generating {} ..."\n'.format(hard_soc, bin_file)) - cmd = const_var.BIN_CMD.format(fun=self.op_intf, soc=hard_soc, param=param_file, impl='""') - fd.write(cmd) - chk = const_var.CHK_CMD.format(res_file=bin_file + '.json') - fd.write(chk) - chk = const_var.CHK_CMD.format(res_file=bin_file + '.o') - fd.write(chk) - fd.write('echo "[{}] Generating {} Done"\n'.format(hard_soc, bin_file)) - - -def gen_bin_param_file(cfgfile: str, out_dir: str, soc: str): - op_descs = opdesc_parser.get_op_desc(cfgfile, [], [], BinParamBuilder, None) - for op_desc in op_descs: - op_desc.set_soc_version(soc) - op_desc.set_out_path(out_dir) - op_desc.gen_input_json() - - -if __name__ == '__main__': - if len(sys.argv) <= 3: - raise RuntimeError('arguments must greater than 3') - gen_bin_param_file(sys.argv[1], sys.argv[2], sys.argv[3]) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_impl_build.py b/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_impl_build.py deleted file mode 100644 index 7fe177da1..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_impl_build.py +++ /dev/null @@ -1,446 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import re -import stat -import opdesc_parser -import const_var - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - -IMPL_HEAD = ''' -import os, sys -import ctypes -import json -import shutil -from tbe.common.platform import get_soc_spec -from tbe.common.utils import para_check -from tbe.tikcpp import compile_op, replay_op, check_op_cap, generalize_op_params, get_code_channel, OpInfo -from tbe.common.buildcfg import get_default_build_config -from impl.util.platform_adapter import tbe_register -from tbe.common.buildcfg import get_current_build_config -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - -DTYPE_MAP = {"float32": ["DT_FLOAT", "float"], - "float16": ["DT_FLOAT16", "half"], - "int8": ["DT_INT8", "int8_t"], - "int16": ["DT_INT16", "int16_t"], - "int32": ["DT_INT32", "int32_t"], - "int64": ["DT_INT64", "int64_t"], - "uint1": ["DT_UINT1", "uint8_t"], - "uint8": ["DT_UINT8", "uint8_t"], - "uint16": ["DT_UINT16", "uint16_t"], - "uint32": ["DT_UINT32", "uint32_t"], - "uint64": ["DT_UINT64", "uint64_t"], - "bool": ["DT_BOOL", "bool"], - "double": ["DT_DOUBLE", "double"], - "dual": ["DT_DUAL", "unknown"], - "dual_sub_int8": ["DT_DUAL_SUB_INT8", "unknown"], - "dual_sub_uint8": ["DT_DUAL_SUB_UINT8", "unknown"], - "string": ["DT_STRING", "unknown"], - "complex64": ["DT_COMPLEX64", "unknown"], - "complex128": ["DT_COMPLEX128", "unknown"], - "qint8": ["DT_QINT8", "unknown"], - "qint16": ["DT_QINT16", "unknown"], - "qint32": ["DT_QINT32", "unknown"], - "quint8": ["DT_QUINT8", "unknown"], - "quint16": ["DT_QUINT16", "unknown"], - "resource": ["DT_RESOURCE", "unknown"], - "string_ref": ["DT_STRING_REF", "unknown"], - "int4": ["DT_INT4", "int8_t"], - "bfloat16": ["DT_BF16", "bfloat16_t"]} - -def get_dtype_fmt_options(__inputs__, __outputs__): - options = [] - for x in __inputs__ + __outputs__: - x_n = x.get("param_name").upper() - x_fmt = x.get("format") - x_dtype = x.get("dtype") - options.append("-DDTYPE_{n}={t}".format(n=x_n, t=DTYPE_MAP.get(x_dtype)[1])) - options.append("-DORIG_DTYPE_{n}={ot}".format(n=x_n, ot=DTYPE_MAP.get(x_dtype)[0])) - options.append("-DFORMAT_{n}=FORMAT_{f}".format(n=x_n, f=x_fmt)) - return options - -def load_dso(so_path): - try: - ctypes.CDLL(so_path) - except OSError as error : - print(error) - raise RuntimeError("cannot open %s" %(so_path)) - else: - print("load so succ ", so_path) - -def get_shortsoc_compile_option(compile_option_list: list, shortsoc:str): - compile_options = [] - if shortsoc in compile_option_list: - compile_options = compile_option_list[shortsoc] - elif '__ALLSOC__' in compile_option_list: - compile_options = compile_option_list['__ALLSOC__'] - return compile_options - -''' - -IMPL_API = ''' -@tbe_register.register_operator("{}") -@para_check.check_op_params({}) -def {}({}, kernel_name="{}", impl_mode=""): - if get_current_build_config("enable_op_prebuild"): - return - __inputs__, __outputs__, __attrs__ = _build_args({}) - options = get_dtype_fmt_options(__inputs__, __outputs__) - options += ["-x", "cce"] - ccec = os.environ.get('CCEC_REAL_PATH') - if ccec is None: - ccec = shutil.which("ccec") - if ccec != None: - ccec_path = os.path.dirname(ccec) - tikcpp_path = os.path.realpath(os.path.join(ccec_path, "..", "..", "tikcpp")) - else: - tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") - options.append("-I" + tikcpp_path) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) - options.append("-I" + os.path.join(PYF_PATH, "..", "ascendc", "common")) - if impl_mode == "high_performance": - options.append("-DHIGH_PERFORMANCE=1") - elif impl_mode == "high_precision": - options.append("-DHIGH_PRECISION=1") - if get_default_build_config("enable_deterministic_mode") == 1: - options.append("-DDETEMINISTIC_MODE=1") - - custom_compile_options = {}, - custom_all_compile_options = {}, - soc_version = get_soc_spec("SOC_VERSION") - soc_short = get_soc_spec("SHORT_SOC_VERSION").lower() - custom_compile_options_soc = get_shortsoc_compile_option(custom_compile_options[0], soc_short) - custom_all_compile_options_soc = get_shortsoc_compile_option(custom_all_compile_options[0], soc_short) - options += custom_all_compile_options_soc - options += custom_compile_options_soc - - origin_func_name = "{}" - ascendc_src_dir = "{}" - ascendc_src_file = "{}" - src = os.path.join(PYF_PATH, "..", "ascendc", ascendc_src_dir, ascendc_src_file) - if not os.path.exists(src): - src = os.path.join(PYF_PATH, ascendc_src_file) -''' - -REPLAY_OP_API = ''' - print("start replay Acend C Operator {}, kernel name is {}") - tikreplay_codegen_path = tikcpp_path + "/tikreplaylib/lib" - tikreplay_stub_path = tikcpp_path + "/tikreplaylib/lib/" + soc_version - print("start load libtikreplaylib_codegen.so and libtikreplaylib_stub.so") - codegen_so_path = tikreplay_codegen_path + "/libtikreplaylib_codegen.so" - replaystub_so_path = tikreplay_stub_path + "/libtikreplaylib_stub.so" - if PYF_PATH.endswith("dynamic"): - op_replay_path = os.path.join(PYF_PATH, "..", "..", "op_replay") - else: - op_replay_path = os.path.join(PYF_PATH, "..", "op_replay") - replayapi_so_path = os.path.join(op_replay_path, "libreplay_{}_" + soc_short + ".so") - load_dso(codegen_so_path) - load_dso(replaystub_so_path) - load_dso(replayapi_so_path) - op_type = "{}" - entry_obj = os.path.join(op_replay_path, "{}_entry_" + soc_short + ".o") - code_channel = get_code_channel(src, kernel_name, op_type, options) - op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ - attrs = __attrs__, impl_mode = impl_mode) - res, msg = replay_op(op_info, entry_obj, code_channel, src, options) - if not res: - print("call replay op failed for %s and get into call compile op" %(msg)) - compile_op(src, origin_func_name, op_info, options, code_channel, '{}') -''' - -COMPILE_OP_API = ''' - print("start compile Ascend C operator {}. kernel name is {}") - op_type = "{}" - code_channel = get_code_channel(src, kernel_name, op_type, options) - op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ - attrs = __attrs__, impl_mode = impl_mode, origin_inputs=[{}], origin_outputs = [{}]) - compile_op(src, origin_func_name, op_info, options, code_channel, '{}') -''' - -SUP_API = ''' -def {}({}, impl_mode=""): - __inputs__, __outputs__, __attrs__ = _build_args({}) - ret_str = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) - ret_dict = json.loads(ret_str) - err_code = ret_dict.get("ret_code") - sup = "Unknown" - reason = "Unknown reason" - if err_code is not None: - if err_code is 0: - sup = "True" - reason = "" - elif err_code is 1: - sup = "False" - reason = ret_dict.get("reason") - else: - sup = "Unknown" - reason = ret_dict.get("reason") - return sup, reason -''' -CAP_API = ''' -def {}({}, impl_mode=""): - __inputs__, __outputs__, __attrs__ = _build_args({}) - result = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) - return result.decode("utf-8") -''' -GLZ_API = ''' -@tbe_register.register_param_generalization("{}") -def {}_generalization({}, generalize_config=None): - __inputs__, __outputs__, __attrs__ = _build_args({}) - ret_str = generalize_op_params("{}", __inputs__, __outputs__, __attrs__, generalize_config) - return [json.loads(ret_str)] -''' - -ATTR_DEFAULT = {'bool': 'False', 'int': '0', 'float': '0.0', 'listInt': '[]', - 'listFloat': '[]', 'listBool': '[]', 'listListInt': '[[]]', 'str': ''} - - -def optype_snake(origin_str): - temp_str = origin_str[0].lower() + origin_str[1:] - new_str = re.sub(r'([A-Z])', r'_\1', temp_str).lower() - return new_str - - -class AdpBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - self.argsname = [] - self.argsdefv = [] - self.op_compile_option:str = '{}' - super().__init__(op_type) - - - def write_adapt(self: any, impl_path, path: str, op_compile_option_all:list = None): - self._build_paradefault() - if impl_path != "": - src_file = os.path.join(impl_path, self.op_file + '.cpp') - if not os.path.exists(src_file): - return - out_path = os.path.abspath(path) - if self.dynamic_shape and not out_path.endswith('dynamic'): - out_path = os.path.join(path, 'dynamic') - os.makedirs(out_path, mode=0o700, exist_ok=True) - adpfile = os.path.join(out_path, self.op_file + '.py') - self._gen_op_compile_option(op_compile_option_all) - with os.fdopen(os.open(adpfile, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - self._write_head(fd) - self._write_argparse(fd) - self._write_impl(fd) - if self.op_chk_support: - self._write_cap('check_supported', fd) - self._write_cap('get_op_support_info', fd) - if self.op_fmt_sel: - self._write_cap('op_select_format', fd) - self._write_cap('get_op_specific_info', fd) - if self.op_range_limit == 'limited' or self.op_range_limit == 'dynamic': - self._write_glz(fd) - - - def _gen_op_compile_option(self:any, op_compile_option_all:list =None): - if op_compile_option_all is not None: - if self.op_type in op_compile_option_all: - self.op_compile_option = op_compile_option_all[self.op_type] - elif "__all__" in op_compile_option_all: - self.op_compile_option = op_compile_option_all["__all__"] - - - def _ip_argpack(self: any, default: bool = True) -> list: - args = [] - for i in range(len(self.input_name)): - arg = self.input_name[i] - if default and self.argsdefv[i] is not None: - arg += '=' + self.argsdefv[i] - args.append(arg) - return args - - def _op_argpack(self: any, default: bool = True) -> list: - args = [] - argidx = len(self.input_name) - for i in range(len(self.output_name)): - arg = self.output_name[i] - if default and self.argsdefv[i + argidx] is not None: - arg += '=' + self.argsdefv[i + argidx] - args.append(arg) - return args - - def _attr_argpack(self: any, default: bool = True) -> list: - args = [] - argidx = len(self.input_name) + len(self.output_name) - for i in range(len(self.attr_list)): - att = self.attr_list[i] - arg = att - if default and self.argsdefv[i + argidx] is not None: - if self.attr_val.get(att).get('type') == 'str': - arg += '="' + self.argsdefv[i + argidx] + '"' - elif self.attr_val.get(att).get('type') == 'bool': - arg += '=' + self.argsdefv[i + argidx].capitalize() - else: - arg += '=' + self.argsdefv[i + argidx] - args.append(arg) - return args - - def _build_paralist(self: any, default: bool = True) -> str: - args = [] - args.extend(self._ip_argpack(default)) - args.extend(self._op_argpack(default)) - args.extend(self._attr_argpack(default)) - return ', '.join(args) - - def _io_parachk(self: any, types: list, type_name: str) -> list: - chk = [] - for iot in types: - if iot == 'optional': - ptype = 'OPTION' - else: - ptype = iot.upper() - chk.append('para_check.{}_{}'.format(ptype, type_name)) - return chk - - def _attr_parachk(self: any) -> list: - chk = [] - for att in self.attr_list: - if self.attr_val.get(att).get('paramType') == 'optional': - pt = 'OPTION' - else: - pt = self.attr_val.get(att).get('paramType').upper() - att_type = self.attr_val.get(att).get('type').upper() - att_type = att_type.replace('LIST', 'LIST_') - chk.append('para_check.{}_ATTR_{}'.format(pt, att_type)) - return chk - - def _build_parachk(self: any) -> str: - chk = [] - chk.extend(self._io_parachk(self.input_type, 'INPUT')) - chk.extend(self._io_parachk(self.output_type, 'OUTPUT')) - chk.extend(self._attr_parachk()) - chk.append('para_check.KERNEL_NAME') - return ', '.join(chk) - - def _build_paradefault(self: any): - optional = False - argtypes = [] - argtypes.extend(self.input_type) - argtypes.extend(self.output_type) - for atype in argtypes: - if atype == 'optional': - optional = True - if optional: - self.argsdefv.append('None') - else: - self.argsdefv.append(None) - for attr in self.attr_list: - atype = self.attr_val.get(attr).get('paramType') - if atype == 'optional': - optional = True - attrval = self.attr_val.get(attr).get('defaultValue') - if attrval is not None: - optional = True - if type == "bool": - attrval = attrval.capitalize() - elif type == "str": - attrval = "\"" + attrval + "\"" - self.argsdefv.append(attrval) - continue - if optional: - self.argsdefv.append(ATTR_DEFAULT.get(self.attr_val.get(attr).get('type'))) - else: - self.argsdefv.append(None) - - def _write_head(self: any, fd: object): - fd.write(IMPL_HEAD) - - def _write_argparse(self: any, fd: object): - args = self._build_paralist(False) - fd.write('def _build_args({}):\n'.format(args)) - fd.write(' __inputs__ = []\n') - fd.write(' for arg in [{}]:\n'.format(', '.join(self.input_name))) - fd.write(' if arg != None:\n') - fd.write(' if isinstance(arg, (list, tuple)):\n') - fd.write(' if len(arg) == 0:\n') - fd.write(' continue\n') - fd.write(' __inputs__.append(arg[0])\n') - fd.write(' else:\n') - fd.write(' __inputs__.append(arg)\n') - fd.write(' __outputs__ = []\n') - fd.write(' for arg in [{}]:\n'.format(', '.join(self.output_name))) - fd.write(' if arg != None:\n') - fd.write(' if isinstance(arg, (list, tuple)):\n') - fd.write(' if len(arg) == 0:\n') - fd.write(' continue\n') - fd.write(' __outputs__.append(arg[0])\n') - fd.write(' else:\n') - fd.write(' __outputs__.append(arg)\n') - fd.write(' __attrs__ = []\n') - for attr in self.attr_list: - fd.write(' if {} != None:\n'.format(attr)) - fd.write(' attr = {}\n') - fd.write(' attr["name"] = "{}"\n'.format(attr)) - fd.write(' attr["dtype"] = "{}"\n'.format(self.attr_val.get(attr).get('type'))) - fd.write(' attr["value"] = {}\n'.format(attr)) - fd.write(' __attrs__.append(attr)\n') - fd.write(' return __inputs__, __outputs__, __attrs__\n') - - def _write_impl(self: any, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - pchk = self._build_parachk() - if len(self.kern_name) > 0: - kern_name = self.kern_name - else: - kern_name = self.op_intf - src = self.op_file + '.cpp' - fd.write(IMPL_API.format(self.op_type, pchk, self.op_intf, argsdef, kern_name, argsval,\ - self.custom_compile_options, self.custom_all_compile_options, self.op_intf,\ - optype_snake(self.op_type), src)) - if self.op_replay_flag: - fd.write(REPLAY_OP_API.format(self.op_type, kern_name, self.op_file, self.op_type, self.op_file,\ - self.op_compile_option)) - else: - fd.write(COMPILE_OP_API.format(self.op_type, kern_name, self.op_type, ', '.join(self.input_name),\ - ', '.join(self.output_name), self.op_compile_option)) - - def _write_cap(self: any, cap_name: str, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - if cap_name == 'check_supported': - fd.write(SUP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) - else: - fd.write(CAP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) - - def _write_glz(self: any, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - fd.write(GLZ_API.format(self.op_type, self.op_intf, argsdef, argsval, self.op_type)) - - -def write_scripts(cfgfile: str, cfgs: dict, dirs: dict, ops: list = None, op_compile_option:list = None): - batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') - iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') - file_map = {} - op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, AdpBuilder,\ - ops, dirs.get(const_var.AUTO_GEN_DIR)) - for op_desc in op_descs: - op_desc.write_adapt(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), op_compile_option) - file_map[op_desc.op_type] = op_desc.op_file - return file_map - -if __name__ == '__main__': - if len(sys.argv) <= 6: - raise RuntimeError('arguments must greater equal than 6') - rep_cfg = {} - rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] - rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] - cfg_dir = {} - cfg_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] - cfg_dir[const_var.CFG_OUT_DIR] = sys.argv[5] - cfg_dir[const_var.AUTO_GEN_DIR] = sys.argv[6] - write_scripts(cfgfile=sys.argv[1], cfgs=rep_cfg, dirs=cfg_dir) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_ops_config.py b/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_ops_config.py deleted file mode 100644 index 7a97180be..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_ops_config.py +++ /dev/null @@ -1,114 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import glob -import json -import argparse -import const_var - - -def load_json(json_file: str): - with open(json_file, encoding='utf-8') as file: - json_content = json.load(file) - return json_content - - -def get_specified_suffix_file(root_dir, suffix): - specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) - all_suffix_files = glob.glob(specified_suffix, recursive=True) - return all_suffix_files - - -def add_simplified_config(op_type, key, core_type, objfile, config): - simple_cfg = config.get('binary_info_config.json') - op_cfg = simple_cfg.get(op_type) - if not op_cfg: - op_cfg = {} - op_cfg['dynamicRankSupport'] = True - op_cfg['simplifiedKeyMode'] = 0 - op_cfg['binaryList'] = [] - simple_cfg[op_type] = op_cfg - bin_list = op_cfg.get('binaryList') - bin_list.append({'coreType': core_type, 'simplifiedKey': key, 'binPath': objfile}) - - -def add_op_config(op_file, bin_info, config): - op_cfg = config.get(op_file) - if not op_cfg: - op_cfg = {} - op_cfg['binList'] = [] - config[op_file] = op_cfg - op_cfg.get('binList').append(bin_info) - - -def gen_ops_config(json_file, soc, config): - core_type_map = {"MIX": 0, "AiCore": 1, "VectorCore": 2} - contents = load_json(json_file) - if ('binFileName' not in contents) or ('supportInfo' not in contents): - return - json_base_name = os.path.basename(json_file) - op_dir = os.path.basename(os.path.dirname(json_file)) - support_info = contents.get('supportInfo') - bin_name = contents.get('binFileName') - bin_suffix = contents.get('binFileSuffix') - core_type = core_type_map.get(contents.get("coreType")) - bin_file_name = bin_name + bin_suffix - op_type = bin_name.split('_')[0] - op_file = op_dir + '.json' - bin_info = {} - keys = support_info.get('simplifiedKey') - if keys: - bin_info['simplifiedKey'] = keys - for key in keys: - add_simplified_config(op_type, key, core_type, os.path.join(soc, op_dir, bin_file_name), config) - bin_info['staticKey'] = support_info.get('staticKey') - bin_info['int64Mode'] = support_info.get('int64Mode') - bin_info['inputs'] = support_info.get('inputs') - bin_info['outputs'] = support_info.get('outputs') - if support_info.get('attrs'): - bin_info['attrs'] = support_info.get('attrs') - bin_info['binInfo'] = {'jsonFilePath': os.path.join(soc, op_dir, json_base_name)} - add_op_config(op_file, bin_info, config) - - -def gen_all_config(root_dir, soc): - suffix = 'json' - config = {} - config['binary_info_config.json'] = {} - all_json_files = get_specified_suffix_file(root_dir, suffix) - for _json in all_json_files: - gen_ops_config(_json, soc, config) - for cfg_key in config.keys(): - cfg_file = os.path.join(root_dir, cfg_key) - with os.fdopen(os.open(cfg_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - json.dump(config.get(cfg_key), fd, indent=' ') - - -def args_prase(): - parser = argparse.ArgumentParser() - parser.add_argument('-p', - '--path', - nargs='?', - required=True, - help='Parse the path of the json file.') - parser.add_argument('-s', - '--soc', - nargs='?', - required=True, - help='Parse the soc_version of ops.') - return parser.parse_args() - - -def main(): - args = args_prase() - gen_all_config(args.path, args.soc) - - -if __name__ == '__main__': - main() diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_replay_build.py b/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_replay_build.py deleted file mode 100644 index 1cac7d911..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/ascendc_replay_build.py +++ /dev/null @@ -1,65 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import opdesc_parser -import replay_codegen -import const_var -from replay_codegen import ReplayCodeGenParams - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - - -class ReplayBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - super().__init__(op_type) - - def gen_replay_source(self: any, impl_path: str, out_path: str, ops_product: str): - if not self.op_replay_flag: - print('{} replay not enabled'.format(self.op_type)) - return - argn = len(self.input_name) + len(self.output_name) + 1 - if self.op_replay_batch: - print('{} replay in batch mode'.format(self.op_type)) - else: - print('{} replay in normal mode'.format(self.op_type)) - if impl_path.endswith('op_kernel'): - implf = os.path.join(impl_path, self.op_file + '.cpp') - tiling_file = os.path.join(impl_path, "../op_host", self.op_file + '_tiling.h') - else: - if self.dynamic_shape: - dyn_path = 'dynamic' - else: - dyn_path = '' - implf = os.path.join(impl_path, dyn_path, self.op_file + '.cpp') - tiling_file = os.path.join(impl_path, "../../op_tiling", self.op_file + '_tiling.h') - rep_conf = replay_codegen.ReplayCodeGen(ReplayCodeGenParams(self.op_type, implf, tiling_file, self.op_file, \ - self.op_intf, argn, self.op_replay_batch, self.max_block_dim, self.max_shape_size)) - rep_conf.set_batch(self.op_replay_batch) - rep_conf.set_outdir(out_path) - rep_conf.gen_replay(ops_product) - - -def gen_replay(cfgfile: str, cfgs: dict, dirs: dict, ops_product: str, ops: list = None): - batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') - iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') - op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, ReplayBuilder, ops) - for op_desc in op_descs: - op_desc.gen_replay_source(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), ops_product) - - -if __name__ == '__main__': - if len(sys.argv) <= 6: - raise RuntimeError('arguments must greater than 6') - rep_cfg = {} - rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] - rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] - rep_dir = {} - rep_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] - rep_dir[const_var.CFG_OUT_DIR] = sys.argv[5] - gen_replay(sys.argv[1], rep_cfg, rep_dir, sys.argv[6]) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/batch_replay_impl.temp b/atb_operator/AtbAddSample/AddOp/cmake/util/batch_replay_impl.temp deleted file mode 100644 index 0e8834664..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/batch_replay_impl.temp +++ /dev/null @@ -1,117 +0,0 @@ -#include -#include -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#include "register/op_check.h" -#define __ASCENDC_REPLAY_CODE__ -#include - -using namespace std; -using namespace optiling; -using namespace AscendCReplay; - -extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); -extern "C" int elf_batch_append(char *elf, uint32_t elfSize, char *jit, int kernum, char *atext[], int alen[], - int atlen, const char* kernelname[]); - -#define KERNEL_N 1 -#define ARG_N (__ARG_NUM__) -#define MAX_L (1024 * 1024 * 100) -#define MAX_E (1024 * 1024) - -int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) -{ - // gen type 1 : direct call codes 0: load .o file - if (param.gentype < 0 || param.gentype > 1) { - printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); - return 0; - } else if (param.gentype == 1 && param.objptr == nullptr) { - printf("Error: call replay with direct call mode, but code obj addr is null\n"); - return 0; - } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { - printf("Error: call replay with object file mode, but object file path is null\n"); - return 0; - } - // core_type 0:MIX 1:CUBE 2:VEC - if (core_type < 0 || core_type > 2) { - printf("Error: call replay core type is %d !\n", core_type); - return 0; - } - g_coreType = __CORE_TYPE__; - g_taskRation = param.task_ration; - g_tilingKey = param.tiling_key; - - unsigned char *buf, *jit; - char *kernel[KERNEL_N]; - int len[KERNEL_N]; - block_idx = 0; - block_num = param.block_dim; - g_ubBase = block_num; - uint8_t *code = (uint8_t *)malloc(MAX_L); - uint8_t *pos = code; - struct timespec tp1, tp2; - - clock_gettime(CLOCK_MONOTONIC, &tp1); - if (block_num > 32) { - printf("Error: block_num > 32\n"); - return 0; - } - //__OP_FOPEN__ - for (int i = 0; i < KERNEL_N; i++) { - //__OP_SET_KERNEL__ - for (int j = 0; j < ARG_N; j++) - AddArg(j, ARG_STEP * (j + 1)); -#ifdef FP_CEILING - SetCtrlFloatEnable(); -#else - SetCtrlFloatDisable(); -#endif - CodeInit(pos, true); - __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); - CodeEnd(); - kernel[i] = (char *)pos; - len[i] = CodeLen(); - pos += len[i]; - } - //__OP_FCLOSE__ - clock_gettime(CLOCK_MONOTONIC, &tp2); - buf = (unsigned char *)malloc(MAX_E); - int fd = open(param.entry_file, O_RDONLY); - if (fd < 0) { - printf("[error]: cannot find entry.o : %s\n", param.entry_file); - return 0; - } - uint32_t bufSize = read(fd, buf, MAX_E); - if (bufSize <= 0) { - printf("[error]: entry.o : %s is too small ! \n", param.entry_file); - } - close(fd); - jit = (unsigned char *)malloc(MAX_L); - printf("total code generated %ld\n", pos - code); - int sz = elf_batch_append((char *)buf, bufSize, (char *)jit, KERNEL_N, kernel, len, pos - code, ¶m.kernel_name); - if (tp1.tv_sec != tp2.tv_sec) { - printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); - } else { - printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); - } - printf("new elf size %d\n", sz); - if (param.gentype == 0) { - fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); - (void)write(fd, jit, sz); - close(fd); - free(jit); - } else if (param.gentype == 1) { - *param.objptr = (char*)jit; - } - free(buf); - free(code); - return sz; -} - -REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/code_channel_infer.py b/atb_operator/AtbAddSample/AddOp/cmake/util/code_channel_infer.py deleted file mode 100644 index 137c9f39f..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/code_channel_infer.py +++ /dev/null @@ -1,115 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" -import os -import stat -import ctypes -import collections -import shutil -import subprocess -import copy - -"""CODE_* is used to cube/vector api is called in operator code -CODE_MIX means both cube and vector api is called -CODE_CUBE means only cube api is called -CODE_VEC means only vector api is called -""" -CODE_MIX = 0 -CODE_CUBE = 1 -CODE_VEC = 2 - - -def _is_v220(op_product: str): - """return if current soc version is V220 - - Returns: - res: True means V220 - """ - if op_product in ["ascend910b"]: - return True - return False - - -InfoCodeChanelParams = collections.namedtuple('InfoCodeChanelParams',\ -['src_file', 'tiling_header', 'kernel_name', 'outdir', 'op_product', 'compile_options']) - - -def infer_code_channel(params: InfoCodeChanelParams): - """get code channel for v220, return CODE_MIX if soc version is not V220 - - Args: - src_file (str): AscendC operator code file - src_file (str): AscendC operator tiling header file - kernel_name (str): kernel function name - optype (str): operator type - compile_options (list): compile options for ccec cmd - - Raises: - Exception: if not exist L1/L0/UB if code, it's not a aicore code - - Returns: - res (int): CODE_MIX/CODE_CUBE/CODE_VEC - """ - if not _is_v220(params.op_product): - return CODE_MIX - return CODE_VEC - if params.compile_options is None: - compile_options = [] - else: - compile_options = params.compile_options - ccec = shutil.which("ccec") - if ccec is not None: - ccec_path = os.path.dirname(ccec) - tikcpp_path = os.path.realpath(os.path.join(ccec_path, "..", "..", "tikcpp")) - else: - tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") - compile_options.append("-I" + tikcpp_path) - compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) - compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) - compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) - compile_options += ["-include", params.tiling_header] - arch = "dav-c220-cube" - sub_core_type = "AIC" - optional_core = "AiCore" - compile_cmd = [shutil.which("ccec"), '-c', '-O3'] - compile_cmd += compile_options - temp_file_name_tag = "_" + str(os.getpid()) + "_temp.o" - dst_file = os.path.join(kernel_meta_dir, kernel_name + temp_file_name_tag) - compile_cmd += [params.src_file, "--cce-aicore-arch={}".format(arch), - "--cce-aicore-only", "-o", dst_file, - "-mllvm", "-cce-aicore-function-stack-size=16000", - "-mllvm", "-cce-aicore-record-overflow=true", - "-mllvm", "-cce-aicore-addr-transform"] - compile_cmd += ["-std=c++17"] - print('get_code_channel: ', ' '.join(compile_cmd)) - proc = subprocess.Popen( - compile_cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) - (out, _) = proc.communicate() - if proc.returncode != 0: - print('get_code_channel coretype compile error: ', out.decode()) - msg = "compile %s error :%s\n" % (params.src_file, out.decode()) - raise Exception(f"get_code_channel coretype error, msg is{msg}") - objdump_cmd = ['objdump', '-s', '-j', '.text', '{}'.format(dst_file)] - - proc = subprocess.Popen( - objdump_cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) - (out, _) = proc.communicate() - if proc.returncode != 0: - print('get_code_channel objdump error: ', out.decode()) - msg = "get_code_channel objdump %s error :%s\n" % (src_file, out.decode()) - raise Exception(f"get_code_channel objdump error, msg is{msg}") - os.remove(dst_file) - lines = out.decode('utf-8').split('\n') - for line in lines: - insts = line.strip().split() - if len(insts) < 5: - continue - for inst in insts[1:]: - if len(inst) != 8: - continue - if inst[6] == 'f': - return CODE_MIX - return CODE_VEC diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/const_var.py b/atb_operator/AtbAddSample/AddOp/cmake/util/const_var.py deleted file mode 100644 index 8b32c3b91..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/const_var.py +++ /dev/null @@ -1,33 +0,0 @@ - -#!/usr/bin/env python -# coding=utf-8 -""" -Function: -The replay funtion entry -Copyright Information: -Huawei Technologies Co., Ltd. All Rights Reserved © 2020 -""" - -import os -import stat - - -REPLAY_BATCH = 'batch' -REPLAY_ITERATE = 'iterate' -CFG_IMPL_DIR = 'impl_dir' -CFG_OUT_DIR = 'out_dir' -AUTO_GEN_DIR = 'auto_gen_dir' -WFLAGS = os.O_WRONLY | os.O_CREAT | os.O_TRUNC -WMODES = stat.S_IWUSR | stat.S_IRUSR -SOC_MAP_EXT = {'ascend310p': 'Ascend310P3', 'ascend310b': 'Ascend310B1', - 'ascend910': 'Ascend910A', 'ascend910b': 'Ascend910B1'} -BIN_CMD = 'opc $1 --main_func={fun} --input_param={param} --soc_version={soc} \ ---output=$2 --impl_mode={impl} --simplified_key_mode=0 --op_mode=dynamic\n' -CHK_CMD = ''' -if ! test -f $2/{res_file} ; then - echo "$2/{res_file} not generated!" - exit 1 -fi -''' -ATTR_DEF_VAL = {'str' : '', 'int': 0, 'float': 0.0, 'bool': False, 'list_bool': [], - 'list_int': [], 'list_float': [], 'list_list_int': [[]]} diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/gen_impl_and_mrege_json.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/gen_impl_and_mrege_json.sh deleted file mode 100644 index 55e12e5ed..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/gen_impl_and_mrege_json.sh +++ /dev/null @@ -1,57 +0,0 @@ -#!/usr/bin/bash - -project_path=$1 -build_path=$2 -vendor_name=customize -if [[ ! -d "$project_path" ]]; then - echo "[ERROR] No projcet path is provided" - exit 1 -fi - -if [[ ! -d "$build_path" ]]; then - echo "[ERROR] No build path is provided" - exit 1 -fi - -# copy ai_core operators implements -# tbe_impl_files_num=$(ls $project_path/tbe/impl/* 2> /dev/null | wc -l) -# if [[ "$tbe_impl_files_num" -gt 0 ]];then -# cp -r ${project_path}/tbe/impl/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/ai_core/tbe/customize_impl -# cp -r ${project_path}/tbe/impl/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/vector_core/tbe/customize_impl -# fi - -# copy aicpu kernel so operators -if [[ -d "${project_path}/cpukernel/aicpu_kernel_lib" ]]; then - cp -f ${project_path}/cpukernel/aicpu_kernel_lib/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/cpu/aicpu_kernel/impl - rm -rf ${project_path}/cpukernel/aicpu_kernel_lib -fi - -# merge aicpu.ini and aicore.ini to generate npu_supported_ops.json -# mkdir -p ${build_path}/framework/op_info_cfg -# mkdir -p ${build_path}/framework/op_info_cfg/aicpu_kernel -# mkdir -p ${build_path}/framework/op_info_cfg/ai_core - -# if [[ -d "${project_path}/tbe/op_info_cfg/ai_core" ]]; then -# bash ${project_path}/cmake/util/gen_ops_filter.sh ${project_path}/tbe/op_info_cfg/ai_core ${build_path}/framework/op_info_cfg/ai_core -# fi - -# if [[ -d "${project_path}/cpukernel/op_info_cfg/aicpu_kernel" ]]; then -# bash ${project_path}/cmake/util/gen_ops_filter.sh ${project_path}/cpukernel/op_info_cfg/aicpu_kernel ${build_path}/framework/op_info_cfg/aicpu_kernel -# fi - -# aicpu_filter_file=${build_path}/framework/op_info_cfg/aicpu_kernel/npu_supported_ops.json -# aicore_filter_file=${build_path}/framework/op_info_cfg/ai_core/npu_supported_ops.json -# if [[ -f "${aicpu_filter_file}" ]] && [[ ! -f "${aicore_filter_file}" ]]; then -# cp $aicpu_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow -# fi -# if [[ -f "${aicore_filter_file}" ]] && [[ ! -f "${aicpu_filter_file}" ]]; then -# cp $aicore_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow -# fi - -# if [[ -f "${aicore_filter_file}" ]] && [[ -f "${aicpu_filter_file}" ]]; then -# chmod u+w ${aicpu_filter_file} -# python3 ${project_path}/cmake/util/insert_op_info.py ${aicore_filter_file} ${aicpu_filter_file} -# chmod u-w ${aicpu_filter_file} -# cp $aicpu_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow -# fi - diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/gen_ops_filter.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/gen_ops_filter.sh deleted file mode 100644 index d4c27d17f..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/gen_ops_filter.sh +++ /dev/null @@ -1,62 +0,0 @@ -#!/bin/bash -# Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -# Description: Generate npu_supported_ops.json -# ============================================================================== - -if [[ -z "$1" ]]; then - echo -e "[ERROR] No source dir provided" - exit 1 -fi - -if [[ -z "$2" ]]; then - echo -e "[ERROR] No destination dir provided" - exit 1 -fi - -src=$1 -dest_file=$2/npu_supported_ops.json - -if [ -f "$dest_file" ];then - chmod u+w $dest_file -fi - -echo $* - -add_ops() { - name=$1 - isHeavy=$2 - file=$3 - grep -w "\"$name\"" ${file} >/dev/null - if [ $? == 0 ];then - return - fi - echo " \"${name}\": {" >> ${file} - echo " \"isGray\": false," >> ${file} - echo " \"isHeavy\": ${isHeavy}" >> ${file} - echo " }," >> ${file} -} - -echo "{" > ${dest_file} -ini_files=$(find ${src} -name "*.ini") -for file in ${ini_files} ; do - name=$(grep '^\[' ${file} | sed 's/\[//g' | sed 's/]//g' | sed 's/\r//g') - grep 'heavyOp.flag' ${file} >/dev/null - if [ $? == 0 ];then - isHeavy=$(grep 'heavyOp.flag' ${file} | awk -F= '{print $2}') - else - isHeavy="false" - fi - for op in ${name} ; do - add_ops ${op} "false" ${dest_file} - done -done -echo "}" >> ${dest_file} -file_count=$(cat ${dest_file} | wc -l) -line=$(($file_count-1)) -sed -i "${line}{s/,//g}" ${dest_file} - -chmod 640 "${dest_file}" -echo -e "[INFO] Succed generated ${dest_file}" - -exit 0 - diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/gen_version_info.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/gen_version_info.sh deleted file mode 100644 index a06cfc78d..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/gen_version_info.sh +++ /dev/null @@ -1,6 +0,0 @@ -ascend_install_dir=$1 -gen_file_dir=$2 - -# create version.info -compiler_version=$(grep "Version" -w ${ascend_install_dir}/compiler/version.info | awk -F = '{print $2}') -echo "custom_opp_compiler_version=${compiler_version}" > ${gen_file_dir}/version.info \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/insert_op_info.py b/atb_operator/AtbAddSample/AddOp/cmake/util/insert_op_info.py deleted file mode 100644 index 28ba08757..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/insert_op_info.py +++ /dev/null @@ -1,36 +0,0 @@ -# -*- coding: utf-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" -import json -import os -import sys -import stat -import const_var - - -if __name__ == '__main__': - if len(sys.argv) != 3: - print(sys.argv) - print('argv error, inert_op_info.py your_op_file lib_op_file') - sys.exit(2) - - with open(sys.argv[1], 'r') as load_f: - insert_operator = json.load(load_f) - - all_operators = {} - if os.path.exists(sys.argv[2]): - if os.path.getsize(sys.argv[2]) != 0: - with open(sys.argv[2], 'r') as load_f: - all_operators = json.load(load_f) - - for k in insert_operator.keys(): - if k in all_operators.keys(): - print('replace op:[', k, '] success') - else: - print('insert op:[', k, '] success') - all_operators[k] = insert_operator[k] - - with os.fdopen(os.open(sys.argv[2], const_var.WFLAGS, const_var.WMODES), 'w') as json_file: - json_file.write(json.dumps(all_operators, indent=4)) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/insert_simplified_keys.py b/atb_operator/AtbAddSample/AddOp/cmake/util/insert_simplified_keys.py deleted file mode 100644 index ace727b90..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/insert_simplified_keys.py +++ /dev/null @@ -1,248 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import re -import glob -import json -import argparse -import const_var - - -DATA_TPYE_DICT = { - 'float32': 0, - 'float16': 1, - 'int8': 2, - 'int16': 6, - 'uint16': 7, - 'uint8': 4, - 'int32': 3, - 'int64': 9, - 'uint32': 8, - 'uint64': 10, - 'bool': 12, - 'double': 11, - 'string': 13, - 'dual': 14, - 'dual': 15, - 'complex64': 16, - 'complex128': 17, - 'qint8': 18, - 'qint16': 19, - 'qint32': 20, - 'quint8': 21, - 'quint16': 22, - 'resource': 23, - 'string': 24, - 'dual': 25, - 'variant': 26, - 'bf16': 27, - 'bfloat16': 27, - 'undefined': 28, - 'int4': 29, - 'uint1': 30, - 'int2': 31 -} - -FORMAT_DICT = { - 'NCHW': 0, - 'NHWC': 1, - 'ND': 2, - 'NC1HWC0': 3, - 'FRACTAL_Z': 4, - 'NC1C0HWPAD': 5, - 'NHWC1C0': 6, - 'FSR_NCHW': 7, - 'FRACTAL_DECONV': 8, - 'C1HWNC0': 9, - 'FRACTAL_DECONV_TRANSPOSE': 10, - 'FRACTAL_DECONV_SP_STRIDE_TRANS': 11, - 'NC1HWC0_C04': 12, - 'FRACTAL_Z_C04': 13, - 'CHWN': 14, - 'FRACTAL_DECONV_SP_STRIDE8_TRANS': 15, - 'HWCN': 16, - 'NC1KHKWHWC0': 17, - 'BN_WEIGHT': 18, - 'FILTER_HWCK': 19, - 'HASHTABLE_LOOKUP_LOOKUPS': 20, - 'HASHTABLE_LOOKUP_KEYS': 21, - 'HASHTABLE_LOOKUP_VALUE': 22, - 'HASHTABLE_LOOKUP_OUTPUT': 23, - 'HASHTABLE_LOOKUP_HITS': 24, - 'C1HWNCoC0': 25, - 'MD': 26, - 'NDHWC': 27, - 'FRACTAL_ZZ': 28, - 'FRACTAL_NZ': 29, - 'NCDHW': 30, - 'DHWCN': 31, - 'NDC1HWC0': 32, - 'FRACTAL_Z_3D': 33, - 'CN': 34, - 'NC': 35, - 'DHWNC': 36, - 'FRACTAL_Z_3D_TRANSPOSE': 37, - 'FRACTAL_ZN_LSTM': 38, - 'FRACTAL_Z_G': 39, - 'RESERVED': 40, - 'ALL': 41, - 'NULL': 42, - 'ND_RNN_BIAS': 43, - 'FRACTAL_ZN_RNN': 44, - 'NYUV': 45, - 'NYUV_A': 46 -} - - -def load_json(json_file: str): - with open(json_file, encoding='utf-8') as file: - json_content = json.load(file) - return json_content - - -def get_specified_suffix_file(root_dir, suffix): - specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) - all_suffix_files = glob.glob(specified_suffix, recursive=True) - return all_suffix_files - - -def get_deterministic_value(support_info): - deterministic_key = 'deterministic' - if deterministic_key not in support_info: - return 0 - deterministic_value = support_info.get(deterministic_key) - if deterministic_value == 'true': - return 1 - else: - return 0 - - -def get_precision_value(support_info): - precision_key = 'implMode' - precision_value = support_info.get(precision_key) - if precision_value == 'high_performance': - _value = 1 - elif precision_value == 'high_precision': - _value = 2 - else: - _value = 0 - return _value - - -def get_overflow_value(support_info): - return 0 - - -def get_parameters(info): - if info: - if 'dtype' in info: - data_type = info['dtype'] - data_type_value = DATA_TPYE_DICT.get(data_type) - else: - data_type_value = 0 - if 'format' in info: - _format = info['format'] - _format_value = FORMAT_DICT.get(_format) - else: - _format_value = 0 - else: - data_type_value = 0 - _format_value = 0 - return str(data_type_value), str(_format_value) - - -def get_dynamic_parameters(info): - # 动态输入时只需获取第一个参数 - return get_parameters(info[0]) - - -def get_all_parameters(support_info, _type): - result_list = list() - info_lists = support_info.get(_type) - if info_lists: - for _info in info_lists: - # 输入为列表时是动态输入 - if isinstance(_info, (list, tuple)): - data_type_value, _format_value = get_dynamic_parameters(_info) - else: - data_type_value, _format_value = get_parameters(_info) - result_list.append("{},{}".format(data_type_value, _format_value)) - return result_list - - -def get_all_input_parameters(support_info): - result = get_all_parameters(support_info, 'inputs') - return '/'.join(result) - - -def insert_content_into_file(input_file, content): - with open(input_file, 'r+') as file: - lines = file.readlines() - for index, line in enumerate(lines): - match_result = re.search(r'"staticKey":', line) - if match_result: - count = len(line) - len(line.lstrip()) - new_content = "{}{}".format(' ' * count, content) - # 插入到前一行,防止插入最后时还需要考虑是否添加逗号 - lines.insert(index, new_content) - break - file.seek(0) - file.write(''.join(lines)) - - -def insert_simplified_keys(json_file): - contents = load_json(json_file) - # 不存在'binFileName'或者'supportInfo'字段时,非需要替换的解析json文件 - if ('binFileName' not in contents) or ('supportInfo' not in contents): - return - support_info = contents.get('supportInfo') - bin_file_name = contents.get('binFileName') - bin_suffix = contents.get('binFileSuffix') - # 'simplifiedKey'字段已经存在时,直接返回,不重复生成 - if 'simplifiedKey' in support_info: - return - op_type = bin_file_name.split('_')[0] - deterministic = str(get_deterministic_value(support_info)) - precision = str(get_precision_value(support_info)) - overflow = str(get_overflow_value(support_info)) - input_parameters = get_all_input_parameters(support_info) - key = '{}/d={},p={},o={}/{}/'.format( - op_type, - deterministic, - precision, - overflow, - input_parameters) - result = '"simplifiedKey": "' + key + '",\n' - insert_content_into_file(json_file, result) - - -def insert_all_simplified_keys(root_dir): - suffix = 'json' - all_json_files = get_specified_suffix_file(root_dir, suffix) - for _json in all_json_files: - insert_simplified_keys(_json) - - -def args_prase(): - parser = argparse.ArgumentParser() - parser.add_argument('-p', - '--path', - nargs='?', - required=True, - help='Parse the path of the json file.') - return parser.parse_args() - - -def main(): - args = args_prase() - insert_all_simplified_keys(args.path) - - -if __name__ == '__main__': - main() diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/kernel_entry.py b/atb_operator/AtbAddSample/AddOp/cmake/util/kernel_entry.py deleted file mode 100644 index 2b77c970d..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/kernel_entry.py +++ /dev/null @@ -1,115 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - - -def gen_fun_def(title, kernel, argn, arg_type, arg_name): - entry = [] - entry.append(title) - entry.append(kernel) - entry.append('(') - args = [] - for i in range(0, argn): - args.append(arg_type + ' ' + arg_name + str(i)) - entry.append(', '.join(args)) - entry.append(')') - return ' '.join(entry) - - -def gen_batch_kernel_body(fname, argn, arg_name): - body = [] - body.append('{') - fun = [] - fun.append(fname) - fun.append('(') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - fun.append(', '.join(args)) - fun.append(');') - body.append(' '.join(fun)) - body.append('}') - return '\n'.join(body) - - -def gen_mc_kernel_body(kn, argn, arg_name, blknum): - body = [] - body.append('{') - body.append(' switch(block_idx) {') - for blk in range(0, blknum): - fun = [] - fun.append('{}_blk{:02d}'.format(kn, blk)) - fun.append('(') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - fun.append(', '.join(args)) - fun.append(')') - body.append(' case {}: {}; break;'.format(blk, ' '.join(fun))) - body.append(' default: break;') - body.append(' }') - body.append('}') - return '\n'.join(body) - - -def gen_proc_body(argn, arg_name): - body = [] - body.append('{') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - body.append('uint64_t __x = (uint64_t)' + ' + (uint64_t)'.join(args) + ';') - body.append('__asm__ ("NOP");') - body.append('__asm__ ("NOP");') - body.append('__asm__ ("NOP");') - body.append('}') - return '\n'.join(body) - - -def batch_code_gen(kn, argn, argt): - codes = [] - kernel_name = kn - proc_name = kernel_name + '_percore' - arg_num = int(argn) - data_type = argt - arg_type = '__gm__ ' + data_type + '* __restrict__' - arg_name = 'arg' - kernel_title = 'extern \"C\" __global__ __aicore__ void' - proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' - codes.append('#ifndef __aicore__') - codes.append('#define __aicore__ [aicore]') - codes.append('#endif') - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') - codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) - codes.append(gen_batch_kernel_body(proc_name, arg_num, arg_name)) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) - codes.append(gen_proc_body(arg_num, arg_name)) - return '\n'.join(codes) + '\n' - - -def mc_code_gen(kn, argn, argt, blknum): - codes = [] - kernel_name = kn - core_num = int(blknum) - arg_num = int(argn) - data_type = argt - arg_type = '__gm__ ' + data_type + '* __restrict__' - arg_name = 'arg' - kernel_title = 'extern \"C\" __global__ __aicore__ void' - proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' - codes.append('#ifndef __aicore__') - codes.append('#define __aicore__ [aicore]') - codes.append('#endif') - for i in range(0, core_num): - proc_name = '{}_blk{:02d}'.format(kernel_name, i) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') - codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) - codes.append(gen_mc_kernel_body(kernel_name, arg_num, arg_name, core_num)) - for i in range(0, core_num): - proc_name = '{}_blk{:02d}'.format(kernel_name, i) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) - codes.append(gen_proc_body(arg_num, arg_name)) - return '\n'.join(codes) + '\n' diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/kernel_impl.temp b/atb_operator/AtbAddSample/AddOp/cmake/util/kernel_impl.temp deleted file mode 100644 index 5079a1043..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/kernel_impl.temp +++ /dev/null @@ -1,10 +0,0 @@ -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#define __ASCENDC_REPLAY_CODE__ -#include "__CCE_FILE__" diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/COPYING b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/COPYING deleted file mode 100644 index d159169d1..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/COPYING +++ /dev/null @@ -1,339 +0,0 @@ - GNU GENERAL PUBLIC LICENSE - Version 2, June 1991 - - Copyright (C) 1989, 1991 Free Software Foundation, Inc., - 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA - Everyone is permitted to copy and distribute verbatim copies - of this license document, but changing it is not allowed. - - Preamble - - The licenses for most software are designed to take away your -freedom to share and change it. By contrast, the GNU General Public -License is intended to guarantee your freedom to share and change free -software--to make sure the software is free for all its users. This -General Public License applies to most of the Free Software -Foundation's software and to any other program whose authors commit to -using it. (Some other Free Software Foundation software is covered by -the GNU Lesser General Public License instead.) You can apply it to -your programs, too. - - When we speak of free software, we are referring to freedom, not -price. Our General Public Licenses are designed to make sure that you -have the freedom to distribute copies of free software (and charge for -this service if you wish), that you receive source code or can get it -if you want it, that you can change the software or use pieces of it -in new free programs; and that you know you can do these things. - - To protect your rights, we need to make restrictions that forbid -anyone to deny you these rights or to ask you to surrender the rights. -These restrictions translate to certain responsibilities for you if you -distribute copies of the software, or if you modify it. - - For example, if you distribute copies of such a program, whether -gratis or for a fee, you must give the recipients all the rights that -you have. You must make sure that they, too, receive or can get the -source code. And you must show them these terms so they know their -rights. - - We protect your rights with two steps: (1) copyright the software, and -(2) offer you this license which gives you legal permission to copy, -distribute and/or modify the software. - - Also, for each author's protection and ours, we want to make certain -that everyone understands that there is no warranty for this free -software. If the software is modified by someone else and passed on, we -want its recipients to know that what they have is not the original, so -that any problems introduced by others will not reflect on the original -authors' reputations. - - Finally, any free program is threatened constantly by software -patents. We wish to avoid the danger that redistributors of a free -program will individually obtain patent licenses, in effect making the -program proprietary. To prevent this, we have made it clear that any -patent must be licensed for everyone's free use or not licensed at all. - - The precise terms and conditions for copying, distribution and -modification follow. - - GNU GENERAL PUBLIC LICENSE - TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION - - 0. This License applies to any program or other work which contains -a notice placed by the copyright holder saying it may be distributed -under the terms of this General Public License. The "Program", below, -refers to any such program or work, and a "work based on the Program" -means either the Program or any derivative work under copyright law: -that is to say, a work containing the Program or a portion of it, -either verbatim or with modifications and/or translated into another -language. (Hereinafter, translation is included without limitation in -the term "modification".) Each licensee is addressed as "you". - -Activities other than copying, distribution and modification are not -covered by this License; they are outside its scope. The act of -running the Program is not restricted, and the output from the Program -is covered only if its contents constitute a work based on the -Program (independent of having been made by running the Program). -Whether that is true depends on what the Program does. - - 1. You may copy and distribute verbatim copies of the Program's -source code as you receive it, in any medium, provided that you -conspicuously and appropriately publish on each copy an appropriate -copyright notice and disclaimer of warranty; keep intact all the -notices that refer to this License and to the absence of any warranty; -and give any other recipients of the Program a copy of this License -along with the Program. - -You may charge a fee for the physical act of transferring a copy, and -you may at your option offer warranty protection in exchange for a fee. - - 2. You may modify your copy or copies of the Program or any portion -of it, thus forming a work based on the Program, and copy and -distribute such modifications or work under the terms of Section 1 -above, provided that you also meet all of these conditions: - - a) You must cause the modified files to carry prominent notices - stating that you changed the files and the date of any change. - - b) You must cause any work that you distribute or publish, that in - whole or in part contains or is derived from the Program or any - part thereof, to be licensed as a whole at no charge to all third - parties under the terms of this License. - - c) If the modified program normally reads commands interactively - when run, you must cause it, when started running for such - interactive use in the most ordinary way, to print or display an - announcement including an appropriate copyright notice and a - notice that there is no warranty (or else, saying that you provide - a warranty) and that users may redistribute the program under - these conditions, and telling the user how to view a copy of this - License. (Exception: if the Program itself is interactive but - does not normally print such an announcement, your work based on - the Program is not required to print an announcement.) - -These requirements apply to the modified work as a whole. If -identifiable sections of that work are not derived from the Program, -and can be reasonably considered independent and separate works in -themselves, then this License, and its terms, do not apply to those -sections when you distribute them as separate works. But when you -distribute the same sections as part of a whole which is a work based -on the Program, the distribution of the whole must be on the terms of -this License, whose permissions for other licensees extend to the -entire whole, and thus to each and every part regardless of who wrote it. - -Thus, it is not the intent of this section to claim rights or contest -your rights to work written entirely by you; rather, the intent is to -exercise the right to control the distribution of derivative or -collective works based on the Program. - -In addition, mere aggregation of another work not based on the Program -with the Program (or with a work based on the Program) on a volume of -a storage or distribution medium does not bring the other work under -the scope of this License. - - 3. You may copy and distribute the Program (or a work based on it, -under Section 2) in object code or executable form under the terms of -Sections 1 and 2 above provided that you also do one of the following: - - a) Accompany it with the complete corresponding machine-readable - source code, which must be distributed under the terms of Sections - 1 and 2 above on a medium customarily used for software interchange; or, - - b) Accompany it with a written offer, valid for at least three - years, to give any third party, for a charge no more than your - cost of physically performing source distribution, a complete - machine-readable copy of the corresponding source code, to be - distributed under the terms of Sections 1 and 2 above on a medium - customarily used for software interchange; or, - - c) Accompany it with the information you received as to the offer - to distribute corresponding source code. (This alternative is - allowed only for noncommercial distribution and only if you - received the program in object code or executable form with such - an offer, in accord with Subsection b above.) - -The source code for a work means the preferred form of the work for -making modifications to it. For an executable work, complete source -code means all the source code for all modules it contains, plus any -associated interface definition files, plus the scripts used to -control compilation and installation of the executable. However, as a -special exception, the source code distributed need not include -anything that is normally distributed (in either source or binary -form) with the major components (compiler, kernel, and so on) of the -operating system on which the executable runs, unless that component -itself accompanies the executable. - -If distribution of executable or object code is made by offering -access to copy from a designated place, then offering equivalent -access to copy the source code from the same place counts as -distribution of the source code, even though third parties are not -compelled to copy the source along with the object code. - - 4. You may not copy, modify, sublicense, or distribute the Program -except as expressly provided under this License. Any attempt -otherwise to copy, modify, sublicense or distribute the Program is -void, and will automatically terminate your rights under this License. -However, parties who have received copies, or rights, from you under -this License will not have their licenses terminated so long as such -parties remain in full compliance. - - 5. You are not required to accept this License, since you have not -signed it. However, nothing else grants you permission to modify or -distribute the Program or its derivative works. These actions are -prohibited by law if you do not accept this License. Therefore, by -modifying or distributing the Program (or any work based on the -Program), you indicate your acceptance of this License to do so, and -all its terms and conditions for copying, distributing or modifying -the Program or works based on it. - - 6. Each time you redistribute the Program (or any work based on the -Program), the recipient automatically receives a license from the -original licensor to copy, distribute or modify the Program subject to -these terms and conditions. You may not impose any further -restrictions on the recipients' exercise of the rights granted herein. -You are not responsible for enforcing compliance by third parties to -this License. - - 7. If, as a consequence of a court judgment or allegation of patent -infringement or for any other reason (not limited to patent issues), -conditions are imposed on you (whether by court order, agreement or -otherwise) that contradict the conditions of this License, they do not -excuse you from the conditions of this License. If you cannot -distribute so as to satisfy simultaneously your obligations under this -License and any other pertinent obligations, then as a consequence you -may not distribute the Program at all. For example, if a patent -license would not permit royalty-free redistribution of the Program by -all those who receive copies directly or indirectly through you, then -the only way you could satisfy both it and this License would be to -refrain entirely from distribution of the Program. - -If any portion of this section is held invalid or unenforceable under -any particular circumstance, the balance of the section is intended to -apply and the section as a whole is intended to apply in other -circumstances. - -It is not the purpose of this section to induce you to infringe any -patents or other property right claims or to contest validity of any -such claims; this section has the sole purpose of protecting the -integrity of the free software distribution system, which is -implemented by public license practices. Many people have made -generous contributions to the wide range of software distributed -through that system in reliance on consistent application of that -system; it is up to the author/donor to decide if he or she is willing -to distribute software through any other system and a licensee cannot -impose that choice. - -This section is intended to make thoroughly clear what is believed to -be a consequence of the rest of this License. - - 8. If the distribution and/or use of the Program is restricted in -certain countries either by patents or by copyrighted interfaces, the -original copyright holder who places the Program under this License -may add an explicit geographical distribution limitation excluding -those countries, so that distribution is permitted only in or among -countries not thus excluded. In such case, this License incorporates -the limitation as if written in the body of this License. - - 9. The Free Software Foundation may publish revised and/or new versions -of the General Public License from time to time. Such new versions will -be similar in spirit to the present version, but may differ in detail to -address new problems or concerns. - -Each version is given a distinguishing version number. If the Program -specifies a version number of this License which applies to it and "any -later version", you have the option of following the terms and conditions -either of that version or of any later version published by the Free -Software Foundation. If the Program does not specify a version number of -this License, you may choose any version ever published by the Free Software -Foundation. - - 10. If you wish to incorporate parts of the Program into other free -programs whose distribution conditions are different, write to the author -to ask for permission. For software which is copyrighted by the Free -Software Foundation, write to the Free Software Foundation; we sometimes -make exceptions for this. Our decision will be guided by the two goals -of preserving the free status of all derivatives of our free software and -of promoting the sharing and reuse of software generally. - - NO WARRANTY - - 11. BECAUSE THE PROGRAM IS LICENSED FREE OF CHARGE, THERE IS NO WARRANTY -FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN -OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES -PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED -OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF -MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS -TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE -PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, -REPAIR OR CORRECTION. - - 12. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING -WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MAY MODIFY AND/OR -REDISTRIBUTE THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, -INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING -OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED -TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY -YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER -PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE -POSSIBILITY OF SUCH DAMAGES. - - END OF TERMS AND CONDITIONS - - How to Apply These Terms to Your New Programs - - If you develop a new program, and you want it to be of the greatest -possible use to the public, the best way to achieve this is to make it -free software which everyone can redistribute and change under these terms. - - To do so, attach the following notices to the program. It is safest -to attach them to the start of each source file to most effectively -convey the exclusion of warranty; and each file should have at least -the "copyright" line and a pointer to where the full notice is found. - - - Copyright (C) - - This program is free software; you can redistribute it and/or modify - it under the terms of the GNU General Public License as published by - the Free Software Foundation; either version 2 of the License, or - (at your option) any later version. - - This program is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the - GNU General Public License for more details. - - You should have received a copy of the GNU General Public License along - with this program; if not, write to the Free Software Foundation, Inc., - 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. - -Also add information on how to contact you by electronic and paper mail. - -If the program is interactive, make it output a short notice like this -when it starts in an interactive mode: - - Gnomovision version 69, Copyright (C) year name of author - Gnomovision comes with ABSOLUTELY NO WARRANTY; for details type `show w'. - This is free software, and you are welcome to redistribute it - under certain conditions; type `show c' for details. - -The hypothetical commands `show w' and `show c' should show the appropriate -parts of the General Public License. Of course, the commands you use may -be called something other than `show w' and `show c'; they could even be -mouse-clicks or menu items--whatever suits your program. - -You should also get your employer (if you work as a programmer) or your -school, if any, to sign a "copyright disclaimer" for the program, if -necessary. Here is a sample; alter the names: - - Yoyodyne, Inc., hereby disclaims all copyright interest in the program - `Gnomovision' (which makes passes at compilers) written by James Hacker. - - , 1 April 1989 - Ty Coon, President of Vice - -This General Public License does not permit incorporating your program into -proprietary programs. If your program is a subroutine library, you may -consider it more useful to permit linking proprietary applications with the -library. If this is what you want to do, use the GNU Lesser General -Public License instead of this License. diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/README.md b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/README.md deleted file mode 100644 index b41f01682..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/README.md +++ /dev/null @@ -1,246 +0,0 @@ -[![License: GPL v2](https://img.shields.io/badge/License-GPL%20v2-blue.svg)](https://www.gnu.org/licenses/old-licenses/gpl-2.0.en.html) -![Build Status](https://github.com/megastep/makeself/workflows/CI/badge.svg) - -# makeself - Make self-extractable archives on Unix - -[makeself.sh][1] is a small shell script that generates a self-extractable -compressed tar archive from a directory. The resulting file appears as a shell script -(many of those have a **.run** suffix), and can be launched as is. The archive -will then uncompress itself to a temporary directory and an optional arbitrary -command will be executed (for example an installation script). This is pretty -similar to archives generated with WinZip Self-Extractor in the Windows world. -Makeself archives also include checksums for integrity self-validation (CRC -and/or MD5/SHA256 checksums). - -The makeself.sh script itself is used only to create the archives from a -directory of files. The resultant archive is actually a compressed (using -gzip, bzip2, or compress) TAR archive, with a small shell script stub at the -beginning. This small stub performs all the steps of extracting the files, -running the embedded command, and removing the temporary files when done. -All the user has to do to install the software contained in such an -archive is to "run" the archive, i.e **sh nice-software.run**. I recommend -using the ".run" (which was introduced by some Makeself archives released by -Loki Software) or ".sh" suffix for such archives not to confuse the users, -so that they will know they are actually shell scripts (with quite a lot of binary data -attached to them though!). - -I am trying to keep the code of this script as portable as possible, i.e it is -not relying on any bash-specific features and only calls commands that are -installed on any functioning UNIX-compatible system. This script as well as -the archives it generates should run on any Unix flavor, with any compatible -Bourne shell, provided of course that the compression programs are available. - -As of version 2.1, Makeself has been rewritten and tested on the following -platforms : - - * Linux (all distributions) - * Sun Solaris (8 and above) - * HP-UX (tested on 11.0 and 11i on HPPA RISC) - * SCO OpenUnix and OpenServer - * IBM AIX 5.1L - * macOS (Darwin) - * SGI IRIX 6.5 - * FreeBSD - * UnicOS / Cray - * Cygwin (Windows) - -If you successfully run Makeself and/or archives created with it on another -system, then please [let me know][2]! - -Examples of publicly available archives made using makeself are : - - * Game patches and installers for [Id Software][3] games like Quake 3 for Linux or Return To Castle Wolfenstein ; - * All game patches released by [Loki Software][4] for the Linux version of popular games ; - * The [nVidia drivers][5] for Linux - * The installer for the Linux version of [Google Earth][6] - * The [VirtualBox][7] installers for Linux - * The [Makeself][1] distribution itself ;-) - * and countless others... - -**Important note for Apache users:** By default, most Web servers will think that Makeself archives are regular text files and thus they may show up as text in a Web browser. The correct way to prevent this is to add a MIME type for this file format, like so (in httpd.conf) : - -`AddType application/x-makeself .run` - -**Important note for certain GNU/Linux distributions:** Archives created with Makeself prior to v2.1.2 were using an old syntax for the _head_ and _tail_ Unix commands that is being progressively obsoleted in their GNU forms. Therefore you may have problems uncompressing some of these archives. A workaround for this is to set the environment variable $_POSIX2_VERSION to enable the old syntax, i.e. : - -`export _POSIX2_VERSION=199209` - -## Usage - -The syntax of makeself is the following: - -``` -makeself.sh [args] archive_dir file_name label startup_script [script_args] -``` - - * _args_ are optional options for Makeself. The available ones are : - - * **`--version`** : Prints the version number on stdout, then exits immediately - * **`--gzip`** : Use gzip for compression (the default on platforms on which gzip is commonly available, like Linux) - * **`--bzip2`** : Use bzip2 instead of gzip for better compression. The bzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. - * **`--pbzip2`** : Use pbzip2 instead of gzip for better and faster compression on machines having multiple CPUs. The pbzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. - * **`--xz`** : Use xz instead of gzip for better compression. The xz command must be available in the command path. It is recommended that the archive prefix be set to something like '.xz.run' for the archive, so that potential users know that they'll need xz to extract it. - * **`--lzo`** : Use lzop instead of gzip for better compression. The lzop command must be available in the command path. It is recommended that the archive prefix be set to something like `.lzo.run` for the archive, so that potential users know that they'll need lzop to extract it. - * **`--lz4`** : Use lz4 instead of gzip for better compression. The lz4 command must be available in the command path. It is recommended that the archive prefix be set to something like '.lz4.run' for the archive, so that potential users know that they'll need lz4 to extract it. - * **`--zstd`** : Use zstd instead of gzip for better compression. The zstd command must be available in the command path. It is recommended that the archive prefix be set to something like '.zstd.run' for the archive, so that potential users know that they'll need zstd to extract it. - * **`--pigz`** : Use pigz for compression. - * **`--base64`** : Encode the archive to ASCII in Base64 format instead of compressing (base64 command required). - * **`--gpg-encrypt`** : Encrypt the archive using `gpg -ac -z $COMPRESS_LEVEL`. This will prompt for a password to encrypt with. Assumes that potential users have `gpg` installed. - * **`--ssl-encrypt`** : Encrypt the archive using `openssl aes-256-cbc -a -salt`. This will prompt for a password to encrypt with. Assumes that the potential users have the OpenSSL tools installed. - * **`--compress`** : Use the UNIX `compress` command to compress the data. This should be the default on all platforms that don't have gzip available. - * **`--nocomp`** : Do not use any compression for the archive, which will then be an uncompressed TAR. - * **`--complevel`** : Specify the compression level for gzip, bzip2, pbzip2, zstd, xz, lzo or lz4. (defaults to 9) - * **`--threads`** : Specify the number of threads to be used by compressors that support parallelization. Omit to use compressor's default. Most useful (and required) for opting into xz's threading, usually with `--threads=0` for all available cores. pbzip2 and pigz are parallel by default, and setting this value allows limiting the number of threads they use. - * **`--notemp`** : The generated archive will not extract the files to a temporary directory, but in a new directory created in the current directory. This is better to distribute software packages that may extract and compile by themselves (i.e. launch the compilation through the embedded script). - * **`--current`** : Files will be extracted to the current directory, instead of in a subdirectory. This option implies `--notemp` above. - * **`--follow`** : Follow the symbolic links inside of the archive directory, i.e. store the files that are being pointed to instead of the links themselves. - * **`--append`** _(new in 2.1.x)_: Append data to an existing archive, instead of creating a new one. In this mode, the settings from the original archive are reused (compression type, label, embedded script), and thus don't need to be specified again on the command line. - * **`--header`** : Makeself uses a separate file to store the header stub, called `makeself-header.sh`. By default, it is assumed that it is stored in the same location as makeself.sh. This option can be used to specify its actual location if it is stored someplace else. - * **`--cleanup`** : Specify a script that is run when execution is interrupted or finishes successfully. The script is executed with the same environment and initial `script_args` as `startup_script`. - * **`--copy`** : Upon extraction, the archive will first extract itself to a temporary directory. The main application of this is to allow self-contained installers stored in a Makeself archive on a CD, when the installer program will later need to unmount the CD and allow a new one to be inserted. This prevents "Filesystem busy" errors for installers that span multiple CDs. - * **`--nox11`** : Disable the automatic spawning of a new terminal in X11. - * **`--nowait`** : When executed from a new X11 terminal, disable the user prompt at the end of the script execution. - * **`--nomd5`** and **`--nocrc`** : Disable the creation of a MD5 / CRC checksum for the archive. This speeds up the extraction process if integrity checking is not necessary. - * **`--sha256`** : Adds a SHA256 checksum for the archive. This is in addition to the MD5 / CRC checksums unless `--nomd5` is also used. - * **`--lsm` _file_** : Provide and LSM file to makeself, that will be embedded in the generated archive. LSM files are describing a software package in a way that is easily parseable. The LSM entry can then be later retrieved using the `--lsm` argument to the archive. An example of a LSM file is provided with Makeself. - * **`--tar-format opt`** : Specify the tar archive format (default is ustar); you may use any value accepted by your tar command (such as posix, v7, etc). - * **`--tar-extra opt`** : Append more options to the tar command line. - - For instance, in order to exclude the `.git` directory from the packaged archive directory using the GNU `tar`, one can use `makeself.sh --tar-extra "--exclude=.git" ...` - - * **`--keep-umask`** : Keep the umask set to shell default, rather than overriding when executing self-extracting archive. - * **`--packaging-date date`** : Use provided string as the packaging date instead of the current date. - * **`--license`** : Append a license file. - * **`--nooverwrite`** : Do not extract the archive if the specified target directory already exists. - * **`--help-header file`** : Add a header to the archive's `--help` output. - * `archive_dir` is the name of the directory that contains the files to be archived - * `file_name` is the name of the archive to be created - * `label` is an arbitrary text string describing the package. It will be displayed while extracting the files. - * `startup_script` is the command to be executed _from within_ the directory of extracted files. Thus, if you wish to execute a program contained in this directory, you must prefix your command with `./`. For example, `./program` will be fine. The `script_args` are additional arguments for this command. - -Here is an example, assuming the user has a package image stored in a **/home/joe/mysoft**, and he wants to generate a self-extracting package named -**mysoft.sh**, which will launch the "setup" script initially stored in /home/joe/mysoft : - -`makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup -` - -Here is also how I created the [makeself.run][9] archive which contains the Makeself distribution : - -`makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" ` - -Archives generated with Makeself can be passed the following arguments: - - * **`--keep`** : Prevent the files to be extracted in a temporary directory that will be removed after the embedded script's execution. The files will then be extracted in the current working directory and will stay here until you remove them. - * **`--verbose`** : Will prompt the user before executing the embedded command - * **`--target dir`** : Allows to extract the archive in an arbitrary place. - * **`--nox11`** : Do not spawn a X11 terminal. - * **`--confirm`** : Prompt the user for confirmation before running the embedded command. - * **`--info`** : Print out general information about the archive (does not extract). - * **`--lsm`** : Print out the LSM entry, if it is present. - * **`--list`** : List the files in the archive. - * **`--check`** : Check the archive for integrity using the embedded checksums. Does not extract the archive. - * **`--nochown`** : By default, a `chown -R` command is run on the target directory after extraction, so that all files belong to the current user. This is mostly needed if you are running as root, as tar will then try to recreate the initial user ownerships. You may disable this behavior with this flag. - * **`--tar`** : Run the tar command on the contents of the archive, using the following arguments as parameter for the command. - * **`--noexec`** : Do not run the embedded script after extraction. - * **`--noexec-cleanup`** : Do not run the embedded cleanup script. - * **`--nodiskspace`** : Do not check for available disk space before attempting to extract. - * **`--cleanup-args`** : Specify arguments to be passed to the cleanup script. Wrap value in quotes to specify multiple arguments. - -Any subsequent arguments to the archive will be passed as additional arguments to the embedded command. You must explicitly use the `--` special command-line construct before any such options to make sure that Makeself will not try to interpret them. - -## Startup Script - -The startup script must be a regular Shell script. - -Within the startup script, you can use the `$USER_PWD` variable to get the path of the folder from which the self-extracting script is executed. This is especially useful to access files that are located in the same folder as the script, as shown in the example below. - -`my-self-extracting-script.sh --fooBarFileParameter foo.bar` - -## Building and Testing - -Clone the git repo and execute `git submodule update --init --recursive` to obtain all submodules. - -* To make a release: `make` -* To run all tests: `make test` - -## Maven Usage - -Makeself is now supported by the following maven plugin [makeself-maven-plugin](https://github.com/hazendaz/makeself-maven-plugin). Please refer to project for usage and report any bugs in regards to maven plugin on that project. - -## License - -Makeself itself is covered by the [GNU General Public License][8] (GPL) version 2 and above. Archives generated by Makeself don't have to be placed under this license (although I encourage it ;-)), since the archive itself is merely data for Makeself. - -## Contributing - -I will gladly consider merging your pull requests on the [GitHub][10] repository. However, please keep the following in mind: - - * One of the main purposes of Makeself is portability. Do not submit patches that will break supported platforms. The more platform-agnostic, the better. - * Please explain clearly what the purpose of the patch is, and how you achieved it. - -## Download - -Get the latest official distribution [here][9] (version 2.4.2). - -The latest development version can be grabbed from [GitHub][10]. Feel free to submit any patches there through the fork and pull request process. - -## Version history - - * **v1.0:** Initial public release - * **v1.1:** The archive can be passed parameters that will be passed on to the embedded script, thanks to John C. Quillan - * **v1.2:** Cosmetic updates, support for bzip2 compression and non-temporary archives. Many ideas thanks to Francois Petitjean. - * **v1.3:** More patches from Bjarni R. Einarsson and Francois Petitjean: Support for no compression (`--nocomp`), script is no longer mandatory, automatic launch in an xterm, optional verbose output, and -target archive option to indicate where to extract the files. - * **v1.4:** Many patches from Francois Petitjean: improved UNIX compatibility, automatic integrity checking, support of LSM files to get info on the package at run time.. - * **v1.5.x:** A lot of bugfixes, and many other patches, including automatic verification through the usage of checksums. Version 1.5.5 was the stable release for a long time, even though the Web page didn't get updated ;-). Makeself was also officially made a part of the [Loki Setup installer][11], and its source is being maintained as part of this package. - * **v2.0:** Complete internal rewrite of Makeself. The command-line parsing was vastly improved, the overall maintenance of the package was greatly improved by separating the stub from makeself.sh. Also Makeself was ported and tested to a variety of Unix platforms. - * **v2.0.1:** First public release of the new 2.0 branch. Prior versions are officially obsoleted. This release introduced the `--copy` argument that was introduced in response to a need for the [UT2K3][12] Linux installer. - * **v2.1.0:** Big change : Makeself can now support multiple embedded tarballs, each stored separately with their own checksums. An existing archive can be updated with the `--append` flag. Checksums are also better managed, and the `--nochown` option for archives appeared. - * **v2.1.1:** Fixes related to the Unix compression (compress command). Some Linux distributions made the insane choice to make it unavailable, even though gzip is capable of uncompressing these files, plus some more bugfixes in the extraction and checksum code. - * **v2.1.2:** Some bug fixes. Use head -n to avoid problems with POSIX conformance. - * **v2.1.3:** Bug fixes with the command line when spawning terminals. Added `--tar`, `--noexec` for archives. Added `--nomd5` and `--nocrc` to avoid creating checksums in archives. The embedded script is now run through "eval". The `--info` output now includes the command used to create the archive. A man page was contributed by Bartosz Fenski. - * **v2.1.4:** Fixed `--info` output. Generate random directory name when extracting files to . to avoid problems. Better handling of errors with wrong permissions for the directory containing the files. Avoid some race conditions, Unset the $CDPATH variable to avoid problems if it is set. Better handling of dot files in the archive directory. - * **v2.1.5:** Made the md5sum detection consistent with the header code. Check for the presence of the archive directory. Added `--encrypt` for symmetric encryption through gpg (Eric Windisch). Added support for the digest command on Solaris 10 for MD5 checksums. Check for available disk space before extracting to the target directory (Andreas Schweitzer). Allow extraction to run asynchronously (patch by Peter Hatch). Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo). - * **v2.1.6:** Replaced one dot per file progress with a realtime progress percentage and a spinning cursor. Added `--noprogress` to prevent showing the progress during the decompression. Added `--target` dir to allow extracting directly to a target directory. (Guy Baconniere) - * **v2.2.0:** First major new release in years! Includes many bugfixes and user contributions. Please look at the [project page on Github][10] for all the details. - * **v2.3.0:** Support for archive encryption via GPG or OpenSSL. Added LZO and LZ4 compression support. Options to set the packaging date and stop the umask from being overriden. Optionally ignore check for available disk space when extracting. New option to check for root permissions before extracting. - * **v2.3.1:** Various compatibility updates. Added unit tests for Travis CI in the GitHub repo. New `--tar-extra`, `--untar-extra`, `--gpg-extra`, `--gpg-asymmetric-encrypt-sign` options. - * **v2.4.0:** Added optional support for SHA256 archive integrity checksums. - * **v2.4.2:** New --cleanup and --cleanup-args arguments for cleanup scripts. Added threading support for supported compressors. Now supports zstd compression. - * **v2.4.3:** Make explicit POSIX tar archives for increased compatibility. - * **v2.4.4:** Fixed various compatibility issues (no longer use POSIX tar archives), Github Actions to check on Solaris and FreeBSD. - * **v2.4.5:** Added `--tar-format` option to set the tar archive format (default is ustar) - -## Links - - * Check out the ["Loki Setup"][11] installer, used to install many Linux games and other applications, and of which I am the co-author. Since the demise of Loki, I am now the official maintainer of the project, and it is now being hosted here on GitHub. - * Bjarni R. Einarsson also wrote the **setup.sh** installer script, inspired by Makeself. [Check it out !][14] - -## Contact - -This script was written by [Stéphane Peter][15] (megastep at megastep.org). Any enhancements and suggestions are welcome. - -Contributions were included from John C. Quillan, Bjarni R. Einarsson, -Francois Petitjean, Ryan C. Gordon, and many contributors on GitHub. If you think I forgot -your name, don't hesitate to contact me. - -This project is now hosted on GitHub. Feel free to submit patches and bug reports on the [project page][10]. - -* * * - -[Stephane Peter][2] - - [1]: http://makeself.io/ - [2]: mailto:megastep@megastep.org - [3]: http://www.idsoftware.com/ - [4]: http://www.lokigames.com/products/myth2/updates.php3 - [5]: http://www.nvidia.com/ - [6]: http://earth.google.com/ - [7]: http://www.virtualbox.org/ - [8]: http://www.gnu.org/copyleft/gpl.html - [9]: https://github.com/megastep/makeself/releases/download/release-2.4.5/makeself-2.4.5.run - [10]: https://github.com/megastep/makeself - [11]: https://github.com/megastep/loki_setup/ - [12]: http://www.unrealtournament2003.com/ - [13]: http://www.icculus.org/ - [14]: http://bre.klaki.net/programs/setup.sh/ - [15]: https://stephanepeter.com/ diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/VERSION b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/VERSION deleted file mode 100644 index 59aa62c1f..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/VERSION +++ /dev/null @@ -1 +0,0 @@ -2.4.5 diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/make-release.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/make-release.sh deleted file mode 100644 index b5692d490..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/make-release.sh +++ /dev/null @@ -1,9 +0,0 @@ -#!/bin/sh -# -# Create a distributable archive of the current version of Makeself - -VER=`cat VERSION` -mkdir -p /tmp/makeself-$VER release -cp -pPR makeself* test README.md COPYING VERSION .gitmodules /tmp/makeself-$VER/ -./makeself.sh --notemp /tmp/makeself-$VER release/makeself-$VER.run "Makeself v$VER" echo "Makeself has extracted itself" - diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself-header.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself-header.sh deleted file mode 100644 index 940903148..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself-header.sh +++ /dev/null @@ -1,660 +0,0 @@ -cat << EOF > "$archname" -#!/bin/bash -# This script was generated using Makeself $MS_VERSION -# The license covering this archive and its contents, if any, is wholly independent of the Makeself license (GPL) -# 2022.3.19-Modified the MS_Help function and some options -# Huawei Technologies Co., Ltd. - -ORIG_UMASK=\`umask\` - -CRCsum="$CRCsum" -MD5="$MD5sum" -SHA="$SHAsum" -SIGNATURE="$Signature" -TMPROOT=\${TMPDIR:="\$HOME"} -if ! test -d "\$TMPROOT"; then - TMPROOT="\$PWD" -fi -export TMPDIR="\$TMPROOT" -USER_PWD="\$PWD" -if ! test -d "\$USER_PWD"; then - exit 1 -fi -export USER_PWD -ARCHIVE_DIR=\`dirname "\$0"\` -export ARCHIVE_DIR - -name_of_file="\$0 " -pwd_of_file="\$PWD" -label="$LABEL" -script="$SCRIPT" -scriptargs="$SCRIPTARGS" -cleanup_script="${CLEANUP_SCRIPT}" -licensetxt="$LICENSE" -helpheader='$HELPHEADER' -targetdir="$archdirname" -filesizes="$filesizes" -totalsize="$totalsize" -keep="$KEEP" -nooverwrite="$NOOVERWRITE" -quiet="n" -accept="n" -nodiskspace="n" -export_conf="$EXPORT_CONF" -decrypt_cmd="$DECRYPT_CMD" -skip="$SKIP" - -print_cmd_arg="" -if type printf > /dev/null; then - print_cmd="printf" -elif test -x /usr/ucb/echo; then - print_cmd="/usr/ucb/echo" -else - print_cmd="echo" -fi - -if test -d /usr/xpg4/bin; then - PATH=/usr/xpg4/bin:\$PATH - export PATH -fi - -if test -d /usr/sfw/bin; then - PATH=\$PATH:/usr/sfw/bin - export PATH -fi - -unset CDPATH - -MS_Printf() -{ - \$print_cmd \$print_cmd_arg "\$1" -} - -MS_PrintLicense() -{ - PAGER=\${PAGER:=more} - if test x"\$licensetxt" != x; then - PAGER_PATH=\`exec <&- 2>&-; which \$PAGER || command -v \$PAGER || type \$PAGER\` - if test -x "\$PAGER_PATH"; then - echo "\$licensetxt" | \$PAGER - else - echo "\$licensetxt" - fi - if test x"\$accept" != xy; then - while true - do - MS_Printf "Please type y to accept, n otherwise: " - read yn - if test x"\$yn" = xn; then - keep=n - eval \$finish; exit 1 - break; - elif test x"\$yn" = xy; then - break; - fi - done - fi - fi -} - -MS_diskspace() -{ - ( - df -kP "\$1" | tail -1 | awk '{ if (\$4 ~ /%/) {print \$3} else {print \$4} }' - ) -} - -MS_dd() -{ - blocks=\`expr \$3 / 1024\` - bytes=\`expr \$3 % 1024\` - # Test for ibs, obs and conv feature - if dd if=/dev/zero of=/dev/null count=1 ibs=512 obs=512 conv=sync 2> /dev/null; then - dd if="\$1" ibs=\$2 skip=1 obs=1024 conv=sync 2> /dev/null | \\ - { test \$blocks -gt 0 && dd ibs=1024 obs=1024 count=\$blocks ; \\ - test \$bytes -gt 0 && dd ibs=1 obs=1024 count=\$bytes ; } 2> /dev/null - else - dd if="\$1" bs=\$2 skip=1 2> /dev/null - fi -} - -MS_dd_Progress() -{ - if test x"\$noprogress" = xy; then - MS_dd "\$@" - return \$? - fi - file="\$1" - offset=\$2 - length=\$3 - pos=0 - bsize=4194304 - while test \$bsize -gt \$length; do - bsize=\`expr \$bsize / 4\` - done - blocks=\`expr \$length / \$bsize\` - bytes=\`expr \$length % \$bsize\` - ( - dd ibs=\$offset skip=1 2>/dev/null - pos=\`expr \$pos \+ \$bsize\` - MS_Printf " 0%% " 1>&2 - if test \$blocks -gt 0; then - while test \$pos -le \$length; do - dd bs=\$bsize count=1 2>/dev/null - pcent=\`expr \$length / 100\` - pcent=\`expr \$pos / \$pcent\` - if test \$pcent -lt 100; then - MS_Printf "\b\b\b\b\b\b\b" 1>&2 - if test \$pcent -lt 10; then - MS_Printf " \$pcent%% " 1>&2 - else - MS_Printf " \$pcent%% " 1>&2 - fi - fi - pos=\`expr \$pos \+ \$bsize\` - done - fi - if test \$bytes -gt 0; then - dd bs=\$bytes count=1 2>/dev/null - fi - MS_Printf "\b\b\b\b\b\b\b" 1>&2 - MS_Printf " 100%% " 1>&2 - ) < "\$file" -} - -MS_Help() -{ - cat << EOH >&2 -Usage: \$0 [options] -Options: - --help | -h Print this message - --info Print embedded info : title, default target directory, embedded script ... - --list Print the list of files in the archive - --check Checks integrity and version dependency of the archive - --quiet Quiet install mode, skip human-computer interactions - --nox11 Do not spawn an xterm - --noexec Do not run embedded script - --extract= Extract directly to a target directory (absolute or relative) - Usually used with --noexec to just extract files without running - --tar arg1 [arg2 ...] Access the contents of the archive through the tar command -\${helpheader} -EOH -} - -MS_Verify_Sig() -{ - GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` - MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` - test -x "\$GPG_PATH" || GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` - test -x "\$MKTEMP_PATH" || MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` - offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` - temp_sig=\`mktemp -t XXXXX\` - echo \$SIGNATURE | base64 --decode > "\$temp_sig" - gpg_output=\`MS_dd "\$1" \$offset \$totalsize | LC_ALL=C "\$GPG_PATH" --verify "\$temp_sig" - 2>&1\` - gpg_res=\$? - rm -f "\$temp_sig" - if test \$gpg_res -eq 0 && test \`echo \$gpg_output | grep -c Good\` -eq 1; then - if test \`echo \$gpg_output | grep -c \$sig_key\` -eq 1; then - test x"\$quiet" = xn && echo "GPG signature is good" >&2 - else - echo "GPG Signature key does not match" >&2 - exit 2 - fi - else - test x"\$quiet" = xn && echo "GPG signature failed to verify" >&2 - exit 2 - fi -} - -MS_Check() -{ - OLD_PATH="\$PATH" - PATH=\${GUESS_MD5_PATH:-"\$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} - MD5_ARG="" - MD5_PATH=\`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum\` - test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which md5 || command -v md5 || type md5\` - test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which digest || command -v digest || type digest\` - PATH="\$OLD_PATH" - - SHA_PATH=\`exec <&- 2>&-; which shasum || command -v shasum || type shasum\` - test -x "\$SHA_PATH" || SHA_PATH=\`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum\` - - if test x"\$quiet" = xn; then - MS_Printf "Verifying archive integrity..." - fi - offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` - fsize=\`cat "\$1" | wc -c | tr -d " "\` - if test \$totalsize -ne \`expr \$fsize - \$offset\`; then - echo " Unexpected archive size." >&2 - exit 2 - fi - verb=\$2 - i=1 - for s in \$filesizes - do - crc=\`echo \$CRCsum | cut -d" " -f\$i\` - if test -x "\$SHA_PATH"; then - if test x"\`basename \$SHA_PATH\`" = xshasum; then - SHA_ARG="-a 256" - fi - sha=\`echo \$SHA | cut -d" " -f\$i\` - if test x"\$sha" = x0000000000000000000000000000000000000000000000000000000000000000; then - test x"\$verb" = xy && echo " \$1 does not contain an embedded SHA256 checksum." >&2 - else - shasum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$SHA_PATH \$SHA_ARG" | cut -b-64\`; - if test x"\$shasum" != x"\$sha"; then - echo "Error in SHA256 checksums: \$shasum is different from \$sha" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " SHA256 checksums are OK." >&2 - fi - crc="0000000000"; - fi - fi - if test -x "\$MD5_PATH"; then - if test x"\`basename \$MD5_PATH\`" = xdigest; then - MD5_ARG="-a md5" - fi - md5=\`echo \$MD5 | cut -d" " -f\$i\` - if test x"\$md5" = x00000000000000000000000000000000; then - test x"\$verb" = xy && echo " \$1 does not contain an embedded MD5 checksum." >&2 - else - md5sum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$MD5_PATH \$MD5_ARG" | cut -b-32\`; - if test x"\$md5sum" != x"\$md5"; then - echo "Error in MD5 checksums: \$md5sum is different from \$md5" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " MD5 checksums are OK." >&2 - fi - crc="0000000000"; verb=n - fi - fi - if test x"\$crc" = x0000000000; then - test x"\$verb" = xy && echo " \$1 does not contain a CRC checksum." >&2 - else - sum1=\`MS_dd_Progress "\$1" \$offset \$s | CMD_ENV=xpg4 cksum | awk '{print \$1}'\` - if test x"\$sum1" != x"\$crc"; then - echo "Error in checksums: \$sum1 is different from \$crc" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " CRC checksums are OK." >&2 - fi - fi - i=\`expr \$i + 1\` - offset=\`expr \$offset + \$s\` - done - if test x"\$quiet" = xn; then - echo " All good." - fi -} - -MS_Decompress() -{ - if test x"\$decrypt_cmd" != x""; then - { eval "\$decrypt_cmd" || echo " ... Decryption failed." >&2; } | eval "$GUNZIP_CMD" - else - eval "$GUNZIP_CMD" - fi - - if test \$? -ne 0; then - echo " ... Decompression failed." >&2 - fi -} - -UnTAR() -{ - if test x"\$quiet" = xn; then - tar \$1vf - $UNTAR_EXTRA 2>&1 || { echo " ... Extraction failed." >&2; kill -15 \$$; } - else - tar \$1f - $UNTAR_EXTRA 2>&1 || { echo Extraction failed. >&2; kill -15 \$$; } - fi -} - -MS_exec_cleanup() { - if test x"\$cleanup" = xy && test x"\$cleanup_script" != x""; then - cleanup=n - cd "\$tmpdir" - eval "\"\$cleanup_script\" \$scriptargs \$cleanupargs" - fi -} - -MS_cleanup() -{ - echo 'Signal caught, cleaning up' >&2 - MS_exec_cleanup - cd "\$TMPROOT" - rm -rf "\$tmpdir" - eval \$finish; exit 15 -} - -Script_Args_Check() -{ - script_supported_args=\$(echo \${helpheader} | grep -o -E "\-\-[^ ]+" | awk -F"=" {'print \$1'}) - arg_to_test=\$(echo \$1|awk -F"=" {'print \$1'}) - - for arg in \${script_supported_args}; - do - if test x"\$arg_to_test" = x"\$arg" ;then - return - fi - done - - MS_Help - exit 1 -} - -finish=true -xterm_loop= -noprogress=$NOPROGRESS -nox11=$NOX11 -copy=$COPY -ownership=$OWNERSHIP -verbose=n -cleanup=y -cleanupargs= -sig_key= - -initargs="\$@" - -while [ -n "\$*" ] -do - case "\$1" in - -h | --help) - MS_Help - exit 0 - ;; - -q | --quiet) - quiet=y - noprogress=y - shift - ;; - --info) - echo Identification: "\$label" - echo Target directory: "\$targetdir" - echo Uncompressed size: $USIZE KB - echo Compression: $COMPRESS - if test x"$ENCRYPT" != x""; then - echo Encryption: $ENCRYPT - fi - echo Date of packaging: $DATE - echo Built with Makeself version $MS_VERSION - echo Build command was: "$MS_COMMAND" - if test x"\$script" != x; then - echo Script run after extraction: - echo " " \$script \$scriptargs - fi - if test x"$copy" = xcopy; then - echo "Archive will copy itself to a temporary location" - fi - if test x"$NEED_ROOT" = xy; then - echo "Root permissions required for extraction" - fi - if test x"$KEEP" = xy; then - echo "directory \$targetdir is permanent" - else - echo "\$targetdir will be removed after extraction" - fi - exit 0 - ;; - --list) - echo Target directory: \$targetdir - offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - for s in \$filesizes - do - MS_dd "\$0" \$offset \$s | MS_Decompress | UnTAR t - offset=\`expr \$offset + \$s\` - done - exit 0 - ;; - --tar) - offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - arg1="\$2" - shift 2 || { MS_Help; exit 1; } - for s in \$filesizes - do - MS_dd "\$0" \$offset \$s | MS_Decompress | tar "\$arg1" - "\$@" - offset=\`expr \$offset + \$s\` - done - exit 0 - ;; - --check) - MS_Check "\$0" y - scriptargs="\$scriptargs \$1" - shift - ;; - --noexec) - script="" - cleanup_script="" - shift - ;; - --extract=*) - keep=y - targetdir=\`echo \$1 | cut -d"=" -f2 \` - if ! shift; then MS_Help; exit 1; fi - ;; - --nox11) - nox11=y - shift - ;; - --xwin) - if test "$NOWAIT" = n; then - finish="echo Press Return to close this window...; read junk" - fi - xterm_loop=1 - shift - ;; - --phase2) - copy=phase2 - shift - ;; - --repack | --repack-path=*) - Script_Args_Check \$1 - scriptargs="\$scriptargs '\$1'" - shift - if [[ ! "\$1" =~ ^-.* ]]; then - scriptargs="\$scriptargs '\$1'" - shift - fi - ;; - *) - Script_Args_Check \$1 - scriptargs="\$scriptargs '\$1'" - shift - ;; - esac -done - -quiet_para="" -if test x"\$quiet" = xy; then - quiet_para="--quiet " -fi -scriptargs="--\$name_of_file""--\"\$pwd_of_file\""" \$quiet_para""\$scriptargs" - -if test x"\$quiet" = xy -a x"\$verbose" = xy; then - echo Cannot be verbose and quiet at the same time. >&2 - exit 1 -fi - -if test x"$NEED_ROOT" = xy -a \`id -u\` -ne 0; then - echo "Administrative privileges required for this archive (use su or sudo)" >&2 - exit 1 -fi - -if test x"\$copy" \!= xphase2; then - MS_PrintLicense -fi - -case "\$copy" in -copy) - tmpdir="\$TMPROOT"/makeself.\$RANDOM.\`date +"%y%m%d%H%M%S"\`.\$\$ - mkdir "\$tmpdir" || { - echo "Could not create temporary directory \$tmpdir" >&2 - exit 1 - } - SCRIPT_COPY="\$tmpdir/makeself" - echo "Copying to a temporary location..." >&2 - cp "\$0" "\$SCRIPT_COPY" - chmod +x "\$SCRIPT_COPY" - cd "\$TMPROOT" - exec "\$SCRIPT_COPY" --phase2 -- \$initargs - ;; -phase2) - finish="\$finish ; rm -rf \`dirname \$0\`" - ;; -esac - -if test x"\$nox11" = xn; then - if tty -s; then # Do we have a terminal? - : - else - if test x"\$DISPLAY" != x -a x"\$xterm_loop" = x; then # No, but do we have X? - if xset q > /dev/null 2>&1; then # Check for valid DISPLAY variable - GUESS_XTERMS="xterm gnome-terminal rxvt dtterm eterm Eterm xfce4-terminal lxterminal kvt konsole aterm terminology" - for a in \$GUESS_XTERMS; do - if type \$a >/dev/null 2>&1; then - XTERM=\$a - break - fi - done - chmod a+x \$0 || echo Please add execution rights on \$0 - if test \`echo "\$0" | cut -c1\` = "/"; then # Spawn a terminal! - exec \$XTERM -e "\$0 --xwin \$initargs" - else - exec \$XTERM -e "./\$0 --xwin \$initargs" - fi - fi - fi - fi -fi - -if test x"\$targetdir" = x.; then - tmpdir="." -else - if test x"\$keep" = xy; then - if test x"\$nooverwrite" = xy && test -d "\$targetdir"; then - echo "Target directory \$targetdir already exists, aborting." >&2 - exit 1 - fi - if test x"\$quiet" = xn; then - echo "Creating directory \$targetdir" >&2 - fi - tmpdir="\$targetdir" - dashp="-p" - else - tmpdir="\$TMPROOT/selfgz\$\$\$RANDOM" - dashp="" - fi - mkdir \$dashp "\$tmpdir" || { - echo 'Cannot create target directory' \$tmpdir >&2 - echo 'You should try option --extract=' >&2 - eval \$finish - exit 1 - } -fi - -location="\`pwd\`" -if test x"\$SETUP_NOCHECK" != x1; then - MS_Check "\$0" -fi -offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - -if test x"\$verbose" = xy; then - MS_Printf "About to extract $USIZE KB in \$tmpdir ... Proceed ? [Y/n] " - read yn - if test x"\$yn" = xn; then - eval \$finish; exit 1 - fi -fi - -if test x"\$quiet" = xn; then - # Decrypting with openssl will ask for password, - # the prompt needs to start on new line - if test x"$ENCRYPT" = x"openssl"; then - echo "Decrypting and uncompressing \$label..." - else - MS_Printf "Uncompressing \$label" - fi -fi -res=3 -if test x"\$keep" = xn; then - trap MS_cleanup 1 2 3 15 -fi - -if test x"\$nodiskspace" = xn; then - leftspace=\`MS_diskspace "\$tmpdir"\` - if test -n "\$leftspace"; then - if test "\$leftspace" -lt $USIZE; then - echo - echo "Not enough space left in "\`dirname \$tmpdir\`" (\$leftspace KB) to decompress \$0 ($USIZE KB)" >&2 - if test x"\$keep" = xn; then - echo "Consider setting TMPDIR to a directory with more free space." - fi - eval \$finish; exit 1 - fi - fi -fi - -for s in \$filesizes -do - if MS_dd_Progress "\$0" \$offset \$s | MS_Decompress | ( cd "\$tmpdir"; umask \$ORIG_UMASK ; UnTAR xp ) 1>/dev/null; then - if test x"\$ownership" = xy; then - (cd "\$tmpdir"; chown -R \`id -u\` .; chgrp -R \`id -g\` .) - fi - else - echo >&2 - echo "Unable to decompress \$0" >&2 - eval \$finish; exit 1 - fi - offset=\`expr \$offset + \$s\` -done -if test x"\$quiet" = xn; then - echo -fi - -cd "\$tmpdir" -res=0 -if test x"\$script" != x; then - if test x"\$export_conf" = x"y"; then - MS_BUNDLE="\$0" - MS_LABEL="\$label" - MS_SCRIPT="\$script" - MS_SCRIPTARGS="\$scriptargs" - MS_ARCHDIRNAME="\$archdirname" - MS_KEEP="\$KEEP" - MS_NOOVERWRITE="\$NOOVERWRITE" - MS_COMPRESS="\$COMPRESS" - MS_CLEANUP="\$cleanup" - export MS_BUNDLE MS_LABEL MS_SCRIPT MS_SCRIPTARGS - export MS_ARCHDIRNAME MS_KEEP MS_NOOVERWRITE MS_COMPRESS - fi - - if test x"\$verbose" = x"y"; then - yn="x" - while test x"\$yn" != x -a x"\$yn" != xy -a x"\$yn" != xY -a x"\$yn" != xn -a x"\$yn" != xN - do - MS_Printf "OK to execute: \$script \$scriptargs \$* ? [Y/n] " - read yn - if test x"\$yn" = x -o x"\$yn" = xy -o x"\$yn" = xY; then - eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$?; - elif test x"\$yn" = xn -o x"\$yn" = xN; then - echo "Unable to decompress \$script ,because of aborting! ";res=\$? - else - echo "Input value is unacceptable,please try again." - fi - done - else - eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$? - fi - if test "\$res" -ne 0; then - test x"\$verbose" = xy && echo "The program '\$script' returned an error code (\$res)" >&2 - fi -fi - -MS_exec_cleanup - -if test x"\$keep" = xn; then - cd "\$TMPROOT" - rm -rf "\$tmpdir" -fi -eval \$finish; exit \$res -EOF diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.1 b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.1 deleted file mode 100644 index 81bf6e4ff..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.1 +++ /dev/null @@ -1,110 +0,0 @@ -.TH "MAKESELF" "1" "2.4.5" -.SH "NAME" -makeself \- An utility to generate self-extractable archives. -.SH "SYNTAX" -.B makeself [\fIoptions\fP] archive_dir file_name label -.B [\fIstartup_script\fP] [\fIargs\fP] -.SH "DESCRIPTION" -This program is a free (GPL) utility designed to create self-extractable -archives from a directory. -.SH "OPTIONS" -The following options are supported. -.TP 15 -.B -v, --version -Prints out the makeself version number and exits. -.TP -.B -h, --help -Print out help information. -.TP -.B --tar-quietly -Suppress verbose output from the tar command -.TP -.B --quiet -Do not print any messages other than errors -.TP -.B --gzip -Compress using gzip (default if detected). -.TP -.B --bzip2 -Compress using bzip2. -.TP -.B --pbzip2 -Compress using pbzip2. -.TP -.B --xz -Compress using xz. -.TP -.B --lzo -Compress using lzop. -.TP -.B --lz4 -Compress using lz4. -.TP -.B --compress -Compress using the UNIX 'compress' command. -.TP -.B --nocomp -Do not compress the data. -.TP -.B --complevel lvl -Specify the compression level for gzip,bzip2,pbzui2,xz,lzo or lz4 -.TP -.B --notemp -The archive will create archive_dir in the current directory and -uncompress in ./archive_dir. -.TP -.B --copy -Upon extraction, the archive will first copy itself to a temporary directory. -.TP -.B --append -Append more files to an existing makeself archive. The label and startup scripts will then be ignored. -.TP -.B --current -Files will be extracted to the current directory. Both --current and --target dir imply --notemp. -.TP -.B --target dir -Extract directly to a target directory. Directory path can be either absolute or relative. -.TP -.B --header file -Specify location of the header script. -.TP -.B --cleanup file -Specify a cleanup script that executes on interrupt and when finished successfully. -.TP -.B --follow -Follow the symlinks in the archive. -.TP -.B --noprogress -Do not show the progress during the decompression. -.TP -.B --nox11 -Disable automatic spawn of an xterm if running in X11. -.TP -.B --nowait -Do not wait for user input after executing embedded program from an xterm. -.TP -.B --nomd5 -Do not create a MD5 checksum for the archive. -.TP -.B --nocrc -Do not create a CRC32 checksum for the archive. -.TP -.B --lsm file -LSM file describing the package. -.B --packaging-date date -Use provided string as the packaging date instead of the current date. -.SH "EXAMPLES" -Here is an example, assuming the user has a package image stored in a /home/joe/mysoft, -and he wants to generate a self-extracting package named mysoft.sh, which will launch -the "setup" script initially stored in /home/joe/mysoft: -.TP -makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup -.TP -Here is also how I created the makeself.run archive which contains the Makeself distribution: -.TP -makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" -.SH "AUTHORS" -Makeself has been written by Stéphane Peter . -.BR -This man page was originally written by Bartosz Fenski for the -Debian GNU/Linux distribution (but it may be used by others). diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.lsm b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.lsm deleted file mode 100644 index 3c4cea8c1..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.lsm +++ /dev/null @@ -1,16 +0,0 @@ -Begin3 -Title: makeself.sh -Version: 2.4.5 -Description: makeself.sh is a shell script that generates a self-extractable - tar.gz archive from a directory. The resulting file appears as a shell - script, and can be launched as is. The archive will then uncompress - itself to a temporary directory and an arbitrary command will be - executed (for example an installation script). This is pretty similar - to archives generated with WinZip Self-Extractor in the Windows world. -Keywords: Installation archive tar winzip -Author: Stephane Peter (megastep@megastep.org) -Maintained-by: Stephane Peter (megastep@megastep.org) -Original-site: https://makeself.io/ -Platform: Unix -Copying-policy: GPL -End diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.sh deleted file mode 100644 index c8ea56597..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/makeself.sh +++ /dev/null @@ -1,822 +0,0 @@ -#!/bin/sh -# -# Makeself version 2.4.x -# by Stephane Peter -# -# Utility to create self-extracting tar.gz archives. -# The resulting archive is a file holding the tar.gz archive with -# a small Shell script stub that uncompresses the archive to a temporary -# directory and then executes a given script from withing that directory. -# -# Makeself home page: https://makeself.io/ -# -# Version 2.0 is a rewrite of version 1.0 to make the code easier to read and maintain. -# -# Version history : -# - 1.0 : Initial public release -# - 1.1 : The archive can be passed parameters that will be passed on to -# the embedded script, thanks to John C. Quillan -# - 1.2 : Package distribution, bzip2 compression, more command line options, -# support for non-temporary archives. Ideas thanks to Francois Petitjean -# - 1.3 : More patches from Bjarni R. Einarsson and Francois Petitjean: -# Support for no compression (--nocomp), script is no longer mandatory, -# automatic launch in an xterm, optional verbose output, and -target -# archive option to indicate where to extract the files. -# - 1.4 : Improved UNIX compatibility (Francois Petitjean) -# Automatic integrity checking, support of LSM files (Francois Petitjean) -# - 1.5 : Many bugfixes. Optionally disable xterm spawning. -# - 1.5.1 : More bugfixes, added archive options -list and -check. -# - 1.5.2 : Cosmetic changes to inform the user of what's going on with big -# archives (Quake III demo) -# - 1.5.3 : Check for validity of the DISPLAY variable before launching an xterm. -# More verbosity in xterms and check for embedded command's return value. -# Bugfix for Debian 2.0 systems that have a different "print" command. -# - 1.5.4 : Many bugfixes. Print out a message if the extraction failed. -# - 1.5.5 : More bugfixes. Added support for SETUP_NOCHECK environment variable to -# bypass checksum verification of archives. -# - 1.6.0 : Compute MD5 checksums with the md5sum command (patch from Ryan Gordon) -# - 2.0 : Brand new rewrite, cleaner architecture, separated header and UNIX ports. -# - 2.0.1 : Added --copy -# - 2.1.0 : Allow multiple tarballs to be stored in one archive, and incremental updates. -# Added --nochown for archives -# Stopped doing redundant checksums when not necesary -# - 2.1.1 : Work around insane behavior from certain Linux distros with no 'uncompress' command -# Cleaned up the code to handle error codes from compress. Simplified the extraction code. -# - 2.1.2 : Some bug fixes. Use head -n to avoid problems. -# - 2.1.3 : Bug fixes with command line when spawning terminals. -# Added --tar for archives, allowing to give arbitrary arguments to tar on the contents of the archive. -# Added --noexec to prevent execution of embedded scripts. -# Added --nomd5 and --nocrc to avoid creating checksums in archives. -# Added command used to create the archive in --info output. -# Run the embedded script through eval. -# - 2.1.4 : Fixed --info output. -# Generate random directory name when extracting files to . to avoid problems. (Jason Trent) -# Better handling of errors with wrong permissions for the directory containing the files. (Jason Trent) -# Avoid some race conditions (Ludwig Nussel) -# Unset the $CDPATH variable to avoid problems if it is set. (Debian) -# Better handling of dot files in the archive directory. -# - 2.1.5 : Made the md5sum detection consistent with the header code. -# Check for the presence of the archive directory -# Added --encrypt for symmetric encryption through gpg (Eric Windisch) -# Added support for the digest command on Solaris 10 for MD5 checksums -# Check for available disk space before extracting to the target directory (Andreas Schweitzer) -# Allow extraction to run asynchronously (patch by Peter Hatch) -# Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo) -# - 2.1.6 : Replaced one dot per file progress with a realtime progress percentage and a spining cursor (Guy Baconniere) -# Added --noprogress to prevent showing the progress during the decompression (Guy Baconniere) -# Added --target dir to allow extracting directly to a target directory (Guy Baconniere) -# - 2.2.0 : Many bugfixes, updates and contributions from users. Check out the project page on Github for the details. -# - 2.3.0 : Option to specify packaging date to enable byte-for-byte reproducibility. (Marc Pawlowsky) -# - 2.4.0 : Optional support for SHA256 checksums in archives. -# - 2.4.2 : Add support for threads for several compressors. (M. Limber) -# Added zstd support. -# - 2.4.3 : Make explicit POSIX tar archives for increased compatibility. -# - 2.4.5 : Added --tar-format to override ustar tar archive format -# -# (C) 1998-2021 by Stephane Peter -# -# This software is released under the terms of the GNU GPL version 2 and above -# Please read the license at http://www.gnu.org/copyleft/gpl.html -# Self-extracting archives created with this script are explictly NOT released under the term of the GPL -# - -MS_VERSION=2.4.5 -MS_COMMAND="$0" -unset CDPATH - -for f in ${1+"$@"}; do - MS_COMMAND="$MS_COMMAND \\\\ - \\\"$f\\\"" -done - -# For Solaris systems -if test -d /usr/xpg4/bin; then - PATH=/usr/xpg4/bin:$PATH - export PATH -fi - -# Procedures - -MS_Usage() -{ - echo "Usage: $0 [args] archive_dir file_name label startup_script [script_args]" - echo "args can be one or more of the following :" - echo " --version | -v : Print out Makeself version number and exit" - echo " --help | -h : Print out this help message" - echo " --tar-quietly : Suppress verbose output from the tar command" - echo " --quiet | -q : Do not print any messages other than errors." - echo " --gzip : Compress using gzip (default if detected)" - echo " --pigz : Compress with pigz" - echo " --zstd : Compress with zstd" - echo " --bzip2 : Compress using bzip2 instead of gzip" - echo " --pbzip2 : Compress using pbzip2 instead of gzip" - echo " --xz : Compress using xz instead of gzip" - echo " --lzo : Compress using lzop instead of gzip" - echo " --lz4 : Compress using lz4 instead of gzip" - echo " --compress : Compress using the UNIX 'compress' command" - echo " --complevel lvl : Compression level for gzip pigz zstd xz lzo lz4 bzip2 and pbzip2 (default 9)" - echo " --threads thds : Number of threads to be used by compressors that support parallelization." - echo " Omit to use compressor's default. Most useful (and required) for opting" - echo " into xz's threading, usually with '--threads=0' for all available cores." - echo " pbzip2 and pigz are parallel by default, and setting this value allows" - echo " limiting the number of threads they use." - echo " --base64 : Instead of compressing, encode the data using base64" - echo " --gpg-encrypt : Instead of compressing, encrypt the data using GPG" - echo " --gpg-asymmetric-encrypt-sign" - echo " : Instead of compressing, asymmetrically encrypt and sign the data using GPG" - echo " --gpg-extra opt : Append more options to the gpg command line" - echo " --ssl-encrypt : Instead of compressing, encrypt the data using OpenSSL" - echo " --ssl-passwd pass : Use the given password to encrypt the data using OpenSSL" - echo " --ssl-pass-src src : Use the given src as the source of password to encrypt the data" - echo " using OpenSSL. See \"PASS PHRASE ARGUMENTS\" in man openssl." - echo " If this option is not supplied, the user will be asked to enter" - echo " encryption password on the current terminal." - echo " --ssl-no-md : Do not use \"-md\" option not supported by older OpenSSL." - echo " --nochown : Do not give the target folder to the current user (default)" - echo " --chown : Give the target folder to the current user recursively" - echo " --nocomp : Do not compress the data" - echo " --notemp : The archive will create archive_dir in the" - echo " current directory and uncompress in ./archive_dir" - echo " --needroot : Check that the root user is extracting the archive before proceeding" - echo " --copy : Upon extraction, the archive will first copy itself to" - echo " a temporary directory" - echo " --append : Append more files to an existing Makeself archive" - echo " The label and startup scripts will then be ignored" - echo " --target dir : Extract directly to a target directory" - echo " directory path can be either absolute or relative" - echo " --nooverwrite : Do not extract the archive if the specified target directory exists" - echo " --current : Files will be extracted to the current directory" - echo " Both --current and --target imply --notemp" - echo " --tar-format opt : Specify a tar archive format (default is ustar)" - echo " --tar-extra opt : Append more options to the tar command line" - echo " --untar-extra opt : Append more options to the during the extraction of the tar archive" - echo " --nomd5 : Don't calculate an MD5 for archive" - echo " --nocrc : Don't calculate a CRC for archive" - echo " --sha256 : Compute a SHA256 checksum for the archive" - echo " --header file : Specify location of the header script" - echo " --cleanup file : Specify a cleanup script that executes on interrupt and when finished successfully." - echo " --follow : Follow the symlinks in the archive" - echo " --noprogress : Do not show the progress during the decompression" - echo " --nox11 : Disable automatic spawn of a xterm" - echo " --nowait : Do not wait for user input after executing embedded" - echo " program from an xterm" - echo " --sign passphrase : Signature private key to sign the package with" - echo " --lsm file : LSM file describing the package" - echo " --license file : Append a license file" - echo " --help-header file : Add a header to the archive's --help output" - echo " --packaging-date date" - echo " : Use provided string as the packaging date" - echo " instead of the current date." - echo - echo " --keep-umask : Keep the umask set to shell default, rather than overriding when executing self-extracting archive." - echo " --export-conf : Export configuration variables to startup_script" - echo - echo "Do not forget to give a fully qualified startup script name" - echo "(i.e. with a ./ prefix if inside the archive)." - exit 1 -} - -# Default settings -if type gzip >/dev/null 2>&1; then - COMPRESS=gzip -elif type compress >/dev/null 2>&1; then - COMPRESS=compress -else - echo "ERROR: missing commands: gzip, compress" >&2 - MS_Usage -fi -ENCRYPT=n -PASSWD="" -PASSWD_SRC="" -OPENSSL_NO_MD=n -COMPRESS_LEVEL=9 -DEFAULT_THREADS=123456 # Sentinel value -THREADS=$DEFAULT_THREADS -KEEP=n -CURRENT=n -NOX11=n -NOWAIT=n -APPEND=n -TAR_QUIETLY=n -KEEP_UMASK=n -QUIET=n -NOPROGRESS=n -COPY=none -NEED_ROOT=n -TAR_ARGS=rvf -TAR_FORMAT=ustar -TAR_EXTRA="" -GPG_EXTRA="" -DU_ARGS=-ks -HEADER=`dirname "$0"`/makeself-header.sh -SIGNATURE="" -TARGETDIR="" -NOOVERWRITE=n -DATE=`LC_ALL=C date` -EXPORT_CONF=n -SHA256=n -OWNERSHIP=n -SIGN=n -GPG_PASSPHRASE="" - -# LSM file stuff -LSM_CMD="echo No LSM. >> \"\$archname\"" - -while true -do - case "$1" in - --version | -v) - echo Makeself version $MS_VERSION - exit 0 - ;; - --pbzip2) - COMPRESS=pbzip2 - shift - ;; - --bzip2) - COMPRESS=bzip2 - shift - ;; - --gzip) - COMPRESS=gzip - shift - ;; - --pigz) - COMPRESS=pigz - shift - ;; - --zstd) - COMPRESS=zstd - shift - ;; - --xz) - COMPRESS=xz - shift - ;; - --lzo) - COMPRESS=lzo - shift - ;; - --lz4) - COMPRESS=lz4 - shift - ;; - --compress) - COMPRESS=compress - shift - ;; - --base64) - COMPRESS=base64 - shift - ;; - --gpg-encrypt) - COMPRESS=gpg - shift - ;; - --gpg-asymmetric-encrypt-sign) - COMPRESS=gpg-asymmetric - shift - ;; - --gpg-extra) - GPG_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-encrypt) - ENCRYPT=openssl - shift - ;; - --ssl-passwd) - PASSWD=$2 - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-pass-src) - PASSWD_SRC=$2 - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-no-md) - OPENSSL_NO_MD=y - shift - ;; - --nocomp) - COMPRESS=none - shift - ;; - --complevel) - COMPRESS_LEVEL="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --threads) - THREADS="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --nochown) - OWNERSHIP=n - shift - ;; - --chown) - OWNERSHIP=y - shift - ;; - --notemp) - KEEP=y - shift - ;; - --copy) - COPY=copy - shift - ;; - --current) - CURRENT=y - KEEP=y - shift - ;; - --tar-format) - TAR_FORMAT="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --tar-extra) - TAR_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --untar-extra) - UNTAR_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --target) - TARGETDIR="$2" - KEEP=y - shift 2 || { MS_Usage; exit 1; } - ;; - --sign) - SIGN=y - GPG_PASSPHRASE="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --nooverwrite) - NOOVERWRITE=y - shift - ;; - --needroot) - NEED_ROOT=y - shift - ;; - --header) - HEADER="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --cleanup) - CLEANUP_SCRIPT="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --license) - # We need to escape all characters having a special meaning in double quotes - LICENSE=$(sed 's/\\/\\\\/g; s/"/\\\"/g; s/`/\\\`/g; s/\$/\\\$/g' "$2") - shift 2 || { MS_Usage; exit 1; } - ;; - --follow) - TAR_ARGS=rvhf - DU_ARGS=-ksL - shift - ;; - --noprogress) - NOPROGRESS=y - shift - ;; - --nox11) - NOX11=y - shift - ;; - --nowait) - NOWAIT=y - shift - ;; - --nomd5) - NOMD5=y - shift - ;; - --sha256) - SHA256=y - shift - ;; - --nocrc) - NOCRC=y - shift - ;; - --append) - APPEND=y - shift - ;; - --lsm) - LSM_CMD="cat \"$2\" >> \"\$archname\"" - shift 2 || { MS_Usage; exit 1; } - ;; - --packaging-date) - DATE="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --help-header) - HELPHEADER=`sed -e "s/'/'\\\\\''/g" $2` - shift 2 || { MS_Usage; exit 1; } - [ -n "$HELPHEADER" ] && HELPHEADER="$HELPHEADER -" - ;; - --tar-quietly) - TAR_QUIETLY=y - shift - ;; - --keep-umask) - KEEP_UMASK=y - shift - ;; - --export-conf) - EXPORT_CONF=y - shift - ;; - -q | --quiet) - QUIET=y - shift - ;; - -h | --help) - MS_Usage - ;; - -*) - echo Unrecognized flag : "$1" - MS_Usage - ;; - *) - break - ;; - esac -done - -if test $# -lt 1; then - MS_Usage -else - if test -d "$1"; then - archdir="$1" - else - echo "Directory $1 does not exist." >&2 - exit 1 - fi -fi -archname="$2" - -if test "$QUIET" = "y" || test "$TAR_QUIETLY" = "y"; then - if test "$TAR_ARGS" = "rvf"; then - TAR_ARGS="rf" - elif test "$TAR_ARGS" = "rvhf"; then - TAR_ARGS="rhf" - fi -fi - -if test "$APPEND" = y; then - if test $# -lt 2; then - MS_Usage - fi - - # Gather the info from the original archive - OLDENV=`sh "$archname" --dumpconf` - if test $? -ne 0; then - echo "Unable to update archive: $archname" >&2 - exit 1 - else - eval "$OLDENV" - OLDSKIP=`expr $SKIP + 1` - fi -else - if test "$KEEP" = n -a $# = 3; then - echo "ERROR: Making a temporary archive with no embedded command does not make sense!" >&2 - echo >&2 - MS_Usage - fi - # We don't want to create an absolute directory unless a target directory is defined - if test "$CURRENT" = y; then - archdirname="." - elif test x"$TARGETDIR" != x; then - archdirname="$TARGETDIR" - else - archdirname=`basename "$1"` - fi - - if test $# -lt 3; then - MS_Usage - fi - - LABEL="$3" - SCRIPT="$4" - test "x$SCRIPT" = x || shift 1 - shift 3 - SCRIPTARGS="$*" -fi - -if test "$KEEP" = n -a "$CURRENT" = y; then - echo "ERROR: It is A VERY DANGEROUS IDEA to try to combine --notemp and --current." >&2 - exit 1 -fi - -case $COMPRESS in -gzip) - GZIP_CMD="gzip -c$COMPRESS_LEVEL" - GUNZIP_CMD="gzip -cd" - ;; -pigz) - GZIP_CMD="pigz -$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD --processes $THREADS" - fi - GUNZIP_CMD="gzip -cd" - ;; -zstd) - GZIP_CMD="zstd -$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD --threads=$THREADS" - fi - GUNZIP_CMD="zstd -cd" - ;; -pbzip2) - GZIP_CMD="pbzip2 -c$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD -p$THREADS" - fi - GUNZIP_CMD="bzip2 -d" - ;; -bzip2) - GZIP_CMD="bzip2 -$COMPRESS_LEVEL" - GUNZIP_CMD="bzip2 -d" - ;; -xz) - GZIP_CMD="xz -c$COMPRESS_LEVEL" - # Must opt-in by specifying a value since not all versions of xz support threads - if test $THREADS -ne $DEFAULT_THREADS; then - GZIP_CMD="$GZIP_CMD --threads=$THREADS" - fi - GUNZIP_CMD="xz -d" - ;; -lzo) - GZIP_CMD="lzop -c$COMPRESS_LEVEL" - GUNZIP_CMD="lzop -d" - ;; -lz4) - GZIP_CMD="lz4 -c$COMPRESS_LEVEL" - GUNZIP_CMD="lz4 -d" - ;; -base64) - GZIP_CMD="base64" - GUNZIP_CMD="base64 --decode -i -" - ;; -gpg) - GZIP_CMD="gpg $GPG_EXTRA -ac -z$COMPRESS_LEVEL" - GUNZIP_CMD="gpg -d" - ENCRYPT="gpg" - ;; -gpg-asymmetric) - GZIP_CMD="gpg $GPG_EXTRA -z$COMPRESS_LEVEL -es" - GUNZIP_CMD="gpg --yes -d" - ENCRYPT="gpg" - ;; -compress) - GZIP_CMD="compress -fc" - GUNZIP_CMD="(type compress >/dev/null 2>&1 && compress -fcd || gzip -cd)" - ;; -none) - GZIP_CMD="cat" - GUNZIP_CMD="cat" - ;; -esac - -if test x"$ENCRYPT" = x"openssl"; then - if test x"$APPEND" = x"y"; then - echo "Appending to existing archive is not compatible with OpenSSL encryption." >&2 - fi - - ENCRYPT_CMD="openssl enc -aes-256-cbc -salt" - DECRYPT_CMD="openssl enc -aes-256-cbc -d" - - if test x"$OPENSSL_NO_MD" != x"y"; then - ENCRYPT_CMD="$ENCRYPT_CMD -md sha256" - DECRYPT_CMD="$DECRYPT_CMD -md sha256" - fi - - if test -n "$PASSWD_SRC"; then - ENCRYPT_CMD="$ENCRYPT_CMD -pass $PASSWD_SRC" - elif test -n "$PASSWD"; then - ENCRYPT_CMD="$ENCRYPT_CMD -pass pass:$PASSWD" - fi -fi - -tmpfile="${TMPDIR:-/tmp}/mkself$$" - -if test -f "$HEADER"; then - oldarchname="$archname" - archname="$tmpfile" - # Generate a fake header to count its lines - SKIP=0 - . "$HEADER" - SKIP=`cat "$tmpfile" |wc -l` - # Get rid of any spaces - SKIP=`expr $SKIP` - rm -f "$tmpfile" - if test "$QUIET" = "n"; then - echo "Header is $SKIP lines long" >&2 - fi - archname="$oldarchname" -else - echo "Unable to open header file: $HEADER" >&2 - exit 1 -fi - -if test "$QUIET" = "n"; then - echo -fi - -if test "$APPEND" = n; then - if test -f "$archname"; then - echo "WARNING: Overwriting existing file: $archname" >&2 - fi -fi - -USIZE=`du $DU_ARGS "$archdir" | awk '{print $1}'` - -if test "." = "$archdirname"; then - if test "$KEEP" = n; then - archdirname="makeself-$$-`date +%Y%m%d%H%M%S`" - fi -fi - -test -d "$archdir" || { echo "Error: $archdir does not exist."; rm -f "$tmpfile"; exit 1; } -if test "$QUIET" = "n"; then - echo "About to compress $USIZE KB of data..." - echo "Adding files to archive named \"$archname\"..." -fi - -# See if we have GNU tar -TAR=`exec <&- 2>&-; which gtar || command -v gtar || type gtar` -test -x "$TAR" || TAR=tar - -tmparch="${TMPDIR:-/tmp}/mkself$$.tar" -( - if test "$APPEND" = "y"; then - tail -n "+$OLDSKIP" "$archname" | eval "$GUNZIP_CMD" > "$tmparch" - fi - cd "$archdir" - # "Determining if a directory is empty" - # https://www.etalabs.net/sh_tricks.html - find . \ - \( \ - ! -type d \ - -o \ - \( -links 2 -exec sh -c ' - is_empty () ( - cd "$1" - set -- .[!.]* ; test -f "$1" && return 1 - set -- ..?* ; test -f "$1" && return 1 - set -- * ; test -f "$1" && return 1 - return 0 - ) - is_empty "$0"' {} \; \ - \) \ - \) -print \ - | LC_ALL=C sort \ - | sed 's/./\\&/g' \ - | xargs $TAR $TAR_EXTRA --format $TAR_FORMAT -$TAR_ARGS "$tmparch" -) || { - echo "ERROR: failed to create temporary archive: $tmparch" - rm -f "$tmparch" "$tmpfile" - exit 1 -} - -USIZE=`du $DU_ARGS "$tmparch" | awk '{print $1}'` - -eval "$GZIP_CMD" <"$tmparch" >"$tmpfile" || { - echo "ERROR: failed to create temporary file: $tmpfile" - rm -f "$tmparch" "$tmpfile" - exit 1 -} -rm -f "$tmparch" - -if test x"$ENCRYPT" = x"openssl"; then - echo "About to encrypt archive \"$archname\"..." - { eval "$ENCRYPT_CMD -in $tmpfile -out ${tmpfile}.enc" && mv -f ${tmpfile}.enc $tmpfile; } || \ - { echo Aborting: could not encrypt temporary file: "$tmpfile".; rm -f "$tmpfile"; exit 1; } -fi - -fsize=`cat "$tmpfile" | wc -c | tr -d " "` - -# Compute the checksums - -shasum=0000000000000000000000000000000000000000000000000000000000000000 -md5sum=00000000000000000000000000000000 -crcsum=0000000000 - -if test "$NOCRC" = y; then - if test "$QUIET" = "n"; then - echo "skipping crc at user request" - fi -else - crcsum=`CMD_ENV=xpg4 cksum < "$tmpfile" | sed -e 's/ /Z/' -e 's/ /Z/' | cut -dZ -f1` - if test "$QUIET" = "n"; then - echo "CRC: $crcsum" - fi -fi - -if test "$SHA256" = y; then - SHA_PATH=`exec <&- 2>&-; which shasum || command -v shasum || type shasum` - if test -x "$SHA_PATH"; then - shasum=`eval "$SHA_PATH -a 256" < "$tmpfile" | cut -b-64` - else - SHA_PATH=`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum` - shasum=`eval "$SHA_PATH" < "$tmpfile" | cut -b-64` - fi - if test "$QUIET" = "n"; then - if test -x "$SHA_PATH"; then - echo "SHA256: $shasum" - else - echo "SHA256: none, SHA command not found" - fi - fi -fi -if test "$NOMD5" = y; then - if test "$QUIET" = "n"; then - echo "Skipping md5sum at user request" - fi -else - # Try to locate a MD5 binary - OLD_PATH=$PATH - PATH=${GUESS_MD5_PATH:-"$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} - MD5_ARG="" - MD5_PATH=`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum` - test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which md5 || command -v md5 || type md5` - test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which digest || command -v digest || type digest` - PATH=$OLD_PATH - if test -x "$MD5_PATH"; then - if test `basename ${MD5_PATH}`x = digestx; then - MD5_ARG="-a md5" - fi - md5sum=`eval "$MD5_PATH $MD5_ARG" < "$tmpfile" | cut -b-32` - if test "$QUIET" = "n"; then - echo "MD5: $md5sum" - fi - else - if test "$QUIET" = "n"; then - echo "MD5: none, MD5 command not found" - fi - fi -fi -if test "$SIGN" = y; then - GPG_PATH=`exec <&- 2>&-; which gpg || command -v gpg || type gpg` - if test -x "$GPG_PATH"; then - SIGNATURE=`$GPG_PATH --pinentry-mode=loopback --batch --yes --passphrase "$GPG_PASSPHRASE" --output - --detach-sig $tmpfile | base64 | tr -d \\\\n` - if test "$QUIET" = "n"; then - echo "Signature: $SIGNATURE" - fi - else - echo "Missing gpg command" >&2 - fi -fi - -totalsize=0 -for size in $fsize; -do - totalsize=`expr $totalsize + $size` -done - -if test "$APPEND" = y; then - mv "$archname" "$archname".bak || exit - - # Prepare entry for new archive - filesizes="$fsize" - CRCsum="$crcsum" - MD5sum="$md5sum" - SHAsum="$shasum" - Signature="$SIGNATURE" - # Generate the header - . "$HEADER" - # Append the new data - cat "$tmpfile" >> "$archname" - - chmod +x "$archname" - rm -f "$archname".bak - if test "$QUIET" = "n"; then - echo "Self-extractable archive \"$archname\" successfully updated." - fi -else - filesizes="$fsize" - CRCsum="$crcsum" - MD5sum="$md5sum" - SHAsum="$shasum" - Signature="$SIGNATURE" - - # Generate the header - . "$HEADER" - - # Append the compressed tar data after the stub - if test "$QUIET" = "n"; then - echo - fi - cat "$tmpfile" >> "$archname" - chmod +x "$archname" - if test "$QUIET" = "n"; then - echo Self-extractable archive \"$archname\" successfully created. - fi -fi -rm -f "$tmpfile" diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/run-tests.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/run-tests.sh deleted file mode 100644 index 31ee16511..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/makeself/run-tests.sh +++ /dev/null @@ -1,8 +0,0 @@ -#!/bin/sh -# Run every available test - Bash needed -cd test -for test in *test; -do - echo "Running test $test ..." - bash $test || { echo "*** ERROR: Test '$test' failed!"; exit 1; } -done diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/merge_aicpu_info_json.sh b/atb_operator/AtbAddSample/AddOp/cmake/util/merge_aicpu_info_json.sh deleted file mode 100644 index a977bd51d..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/merge_aicpu_info_json.sh +++ /dev/null @@ -1,31 +0,0 @@ -#!/bin/bash - -project_path=$1 -build_path=$2 -vendor_name=customize -echo $@ -if [[ ! -d "$project_path" ]]; then - echo "[ERROR] No projcet path is provided" - exit 1 -fi - -if [[ ! -d "$build_path" ]]; then - echo "[ERROR] No build path is provided" - exit 1 -fi - -if [[ ! -d "$ASCEND_OPP_PATH" ]]; then - echo "[ERROR] No opp install path is provided" - exit 1 -fi -custom_exist_info_json=$ASCEND_OPP_PATH/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json -custom_new_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json -temp_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/temp_cust_aicpu_kernel.json - -if [[ -f "$custom_exist_info_json" ]] && [[ -f "$custom_new_info_json" ]]; then - cp -f $custom_exist_info_json $temp_info_json - chmod +w $temp_info_json - python3 ${project_path}/cmake/util/insert_op_info.py ${custom_new_info_json} ${temp_info_json} - cp -f $temp_info_json $custom_new_info_json - rm -f $temp_info_json -fi diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/opdesc_parser.py b/atb_operator/AtbAddSample/AddOp/cmake/util/opdesc_parser.py deleted file mode 100644 index c38e79edc..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/opdesc_parser.py +++ /dev/null @@ -1,260 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os - - -OP_ALL = '__ALLOP__' -SOC_ALL = '__ALLSOC__' -SOC_TO_SHORT_SOC_MAP = { - "ascend910a": "ascend910", - "ascend910proa": "ascend910", - "ascend910b": "ascend910", - "ascend910prob": "ascend910", - "ascend910premiuma": "ascend910", - "ascend910b1": "ascend910b", - "ascend910b2": "ascend910b", - "ascend910b3": "ascend910b", - "ascend910b4": "ascend910b", - "ascend310p1": "ascend310p", - "ascend310p3": "ascend310p", - "ascend310p3vir01": "ascend310p", - "ascend310p3vir02": "ascend310p", - "ascend310p3vir04": "ascend310p", - "ascend310p3vir08": "ascend310p", - "ascend310b1": "ascend310b", - "bs9sx1aa": "bs9sx1a" -} - - -class OpDesc: - def __init__(self: any, op_type: str): - self.op_type = op_type - self.attr_list = [] - self.attr_val = {} - self.input_name = [] - self.input_type = [] - self.input_dtype = [] - self.input_fmt = [] - self.output_name = [] - self.output_type = [] - self.output_dtype = [] - self.output_fmt = [] - self.op_fmt_sel = False - self.op_chk_support = False - self.op_intf = '' - self.kern_name = '' - self.op_file = '' - self.op_replay_flag = False - self.op_replay_batch = False - self.input_idx = -1 - self.output_idx = -1 - self.max_block_dim = 32 - self.max_shape_size = 268435456 - self.dynamic_shape = False - self.op_range_limit = '' - self.custom_compile_options = {} - self.custom_all_compile_options = {} - - @staticmethod - def _parse_digit(conf: str) -> int: - return int(conf.split('=')[1]) - - @staticmethod - def _parse_flag(conf: str) -> bool: - if 'true' == conf.split('=')[1]: - return True - return False - - @staticmethod - def _parse_str(conf: str) -> str: - return conf.split('=')[1] - - @staticmethod - def _parse_list(conf: str) -> list: - return conf.split('=')[1].split(',') - - def parse_input(self: any, conf: str): - if conf.startswith('input{}.name'.format(int(self.input_idx) + 1)): - self.input_idx += 1 - self.input_name.append(self._parse_str(conf)) - elif conf.startswith('input{}.paramType'.format(int(self.input_idx))): - self.input_type.append(self._parse_str(conf)) - elif conf.startswith('input{}.dtype'.format(int(self.input_idx))): - self.input_dtype.append(self._parse_str(conf)) - elif conf.startswith('input{}.format'.format(int(self.input_idx))): - self.input_fmt.append(self._parse_str(conf)) - else: - return - - def parse_output(self: any, conf: str): - if conf.startswith('output{}.name'.format(int(self.output_idx) + 1)): - self.output_idx += 1 - self.output_name.append(self._parse_str(conf)) - elif conf.startswith('output{}.paramType'.format(int(self.output_idx))): - self.output_type.append(self._parse_str(conf)) - elif conf.startswith('output{}.dtype'.format(int(self.output_idx))): - self.output_dtype.append(self._parse_str(conf)) - elif conf.startswith('output{}.format'.format(int(self.output_idx))): - self.output_fmt.append(self._parse_str(conf)) - else: - return - - def parse_op_format(self: any, conf: str): - self.op_fmt_sel = self._parse_flag(conf) - - def parse_check_support(self: any, conf: str): - self.op_chk_support = self._parse_flag(conf) - - def parse_range_limit(self: any, conf: str): - self.op_range_limit = self._parse_str(conf) - - def parse_kern_name(self: any, conf: str): - self.kern_name = self._parse_str(conf) - - def parse_op_intf(self: any, conf: str): - self.op_intf = self._parse_str(conf) - - def parse_op_file(self: any, conf: str): - self.op_file = self._parse_str(conf) - - def parse_dynamic_shape(self: any, conf: str): - self.dynamic_shape = self._parse_flag(conf) - - def parse_attr_list(self: any, conf: str): - self.attr_list = self._parse_list(conf) - - def parse_attr_val(self: any, conf: str): - for attr in self.attr_list: - if self.attr_val.get(attr) is None: - self.attr_val[attr] = {} - if conf.startswith('attr_{}.type'.format(attr)): - self.attr_val.get(attr)['type'] = self._parse_str(conf) - elif conf.startswith('attr_{}.paramType'.format(attr)): - self.attr_val.get(attr)['paramType'] = self._parse_str(conf) - elif conf.startswith('attr_{}.defaultValue'.format(attr)): - self.attr_val.get(attr)['defaultValue'] = self._parse_str(conf) - - def parse_replay_val(self: any, batch_list: list, iterator_list: list): - if self.op_type in batch_list: - self.op_replay_flag = True - self.op_replay_batch = True - elif self.op_type in iterator_list: - self.op_replay_flag = True - self.op_replay_batch = False - - -def _is_op_type_in_opdesc(op_descs: list, op_type: str): - for op in op_descs: - if op_type == op.op_type: - return True - return False - - -def _set_all_options_to_opdescs(op_descs, soc_ver_compile_options): - for op in op_descs: - op.custom_all_compile_options = soc_ver_compile_options - - -def _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options): - for op in op_descs: - if op.op_type != op_type: - continue - op.custom_compile_options = soc_ver_compile_options - - -def _trans_soc_ver_to_short(soc_ver: str): - low_soc_ver = soc_ver.lower() - if low_soc_ver not in SOC_TO_SHORT_SOC_MAP: - print(f'WARNING: caution: {soc_ver} will trans into ascend910, if not your intention,' - f'use ascend910b1~4 instead') - return SOC_TO_SHORT_SOC_MAP[low_soc_ver] - - -def _get_op_custom_options(op_descs: list, auto_gen_dir: str): - if auto_gen_dir is None: - return {} - file = os.path.join(auto_gen_dir, "custom_compile_options.ini") - if not os.path.exists(file): - print(f'WARNING: cannot find {auto_gen_dir}/custom_compile_options.ini') - return {} - with open (file, 'r') as fd: - lines = fd.readlines() - for line in lines: - param_list = str.split(line.rstrip('\n'), ',') - if len(param_list) != 3: - raise Exception(f'ERROR: custom compile option {param_list} len is not 3') - op_type = param_list[0] - if op_type.upper() == 'ALL': - op_type = OP_ALL - if op_type != OP_ALL and _is_op_type_in_opdesc(op_descs, op_type) == False: - print(f'WARNING: op: {op_type} are not exists in this project') - continue - soc_ver_compile_options = {} - soc_ver = param_list[1] - options_str = param_list[2] - options = str.split(options_str, ';') - if soc_ver == '': - soc_ver_compile_options[SOC_ALL] = options - else: - soc_ver_list = str.split(soc_ver, ';') - for ver in soc_ver_list: - short_ver = _trans_soc_ver_to_short(ver) - soc_ver_compile_options[short_ver] = options - if op_type == OP_ALL: - _set_all_options_to_opdescs(op_descs, soc_ver_compile_options) - else: - _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options) - - -def get_op_desc(file: str, batch_list: list, iterator_list: list, builder: any, - op_type: list, auto_gen_dir: str = None) -> list: - op_descs = [] - op_match = False - with open (file, 'r') as fd: - lines = fd.readlines() - for line in lines: - line = line.strip() - if line.startswith('['): - name = line[1:-1] - if op_type is None or name in op_type: - op_match = True - op_desc = builder(name) - op_desc.parse_replay_val(batch_list, iterator_list) - op_descs.append(op_desc) - else: - op_match = False - if op_type is not None and len(op_descs) == len(op_type): - return op_descs - continue - if not op_match: - continue - if line.startswith('input'): - op_desc.parse_input(line) - elif line.startswith('output'): - op_desc.parse_output(line) - elif line.startswith('dynamicFormat.flag'): - op_desc.parse_op_format(line) - elif line.startswith('needCheckSupport.flag'): - op_desc.parse_check_support(line) - elif line.startswith('rangeLimit.value'): - op_desc.parse_range_limit(line) - elif line.startswith('opInterface.value'): - op_desc.parse_op_intf(line) - elif line.startswith('kernel.name'): - op_desc.parse_kern_name(line) - elif line.startswith('opFile.value'): - op_desc.parse_op_file(line) - elif line.startswith('dynamicShapeSupport.flag'): - op_desc.parse_dynamic_shape(line) - elif line.startswith('attr.list'): - op_desc.parse_attr_list(line) - elif line.startswith('attr_'): - op_desc.parse_attr_val(line) - _get_op_custom_options(op_descs, auto_gen_dir) - return op_descs diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/parse_ini_to_json.py b/atb_operator/AtbAddSample/AddOp/cmake/util/parse_ini_to_json.py deleted file mode 100644 index 721465fee..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/parse_ini_to_json.py +++ /dev/null @@ -1,338 +0,0 @@ -# Copyright 2020-2021 Huawei Technologies Co., Ltd -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -""" -parser ini to json -""" - -import json -import os -import stat -import sys - - -ATTR_TYPE_LIST = ["int", "float", "bool", "str", "listInt", "listFloat", "listBool", "listStr", "listListInt", - "type", "listType", "tensor", "listTensor"] -ATTR_PARAMTYPE_LIST = ["optional", "required"] -BOOL_FLAG_KEY = ["dynamicFormat", "dynamicShapeSupport", "dynamicRankSupport", "precision_reduce", "heavyOp", - "needCheckSupport"] -BOOL_LIST = ["true", "false"] -DTYPE_LIST = ["float16", "float", "float32", "int8", "int16", "int32", "uint8", "uint16", "uint32", "bool", - "int64", "uint64", "qint8", "qint16", "qint32", "quint8", "quint16", "double", "complex64", - "complex128", "string", "resource", "dual", "dual_sub_int8", "dual_sub_uint8", "string_ref", - "int4", "bfloat16", "uint1"] -FORMAT_LIST = ["NCHW", "NHWC", "ND", "NC1HWC0", "FRACTAL_Z", "NC1C0HWPAD", "NHWC1C0", "FSR_NCHW", "FRACTAL_DECONV", - "C1HWNC0", "FRACTAL_DECONV_TRANSPOSE", "FRACTAL_DECONV_SP_STRIDE_TRANS", "NC1HWC0_C04", - "FRACTAL_Z_C04", "CHWN", "FRACTAL_DECONV_SP_STRIDE8_TRANS", "HWCN", "NC1KHKWHWC0", "BN_WEIGHT", - "FILTER_HWCK", "HASHTABLE_LOOKUP_LOOKUPS", "HASHTABLE_LOOKUP_KEYS", "HASHTABLE_LOOKUP_VALUE", - "HASHTABLE_LOOKUP_OUTPUT", "HASHTABLE_LOOKUP_HITS", "C1HWNCoC0", "MD", "NDHWC", "FRACTAL_ZZ", - "FRACTAL_NZ", "NCDHW", "DHWCN", "NDC1HWC0", "FRACTAL_Z_3D", "CN", "NC", "DHWNC", - "FRACTAL_Z_3D_TRANSPOSE", "FRACTAL_ZN_LSTM", "FRACTAL_ZN_RNN", "FRACTAL_Z_G", "NULL"] - - -def parse_ini_files(ini_files): - """ - parse ini files to json - Parameters: - ---------------- - ini_files:input file list - return:ops_info - ---------------- - """ - tbe_ops_info = {} - for ini_file in ini_files: - check_file_size(ini_file) - parse_ini_to_obj(ini_file, tbe_ops_info) - return tbe_ops_info - - -def check_file_size(input_file): - try: - file_size = os.path.getsize(input_file) - except OSError as os_error: - print('[ERROR] Failed to open "%s". %s' % (input_file, str(os_error))) - raise OSError from os_error - if file_size > 10*1024*1024: - print('[WARN] The size of %s exceeds 10MB, it may take more time to run, please wait.' % input_file) - - -def parse_ini_to_obj(ini_file, tbe_ops_info): - """ - parse ini file to json obj - Parameters: - ---------------- - ini_file:ini file path - tbe_ops_info:ops_info - ---------------- - """ - with open(ini_file) as ini_file: - lines = ini_file.readlines() - op_dict = {} - op_name = "" - find_op_type = False - for line in lines: - line = line.rstrip() - if line == "": - continue - if line.startswith("["): - if line.endswith("]"): - op_name = line[1:-1] - op_dict = {} - tbe_ops_info[op_name] = op_dict - find_op_type = True - elif "=" in line: - key1 = line[:line.index("=")] - key2 = line[line.index("=")+1:] - key1_0, key1_1 = key1.split(".") - if key1_0 not in op_dict: - op_dict[key1_0] = {} - if key1_1 in op_dict.get(key1_0): - raise RuntimeError("Op:" + op_name + " " + key1_0 + " " + - key1_1 + " is repeated!") - dic_key = op_dict.get(key1_0) - dic_key[key1_1] = key2 - else: - continue - if not find_op_type: - raise RuntimeError("Not find OpType in .ini file.") - - -def check_output_exist(op_dict, is_valid): - """ - Function Description: - Check output is exist - Parameter: op_dict - Parameter: is_valid - """ - if "output0" in op_dict: - output0_dict = op_dict.get("output0") - if output0_dict.get("name", None) is None: - is_valid = False - print("output0.name is required in .ini file!") - else: - is_valid = False - print("output0 is required in .ini file!") - return is_valid - - -def check_attr_dict(attr_dict, is_valid, attr): - """ - Function Description: - Check attr_dict - Parameter: attr_dict - Parameter: is_valid - Parameter: attr - """ - attr_type = attr_dict.get("type") - value = attr_dict.get("value") - param_type = attr_dict.get("paramType") - if attr_type is None or value is None: - is_valid = False - print("If attr.list is exist, {0}.type and {0}.value is required".format(attr)) - if param_type and param_type not in ATTR_PARAMTYPE_LIST: - is_valid = False - print("{0}.paramType only support {1}.".format(attr, ATTR_PARAMTYPE_LIST)) - if attr_type and attr_type not in ATTR_TYPE_LIST: - is_valid = False - print("{0}.type only support {1}.".format(attr, ATTR_TYPE_LIST)) - return is_valid - - -def check_attr(op_dict, is_valid): - """ - Function Description: - Check attr - Parameter: op_dict - Parameter: is_valid - """ - if "attr" in op_dict: - attr_dict = op_dict.get("attr") - attr_list_str = attr_dict.get("list", None) - if attr_list_str is None: - is_valid = False - print("attr.list is required in .ini file!") - else: - attr_list = attr_list_str.split(",") - for attr_name in attr_list: - attr = "attr_" + attr_name.strip() - attr_dict = op_dict.get(attr) - if attr_dict: - is_valid = check_attr_dict(attr_dict, is_valid, attr) - else: - is_valid = False - print("%s is required in .ini file, when attr.list is %s!" % (attr, attr_list_str)) - return is_valid - - -def check_bool_flag(op_dict, is_valid): - """ - Function Description: - check_bool_flag - Parameter: op_dict - Parameter: is_valid - """ - for key in BOOL_FLAG_KEY: - if key in op_dict: - op_bool_key = op_dict.get(key) - if op_bool_key.get("flag").strip() not in BOOL_LIST: - is_valid = False - print("{0}.flag only support {1}.".format(key, BOOL_LIST)) - return is_valid - - -def check_type_format(op_info, is_valid, op_info_key): - """ - Function Description: - Check type and format - Parameter: op_info - Parameter: is_valid - Parameter: op_info_key - """ - op_info_dtype_str = op_info.get("dtype") - op_info_dtype_num = 0 - op_info_format_num = 0 - if op_info_dtype_str: - op_info_dtype = op_info_dtype_str.split(",") - op_info_dtype_num = len(op_info_dtype) - for dtype in op_info_dtype: - if dtype.strip() not in DTYPE_LIST: - is_valid = False - print("{0}.dtype not support {1}.".format(op_info_key, dtype)) - op_info_format_str = op_info.get("format") - if op_info_format_str: - op_info_format = op_info_format_str.split(",") - op_info_format_num = len(op_info_format) - for op_format in op_info_format: - if op_format.strip() not in FORMAT_LIST: - is_valid = False - print("{0}.format not support {1}.".format(op_info_key, op_format)) - if op_info_dtype_num > 0 and op_info_format_num > 0: - if op_info_dtype_num != op_info_format_num: - is_valid = False - print("The number of {0}.dtype not match the number of {0}.format.".format(op_info_key)) - return is_valid - - -def check_op_info(tbe_ops): - """ - Function Description: - Check info. - Parameter: tbe_ops - Return Value: is_valid - """ - print("\n\n==============check valid for ops info start==============") - required_op_input_info_keys = ["paramType", "name"] - required_op_output_info_keys = ["paramType", "name"] - param_type_valid_value = ["dynamic", "optional", "required"] - is_valid = True - for op_key in tbe_ops: - op_dict = tbe_ops[op_key] - for op_info_key in op_dict: - if op_info_key.startswith("input"): - op_input_info = op_dict[op_info_key] - missing_keys = [] - for required_op_input_info_key in required_op_input_info_keys: - if required_op_input_info_key not in op_input_info: - missing_keys.append(required_op_input_info_key) - if len(missing_keys) > 0: - print("op: " + op_key + " " + op_info_key + " missing: " + - ",".join(missing_keys)) - is_valid = False - else: - if not op_input_info["paramType"] in param_type_valid_value: - print("op: " + op_key + " " + op_info_key + \ - " paramType not valid, valid key:[dynamic, " - "optional, required]") - is_valid = False - is_valid = check_type_format(op_input_info, is_valid, op_info_key) - if op_info_key.startswith("output"): - op_input_info = op_dict[op_info_key] - missing_keys = [] - for required_op_input_info_key in required_op_output_info_keys: - if required_op_input_info_key not in op_input_info: - missing_keys.append(required_op_input_info_key) - if len(missing_keys) > 0: - print("op: " + op_key + " " + op_info_key + " missing: " + - ",".join(missing_keys)) - is_valid = False - else: - if not op_input_info["paramType"] in param_type_valid_value: - print("op: " + op_key + " " + op_info_key + - " paramType not valid, valid key:[dynamic, " - "optional, required]") - is_valid = False - is_valid = check_type_format(op_input_info, is_valid, op_info_key) - is_valid = check_attr(op_dict, is_valid) - is_valid = check_bool_flag(op_dict, is_valid) - print("==============check valid for ops info end================\n\n") - return is_valid - - -def write_json_file(tbe_ops_info, json_file_path): - """ - Save info to json file - Parameters: - ---------------- - tbe_ops_info: ops_info - json_file_path: json file path - ---------------- - """ - json_file_real_path = os.path.realpath(json_file_path) - wr_flag = os.O_WRONLY | os.O_CREAT - wr_mode = stat.S_IWUSR | stat.S_IRUSR - with os.fdopen(os.open(json_file_real_path, wr_flag, wr_mode), 'w') as file_path: - # Only the owner and group have rights - os.chmod(json_file_real_path, stat.S_IWGRP + stat.S_IWUSR + stat.S_IRGRP - + stat.S_IRUSR) - json.dump(tbe_ops_info, file_path, sort_keys=True, indent=4, - separators=(',', ':')) - print("Compile op info cfg successfully.") - - -def parse_ini_to_json(ini_file_paths, outfile_path): - """ - parse ini files to json file - Parameters: - ---------------- - ini_file_paths: list of ini file path - outfile_path: output file path - ---------------- - """ - tbe_ops_info = parse_ini_files(ini_file_paths) - if not check_op_info(tbe_ops_info): - print("Compile op info cfg failed.") - return False - write_json_file(tbe_ops_info, outfile_path) - return True - - -if __name__ == '__main__': - args = sys.argv - - OUTPUT_FILE_PATH = "tbe_ops_info.json" - ini_file_path_list = [] - - for arg in args: - if arg.endswith("ini"): - ini_file_path_list.append(arg) - OUTPUT_FILE_PATH = arg.replace(".ini", ".json") - if arg.endswith("json"): - OUTPUT_FILE_PATH = arg - - if len(ini_file_path_list) == 0: - ini_file_path_list.append("tbe_ops_info.ini") - - if not parse_ini_to_json(ini_file_path_list, OUTPUT_FILE_PATH): - sys.exit(1) - sys.exit(0) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/preset_parse.py b/atb_operator/AtbAddSample/AddOp/cmake/util/preset_parse.py deleted file mode 100644 index 8f1124b1d..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/preset_parse.py +++ /dev/null @@ -1,23 +0,0 @@ -import json -import sys -import os - - -def get_config_opts(file): - src_dir = os.path.abspath(os.path.dirname(file)) - opts = '' - with open(file, 'r') as fd: - config = json.load(fd) - for conf in config: - if conf == 'configurePresets': - for node in config[conf]: - macros = node.get('cacheVariables') - if macros is not None: - for key in macros: - opts += '-D{}={} '.format(key, macros[key]['value']) - opts = opts.replace('${sourceDir}', src_dir) - print(opts) - - -if __name__ == "__main__": - get_config_opts(sys.argv[1]) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/replay_codegen.py b/atb_operator/AtbAddSample/AddOp/cmake/util/replay_codegen.py deleted file mode 100644 index 1baa364ef..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/replay_codegen.py +++ /dev/null @@ -1,105 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import os -import stat -import collections -import kernel_entry as keb -from tiling_data_def_build import gen_tiling -import code_channel_infer -import const_var - -PYF_PATH = os.path.dirname(__file__) - -ReplayCodeGenParams = collections.namedtuple('ReplayCodeGenParams',\ -['op_type', 'impl', 'tiling_file', 'kernel', 'entry', 'argn', 'op_replay_batch', 'max_block_dim', 'max_shape_size']) - - -class ReplayCodeGen: - def __init__(self, replayCodeGenParams): - self.op_type = replayCodeGenParams.op_type - self.impl = replayCodeGenParams.impl - self.tiling_file = replayCodeGenParams.tiling_file - self.tiling_data_file = '' - self.kernel = replayCodeGenParams.kernel - self.entry = replayCodeGenParams.entry - self.argn = replayCodeGenParams.argn - self.batch = False - self.outdir = '' - self.data_type = 'uint8_t' - self.blknum = 32 - self.op_replay_batch = replayCodeGenParams.op_replay_batch - self.max_block_dim = replayCodeGenParams.max_block_dim - self.max_shape_size = replayCodeGenParams.max_shape_size - - def set_batch(self, is_batch): - self.batch = is_batch - - def set_outdir(self, outdir): - self.outdir = outdir - - def gen_replay(self, ops_product: str): - kerentry = os.path.join(self.outdir, self.kernel + '_entry.cce') - kerimpl = os.path.join(self.outdir, self.kernel + '_impl.cpp') - replayimpl = os.path.join(self.outdir, self.kernel + '_replay.cpp') - if self.batch: - reptmp = os.path.join(PYF_PATH, 'batch_replay_impl.temp') - else: - reptmp = os.path.join(PYF_PATH, 'replay_impl.temp') - kertmp = os.path.join(PYF_PATH, 'kernel_impl.temp') - self._gen_kentry(kerentry) - self._gen_kimpl_code(kerimpl, kertmp) - self._gen_tiling_data_header() - self._gen_replay_code(replayimpl, reptmp, ops_product) - - def _gen_tiling_data_header(self): - self.tiling_data_file = os.path.join(self.outdir, self.kernel + '_tiling_data.h') - gen_tiling(self.tiling_file, self.tiling_data_file) - - def _gen_kimpl_code(self, src, tmpfile): - with open(tmpfile, 'r') as fd: - temp = fd.read() - temp = temp.replace('__CCE_FILE__', self.impl) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(temp) - - def _gen_replay_code(self, src, tmpfile, ops_product: str): - with open(tmpfile, 'r') as fd: - temp = fd.read() - temp = temp.replace('__ARG_NUM__', str(self.argn)) - argdef = [] - kargs = [] - for i in range(0, self.argn): - argdef.append('{} *'.format(self.data_type)) - kargs.append('({} *)GetArg({})'.format(self.data_type, i)) - temp = temp.replace('__ARGS_DEF__', ', '.join(argdef)) - temp = temp.replace('__KERNEL_ARGS__', ', '.join(kargs)) - temp = temp.replace('__KERNEL_FUN__', self.entry) - core_type_infer = 'core_type' - code_channel = code_channel_infer.infer_code_channel(code_channel_infer.InfoCodeChanelParams(self.impl,\ - self.tiling_data_file, self.kernel, self.outdir, ops_product, None)) - if code_channel == code_channel_infer.CODE_VEC: - core_type_infer = '0' - elif code_channel == code_channel_infer.CODE_CUBE: - core_type_infer = '1' - temp = temp.replace('__CORE_TYPE__', core_type_infer) - # regist function - temp = temp.replace('__OPS_PRODUCT__', ops_product) - temp = temp.replace('__OPTYPE__', self.op_type) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(temp) - - def _gen_kentry(self, src): - kf = '' - pre_alloc_str = 'A' * 256 - if self.batch: - kf += keb.batch_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str), self.argn, self.data_type) - else: - kf += keb.mc_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str),\ - self.argn, self.data_type, self.blknum) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(kf) diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/replay_impl.temp b/atb_operator/AtbAddSample/AddOp/cmake/util/replay_impl.temp deleted file mode 100644 index 1d30dd865..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/replay_impl.temp +++ /dev/null @@ -1,120 +0,0 @@ -#include -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#include "register/op_check.h" -#define __ASCENDC_REPLAY_CODE__ -using namespace std; -using namespace optiling; -using namespace AscendCReplay; - -extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); -extern "C" int elf_append(char *elf, uint32_t elfSize, char *jit, int kernum, int blknum[], char *atext[], - int alen[], int atlen, const char* kernelname[]); - -#define KERNEL_N 1 -#define ARG_N (__ARG_NUM__) -#define MAX_L (1024 * 1024 * 100) -#define MAX_E (1024 * 1024) - -int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) -{ - // gen type 1 : direct call codes 0: load .o file - if (param.gentype < 0 || param.gentype > 1) { - printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); - return 0; - } else if (param.gentype == 1 && param.objptr == nullptr) { - printf("Error: call replay with direct call mode, but code obj addr is null\n"); - return 0; - } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { - printf("Error: call replay with object file mode, but object file path is null\n"); - return 0; - } - // core_type 0:MIX 1:CUBE 2:VEC - if (core_type < 0 || core_type > 2) { - printf("Error: call replay core type is %d !\n", core_type); - return 0; - } - g_coreType = __CORE_TYPE__; - g_taskRation = param.task_ration; - g_tilingKey = param.tiling_key; - - unsigned char *buf, *jit; - char *kernel[KERNEL_N * 32]; - int len[KERNEL_N * 32]; - int blknum[KERNEL_N]; - int max; - block_num = param.block_dim; - g_ubBase = block_num; - uint8_t *code = (uint8_t *)malloc(MAX_L); - uint8_t *pos = code; - struct timespec tp1, tp2; - - clock_gettime(CLOCK_MONOTONIC, &tp1); - if (block_num > 32) { - printf("Error: block_num > 32\n"); - return 0; - } - //__OP_FOPEN__ - for (int i = 0; i < KERNEL_N; i++) { - for (int j = 0; j < ARG_N; j++) - AddArg(j, ARG_STEP * (j + 1)); - for (block_idx = 0; block_idx < block_num; block_idx++) { - //__OP_SET_KERNEL__ - int code_idx = i * block_num + block_idx; -#ifdef FP_CEILING - SetCtrlFloatEnable(); -#else - SetCtrlFloatDisable(); -#endif - CodeInit(pos, false); - __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); - CodeEnd(); - kernel[code_idx] = (char *)pos; - len[code_idx] = CodeLen(); - pos += len[code_idx]; - printf("kernel %d core %ld code generated len %d\n", i, block_idx, len[code_idx]); - } - blknum[i] = block_num; - } - //__OP_FCLOSE__ - clock_gettime(CLOCK_MONOTONIC, &tp2); - buf = (unsigned char *)malloc(MAX_E); - int fd = open(param.entry_file, O_RDONLY); - if (fd < 0) { - printf("[error]: cannot find entry.o : %s\n", param.entry_file); - return 0; - } - uint32_t bufSize = read(fd, buf, MAX_E); - if (bufSize <= 0) { - printf("[error]: entry.o : %s is too small ! \n", param.entry_file); - } - close(fd); - jit = (unsigned char *)malloc(MAX_L); - printf("total code generated %ld\n", pos - code); - int sz = elf_append((char *)buf, bufSize, (char *)jit, KERNEL_N, blknum, kernel, len, pos - code, ¶m.kernel_name); - if (tp1.tv_sec != tp2.tv_sec) { - printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); - } else { - printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); - } - printf("new elf size %d\n", sz); - if (param.gentype == 0) { - fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); - (void)write(fd, jit, sz); - close(fd); - free(jit); - } else if (param.gentype == 1) { - *param.objptr = (char*)jit; - } - free(buf); - free(code); - return sz; -} - -REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbAddSample/AddOp/cmake/util/tiling_data_def_build.py b/atb_operator/AtbAddSample/AddOp/cmake/util/tiling_data_def_build.py deleted file mode 100644 index a96304261..000000000 --- a/atb_operator/AtbAddSample/AddOp/cmake/util/tiling_data_def_build.py +++ /dev/null @@ -1,84 +0,0 @@ -#!/usr/bin/env python -# coding=utf-8 -""" -Function: -The replay funtion entry -Copyright Information: -Huawei Technologies Co., Ltd. All Rights Reserved © 2020 -""" - -import sys -import os -import stat -import re -import const_var - - -def gen_tiling(tiling_header_file: str, tiling_file_out: str): - if not os.path.exists(tiling_header_file): - print("warning: no userdef tiling header file: ", tiling_header_file) - return - print("generate tiling def header file: ", tiling_file_out) - tmp_name = os.path.splitext(os.path.basename(tiling_header_file))[0].upper() - tiling_source = '#ifndef __{}_H__\n'.format(tmp_name) - tiling_source += '#define __{}_H__\n\n'.format(tmp_name) - tiling_source += '#include \n' - tiling_source += '#include \n\n' - tiling_source += '#include "kernel_tiling/kernel_tiling.h"\n\n' - end_source = "" - pattern = re.compile(r'[(](.*)[)]', re.S) - with open(tiling_header_file, 'r') as fd: - lines = fd.readlines() - for line in lines: - line = line.strip() - if (line.startswith('BEGIN_TILING_DATA_DEF')): - tiling_source += '#pragma pack(1)\n' - tiling_source += 'struct ' - struct_def = re.findall(pattern, line)[0] - tiling_source += struct_def + ' {\n' - elif (line.startswith('TILING_DATA_FIELD_DEF_ARR')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {}[{}] = {{}};\n'.format(fds[0].strip(), fds[2].strip(), fds[1].strip()) - elif (line.startswith('TILING_DATA_FIELD_DEF_STRUCT')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {};\n'.format(fds[0].strip(), fds[1].strip()) - elif (line.startswith('TILING_DATA_FIELD_DEF')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {} = 0;\n'.format(fds[0].strip(), fds[1].strip()) - elif (line.startswith('END_TILING_DATA_DEF')): - tiling_source += '};\n' - tiling_source += '#pragma pack()\n\n' - tiling_source += '#ifdef __NPU_TILING__\n' - tiling_source += \ - 'inline [aicore] void Init{stru}(const __gm__ uint8_t* tiling, {stru}* const_data)\n'\ - .format(stru=struct_def) - tiling_source += '{\n' - tiling_source += ' const __gm__ uint32_t *src = (const __gm__ uint32_t *)tiling;\n' - tiling_source += ' uint32_t *dst = (uint32_t *)const_data;\n' - tiling_source += ' for (auto i = 0; i < sizeof({}) / 4; i++) *(dst + i) = *(src + i);\n'\ - .format(struct_def) - tiling_source += '}\n' - tiling_source += '#else\n' - tiling_source += 'inline void Init{stru}(uint8_t* tiling, {stru}* const_data)\n'.format(stru=struct_def) - tiling_source += '{\n' - tiling_source += ' memcpy(const_data, tiling, sizeof({}));\n'.format(struct_def) - tiling_source += '}\n' - tiling_source += '#endif\n\n' - end_source = ''' -#define GET_TILING_DATA(tiling_data, tiling_arg) \\ -{stru} tiling_data; \\ -Init{stru}(tiling_arg, &tiling_data)\n -'''.format(stru=struct_def) - tiling_source += end_source - tiling_source += '#endif' - with os.fdopen(os.open(tiling_file_out, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(tiling_source) - - -if __name__ == '__main__': - if len(sys.argv) <= 2: - raise RuntimeError('arguments must greater than 2') - gen_tiling(sys.argv[1], sys.argv[2]) diff --git a/atb_operator/AtbAddSample/AddOp/framework/CMakeLists.txt b/atb_operator/AtbAddSample/AddOp/framework/CMakeLists.txt deleted file mode 100644 index b6be9b492..000000000 --- a/atb_operator/AtbAddSample/AddOp/framework/CMakeLists.txt +++ /dev/null @@ -1,11 +0,0 @@ -if(NOT EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/mindspore") - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/caffe_plugin") - add_subdirectory(caffe_plugin) - endif() - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/tf_plugin") - add_subdirectory(tf_plugin) - endif() - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/onnx_plugin") - add_subdirectory(onnx_plugin) - endif() -endif() diff --git a/atb_operator/AtbAddSample/AddOp/framework/tf_plugin/CMakeLists.txt b/atb_operator/AtbAddSample/AddOp/framework/tf_plugin/CMakeLists.txt deleted file mode 100644 index a6aba5c20..000000000 --- a/atb_operator/AtbAddSample/AddOp/framework/tf_plugin/CMakeLists.txt +++ /dev/null @@ -1,14 +0,0 @@ - -aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} plugin_srcs) -add_library(cust_tf_parsers SHARED ${plugin_srcs}) -target_compile_definitions(cust_tf_parsers PRIVATE google=ascend_private) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_tf_parsers PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_tf_parsers PRIVATE intf_pub graph) -install(TARGETS cust_tf_parsers - LIBRARY DESTINATION packages/vendors/${vendor_name}/framework/tensorflow -) diff --git a/atb_operator/AtbAddSample/AddOp/framework/tf_plugin/tensorflow_add_custom_plugin.cc b/atb_operator/AtbAddSample/AddOp/framework/tf_plugin/tensorflow_add_custom_plugin.cc deleted file mode 100644 index 2cd837ce5..000000000 --- a/atb_operator/AtbAddSample/AddOp/framework/tf_plugin/tensorflow_add_custom_plugin.cc +++ /dev/null @@ -1,23 +0,0 @@ -/* Copyright (C) 2020-2021. Huawei Technologies Co., Ltd. All -rights reserved. - * - * This program is free software; you can redistribute it and/or modify - * it under the terms of the Apache License Version 2.0. - * You may not use this file except in compliance with the License. - * - * This program is distributed in the hope that it will be useful, - * but WITHOUT ANY WARRANTY; without even the implied warranty of - * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the - * Apache License for more details at - * http://www.apache.org/licenses/LICENSE-2.0 - */ - -#include "register/register.h" - -namespace domi { -// register op info to GE -REGISTER_CUSTOM_OP("AddCustom") - .FrameworkType(TENSORFLOW) // type: CAFFE, TENSORFLOW - .OriginOpType("Add") // name in tf module - .ParseParamsByOperatorFn(AutoMappingByOpFn); -} // namespace domi diff --git a/atb_operator/AtbAddSample/AddOp/op_host/CMakeLists.txt b/atb_operator/AtbAddSample/AddOp/op_host/CMakeLists.txt deleted file mode 100644 index 40dd51cfa..000000000 --- a/atb_operator/AtbAddSample/AddOp/op_host/CMakeLists.txt +++ /dev/null @@ -1,82 +0,0 @@ - -aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} ops_srcs) - -opbuild(OPS_SRC ${ops_srcs} - OUT_DIR ${ASCEND_AUTOGEN_PATH} -) - -add_library(cust_op_proto SHARED ${ops_srcs} ${ASCEND_AUTOGEN_PATH}/op_proto.cc) -target_compile_definitions(cust_op_proto PRIVATE OP_PROTO_LIB) -target_compile_options(cust_op_proto PRIVATE - -fvisibility=hidden -) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_op_proto PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_op_proto PRIVATE - intf_pub - exe_graph - register - tiling_api - -Wl,--whole-archive - rt2_registry - -Wl,--no-whole-archive -) -set_target_properties(cust_op_proto PROPERTIES OUTPUT_NAME - cust_opsproto_rt2.0 -) -add_library(cust_optiling SHARED ${ops_srcs}) -target_compile_definitions(cust_optiling PRIVATE OP_TILING_LIB) -target_compile_options(cust_optiling PRIVATE - -fvisibility=hidden -) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_optiling PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_optiling PRIVATE - intf_pub - exe_graph - register - tiling_api - -Wl,--whole-archive - rt2_registry - -Wl,--no-whole-archive -) -set_target_properties(cust_optiling PROPERTIES OUTPUT_NAME - cust_opmaster_rt2.0 -) - -file(GLOB aclnn_src ${ASCEND_AUTOGEN_PATH}/aclnn_*.cpp) -file(GLOB aclnn_inc ${ASCEND_AUTOGEN_PATH}/aclnn_*.h) -add_library(cust_opapi SHARED ${aclnn_src}) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_opapi PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_opapi PRIVATE intf_pub ascendcl nnopbase) - -add_custom_target(optiling_compat ALL - COMMAND ln -sf lib/linux/${CMAKE_SYSTEM_PROCESSOR}/$ - ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so -) - -install(TARGETS cust_op_proto - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_proto/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) -install(FILES ${ASCEND_AUTOGEN_PATH}/op_proto.h - DESTINATION packages/vendors/${vendor_name}/op_proto/inc) -install(TARGETS cust_optiling - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) -install(FILES ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so - DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling) -install(TARGETS cust_opapi - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_api/lib) -install(FILES ${aclnn_inc} - DESTINATION packages/vendors/${vendor_name}/op_api/include) diff --git a/atb_operator/AtbAddSample/AddOp/op_host/add_custom.cpp b/atb_operator/AtbAddSample/AddOp/op_host/add_custom.cpp deleted file mode 100644 index b61a67f11..000000000 --- a/atb_operator/AtbAddSample/AddOp/op_host/add_custom.cpp +++ /dev/null @@ -1,117 +0,0 @@ -#include "add_custom_tiling.h" -#include "register/op_def_registry.h" -#include "graph/utils/type_utils.h" -#include "tiling/platform/platform_ascendc.h" - -namespace optiling { -const uint32_t BLOCK_SIZE = 32; -const uint32_t BUFFER_NUM = 2; -static ge::graphStatus TilingFunc(gert::TilingContext* context) -{ - TilingData tiling; - uint64_t ubSize; - auto ascendcPlatform = platform_ascendc::PlatformAscendC(context->GetPlatformInfo()); - ascendcPlatform.GetCoreMemSize(platform_ascendc::CoreMemType::UB, ubSize); - auto coreNum = ascendcPlatform.GetCoreNum(); - - // Based on the input length and the number of inputs, the number of bytes of the input data type is obtained - uint32_t inputNum = context->GetInputShape(0)->GetStorageShape().GetShapeSize(); - uint32_t typeLength = 0; - ge::TypeUtils::GetDataTypeLength(context->GetInputDesc(0)->GetDataType(), typeLength); - uint32_t inputLength = inputNum * typeLength; - uint32_t inputBytes = inputLength / inputNum; - - // There are a total of 3 shared UB spaces in the input and output. If it's int8, there are 2 more TBUFs - uint32_t ubDataNumber = (inputBytes == 1) ? 5 : 3; - // The number of 32B data blocks that can be used for each data. DOUBLE BUFFER is already counted here - uint32_t tileBlockNum = (ubSize / BLOCK_SIZE / BUFFER_NUM) / ubDataNumber; - uint32_t tileDataNum = (tileBlockNum * BLOCK_SIZE) / inputBytes; - - // Input data for 32B alignment - uint32_t inputLengthAlgin32 = (((inputLength + BLOCK_SIZE - 1) / BLOCK_SIZE) * BLOCK_SIZE); - // There is at least 32B of data on each core, satisfying several settings for several cores. The maximum number of audits is the actual number of audits - coreNum = (coreNum < inputLengthAlgin32 / BLOCK_SIZE) ? coreNum : inputLengthAlgin32 / BLOCK_SIZE; - coreNum = (coreNum >= 1) ? coreNum : 1; - uint32_t everyCoreInputBlockNum = inputLengthAlgin32 / BLOCK_SIZE / coreNum; - uint32_t tailBlockNum = (inputLengthAlgin32 / BLOCK_SIZE) % coreNum; - - // Small chunks are calculated and sliced several times using the number of data on each core - uint32_t smallCoreDataNum = everyCoreInputBlockNum * BLOCK_SIZE / inputBytes; - uint32_t smallTileNum = everyCoreInputBlockNum / tileBlockNum; - uint32_t finalSmallTileNum = (everyCoreInputBlockNum % tileBlockNum) == 0 ? smallTileNum : smallTileNum + 1; - // Tail block calculation for small chunks of data - uint32_t smallTailDataNum = smallCoreDataNum - (tileDataNum * smallTileNum); - smallTailDataNum = smallTailDataNum == 0 ? tileDataNum : smallTailDataNum; - - // The total length of a large block of data is 32B larger than that of a small block of data - everyCoreInputBlockNum += 1; - uint32_t bigCoreDataNum = everyCoreInputBlockNum * BLOCK_SIZE / inputBytes; - uint32_t bigTileNum = everyCoreInputBlockNum / tileBlockNum; - uint32_t finalBigTileNum = (everyCoreInputBlockNum % tileBlockNum) == 0 ? bigTileNum : bigTileNum + 1; - uint32_t bigTailDataNum = bigCoreDataNum - tileDataNum * bigTileNum; - bigTailDataNum = bigTailDataNum == 0 ? tileDataNum : bigTailDataNum; - - tiling.set_smallCoreDataNum(smallCoreDataNum); - tiling.set_bigCoreDataNum(bigCoreDataNum); - tiling.set_tileDataNum(tileDataNum); - tiling.set_smallTailDataNum(smallTailDataNum); - tiling.set_bigTailDataNum(bigTailDataNum); - tiling.set_finalSmallTileNum(finalSmallTileNum); - tiling.set_finalBigTileNum(finalBigTileNum); - tiling.set_tailBlockNum(tailBlockNum); - - context->SetBlockDim(coreNum); - tiling.SaveToBuffer(context->GetRawTilingData()->GetData(), context->GetRawTilingData()->GetCapacity()); - context->GetRawTilingData()->SetDataSize(tiling.GetDataSize()); - size_t *currentWorkspace = context->GetWorkspaceSizes(1); - currentWorkspace[0] = 0; - return ge::GRAPH_SUCCESS; -} -} - -namespace ge { -static ge::graphStatus InferShape(gert::InferShapeContext* context) -{ - const gert::Shape* x1_shape = context->GetInputShape(0); - gert::Shape* y_shape = context->GetOutputShape(0); - *y_shape = *x1_shape; - return GRAPH_SUCCESS; -} -static graphStatus InferDataType(gert::InferDataTypeContext* context) -{ - const auto inputDataType = context->GetInputDataType(0); - context->SetOutputDataType(0, inputDataType); - return ge::GRAPH_SUCCESS; -} -} - -namespace ops { -class AddCustom : public OpDef { -public: - explicit AddCustom(const char* name) : OpDef(name) - { - this->Input("x") - .ParamType(REQUIRED) - .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) - .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) - .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); - this->Input("y") - .ParamType(REQUIRED) - .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) - .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) - .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); - this->Output("z") - .ParamType(REQUIRED) - .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) - .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) - .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); - - this->SetInferShape(ge::InferShape).SetInferDataType(ge::InferDataType); - this->AICore() - .SetTiling(optiling::TilingFunc) - .AddConfig("ascend310b") - .AddConfig("ascend910b"); - } -}; -OP_ADD(AddCustom); -} diff --git a/atb_operator/AtbAddSample/AddOp/op_host/add_custom_tiling.h b/atb_operator/AtbAddSample/AddOp/op_host/add_custom_tiling.h deleted file mode 100644 index 28fd00a37..000000000 --- a/atb_operator/AtbAddSample/AddOp/op_host/add_custom_tiling.h +++ /dev/null @@ -1,22 +0,0 @@ -/* - * Copyright (c) Huawei Technologies Co., Ltd. 2022-2023. All rights reserved. - */ -#ifndef ADD_CUSTOM_TILING_H -#define ADD_CUSTOM_TILING_H -#include "register/tilingdata_base.h" - -namespace optiling { -BEGIN_TILING_DATA_DEF(TilingData) - TILING_DATA_FIELD_DEF(uint32_t, smallCoreDataNum); - TILING_DATA_FIELD_DEF(uint32_t, bigCoreDataNum); - TILING_DATA_FIELD_DEF(uint32_t, finalBigTileNum); - TILING_DATA_FIELD_DEF(uint32_t, finalSmallTileNum); - TILING_DATA_FIELD_DEF(uint32_t, tileDataNum); - TILING_DATA_FIELD_DEF(uint32_t, smallTailDataNum); - TILING_DATA_FIELD_DEF(uint32_t, bigTailDataNum); - TILING_DATA_FIELD_DEF(uint32_t, tailBlockNum); -END_TILING_DATA_DEF; - -REGISTER_TILING_DATA_CLASS(AddCustom, TilingData) -} -#endif // ADD_CUSTOM_TILING_H \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AddOp/op_kernel/CMakeLists.txt b/atb_operator/AtbAddSample/AddOp/op_kernel/CMakeLists.txt deleted file mode 100644 index 0d31a444c..000000000 --- a/atb_operator/AtbAddSample/AddOp/op_kernel/CMakeLists.txt +++ /dev/null @@ -1,61 +0,0 @@ -# set custom compile options -if ("${CMAKE_BUILD_TYPE}x" STREQUAL "Debugx") - add_ops_compile_options(ALL OPTIONS -g -O0) -endif() - -foreach(compute_unit ${ASCEND_COMPUTE_UNIT}) - - # generate aic-${compute_unit}-ops-info.json - add_ops_info_target(TARGET ops_info_gen_${compute_unit} - OUTPUT ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core/${compute_unit}/aic-${compute_unit}-ops-info.json - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/config/${compute_unit} - ) - - # generate ascendc impl py once - if (NOT TARGET ascendc_impl_gen) - add_ops_impl_target(TARGET ascendc_impl_gen - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl - ) - endif() - - # dynamic shape binary compile - if (${ENABLE_BINARY_PACKAGE}) - add_bin_compile_target(TARGET ascendc_bin_${compute_unit} - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} - ADP_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/dynamic - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/binary/${compute_unit} - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/kernel - COMPUTE_UNIT ${compute_unit} - ) - add_dependencies(ascendc_bin_${compute_unit} ascendc_impl_gen) - endif() - -endforeach() - -# generate npu_supported_ops.json -add_npu_support_target(TARGET npu_supported_ops - OPS_INFO_DIR ${ASCEND_AUTOGEN_PATH} - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core - INSTALL_DIR packages/vendors/${vendor_name}/framework/${ASCEND_FRAMEWORK_TYPE} -) - -if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) - add_subdirectory(testcases) -endif() - -# install kernel file -if (${ENABLE_SOURCE_PACKAGE}) - file(GLOB KERNEL_FILES - ${CMAKE_CURRENT_SOURCE_DIR}/*.cpp - ${CMAKE_CURRENT_SOURCE_DIR}/*.h - ${CMAKE_CURRENT_SOURCE_DIR}/*.py - ) - install(FILES ${KERNEL_FILES} - DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl/dynamic - ) -endif() diff --git a/atb_operator/AtbAddSample/AddOp/op_kernel/add_custom.cpp b/atb_operator/AtbAddSample/AddOp/op_kernel/add_custom.cpp deleted file mode 100644 index 37c80a825..000000000 --- a/atb_operator/AtbAddSample/AddOp/op_kernel/add_custom.cpp +++ /dev/null @@ -1,134 +0,0 @@ -/* - * Copyright (c) Huawei Technologies Co., Ltd. 2022-2023. All rights reserved. - * - * Function : z = x + y - * This sample is a very basic sample that implements vector add on Ascend plaform. - */ -#include "kernel_operator.h" -// tensor num for each queue -constexpr int32_t BUFFER_NUM = 2; - -template class KernelAdd { - using T = TYPE_X; -public: - __aicore__ inline KernelAdd() {} - __aicore__ inline void Init(GM_ADDR x, GM_ADDR y, GM_ADDR z, uint32_t smallCoreDataNum, - uint32_t bigCoreDataNum, uint32_t finalBigTileNum, - uint32_t finalSmallTileNum, uint32_t tileDataNum, - uint32_t smallTailDataNum, uint32_t bigTailDataNum, - uint32_t tailBlockNum) - { - ASSERT(AscendC::GetBlockNum() != 0 && "block dim can not be zero!"); - uint32_t coreNum = AscendC::GetBlockIdx(); - uint32_t globalBufferIndex = bigCoreDataNum * AscendC::GetBlockIdx(); - this->tileDataNum = tileDataNum; - if (coreNum < tailBlockNum) { - this->coreDataNum = bigCoreDataNum; - this->tileNum = finalBigTileNum; - this->tailDataNum = bigTailDataNum; - } - else { - this->coreDataNum = smallCoreDataNum; - this->tileNum = finalSmallTileNum; - this->tailDataNum = smallTailDataNum; - globalBufferIndex -= (bigCoreDataNum - smallCoreDataNum) * (AscendC::GetBlockIdx() - tailBlockNum); - } - xGm.SetGlobalBuffer((__gm__ TYPE_X*)x + globalBufferIndex, this->coreDataNum); - yGm.SetGlobalBuffer((__gm__ TYPE_Y*)y + globalBufferIndex, this->coreDataNum); - zGm.SetGlobalBuffer((__gm__ TYPE_Z*)z + globalBufferIndex, this->coreDataNum); - pipe.InitBuffer(inQueueX, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_X)); - pipe.InitBuffer(inQueueY, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_Y)); - pipe.InitBuffer(outQueueZ, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_Z)); - pipe.InitBuffer(tmp1, this->tileDataNum * sizeof(half)); - pipe.InitBuffer(tmp2, this->tileDataNum * sizeof(half)); - } - __aicore__ inline void Process() - { - int32_t loopCount = this->tileNum; - this->processDataNum = this->tileDataNum; - for (int32_t i = 0; i < loopCount; i++) { - if (i == this->tileNum - 1) { - this->processDataNum = this->tailDataNum; - } - CopyIn(i); - Compute(i); - CopyOut(i); - } - } - -private: - __aicore__ inline void CopyIn(int32_t progress) - { - AscendC::LocalTensor xLocal = inQueueX.AllocTensor(); - AscendC::LocalTensor yLocal = inQueueY.AllocTensor(); - AscendC::DataCopy(xLocal, xGm[progress * this->tileDataNum], this->processDataNum); - AscendC::DataCopy(yLocal, yGm[progress * this->tileDataNum], this->processDataNum); - inQueueX.EnQue(xLocal); - inQueueY.EnQue(yLocal); - } - __aicore__ inline void Compute(int32_t progress) - { - AscendC::LocalTensor xLocal = inQueueX.DeQue(); - AscendC::LocalTensor yLocal = inQueueY.DeQue(); - AscendC::LocalTensor zLocal = outQueueZ.AllocTensor(); - if constexpr (std::is_same_v) { - auto p1 = tmp1.Get(); - auto p2 = tmp2.Get(); - AscendC::Cast(p1, xLocal, AscendC::RoundMode::CAST_NONE, this->processDataNum); - AscendC::Cast(p2, yLocal, AscendC::RoundMode::CAST_NONE, this->processDataNum); - AscendC::Add(p2, p1, p2, this->processDataNum); - AscendC::Cast(p1.ReinterpretCast(), p2, AscendC::RoundMode::CAST_RINT, this->processDataNum); - AscendC::ShiftLeft(p1.ReinterpretCast(), p1.ReinterpretCast(), int16_t(8), this->processDataNum); - AscendC::ShiftRight(p1.ReinterpretCast(), p1.ReinterpretCast(), int16_t(8), this->processDataNum); - AscendC::Cast(p2, p1.ReinterpretCast(), AscendC::RoundMode::CAST_NONE, this->processDataNum); - AscendC::Cast(zLocal, p2, AscendC::RoundMode::CAST_NONE, this->processDataNum); - } - else { - AscendC::Add(zLocal, xLocal, yLocal, this->processDataNum); - } - outQueueZ.EnQue(zLocal); - inQueueX.FreeTensor(xLocal); - inQueueY.FreeTensor(yLocal); - } - __aicore__ inline void CopyOut(int32_t progress) - { - AscendC::LocalTensor zLocal = outQueueZ.DeQue(); - AscendC::DataCopy(zGm[progress * this->tileDataNum], zLocal, this->processDataNum); - outQueueZ.FreeTensor(zLocal); - } - -private: - AscendC::TPipe pipe; - AscendC::TQue inQueueX, inQueueY; - AscendC::TQue outQueueZ; - AscendC::TBuf tmp1, tmp2; - AscendC::GlobalTensor xGm; - AscendC::GlobalTensor yGm; - AscendC::GlobalTensor zGm; - uint32_t coreDataNum; - uint32_t tileNum; - uint32_t tileDataNum; - uint32_t tailDataNum; - uint32_t processDataNum; -}; - -extern "C" __global__ __aicore__ void add_custom(GM_ADDR x, GM_ADDR y, GM_ADDR z, GM_ADDR workspace, GM_ADDR tiling) -{ - GET_TILING_DATA(tiling_data, tiling); - KernelAdd op; - op.Init(x, y, z, tiling_data.smallCoreDataNum, - tiling_data.bigCoreDataNum, tiling_data.finalBigTileNum, - tiling_data.finalSmallTileNum, tiling_data.tileDataNum, - tiling_data.smallTailDataNum, tiling_data.bigTailDataNum, - tiling_data.tailBlockNum); - op.Process(); -} - -#ifndef ASCENDC_CPU_DEBUG -// call of kernel function -void add_custom_do(uint32_t blockDim, void* l2ctrl, void* stream, uint8_t* x, uint8_t* y, uint8_t* z, - uint8_t* workspace, uint8_t* tiling) -{ - add_custom<<>>(x, y, z, workspace, tiling); -} -#endif diff --git a/atb_operator/AtbAddSample/AddOp/scripts/install.sh b/atb_operator/AtbAddSample/AddOp/scripts/install.sh deleted file mode 100644 index 9d45a8052..000000000 --- a/atb_operator/AtbAddSample/AddOp/scripts/install.sh +++ /dev/null @@ -1,318 +0,0 @@ -#!/bin/bash -vendor_name=customize -targetdir=/usr/local/Ascend/opp -target_custom=0 - -sourcedir=$PWD/packages -vendordir=vendors/$vendor_name - -QUIET="y" - -while true -do - case $1 in - --quiet) - QUIET="y" - shift - ;; - --install-path=*) - INSTALL_PATH=$(echo $1 | cut -d"=" -f2-) - INSTALL_PATH=${INSTALL_PATH%*/} - shift - ;; - --*) - shift - ;; - *) - break - ;; - esac -done - -log() { - cur_date=`date +"%Y-%m-%d %H:%M:%S"` - echo "[runtime] [$cur_date] "$1 -} - -if [ -n "${INSTALL_PATH}" ]; then - if [[ ! "${INSTALL_PATH}" = /* ]]; then - log "[ERROR] use absolute path for --install-path argument" - exit 1 - fi - if [ ! -d ${INSTALL_PATH} ]; then - mkdir ${INSTALL_PATH} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${INSTALL_PATH} failed" - exit 1 - fi - fi - targetdir=${INSTALL_PATH} -elif [ -n "${ASCEND_CUSTOM_OPP_PATH}" ]; then - if [ ! -d ${ASCEND_CUSTOM_OPP_PATH} ]; then - mkdir -p ${ASCEND_CUSTOM_OPP_PATH} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${ASCEND_CUSTOM_OPP_PATH} failed" - fi - fi - targetdir=${ASCEND_CUSTOM_OPP_PATH} -else - if [ "x${ASCEND_OPP_PATH}" == "x" ]; then - log "[ERROR] env ASCEND_OPP_PATH no exist" - exit 1 - fi - targetdir="${ASCEND_OPP_PATH}" -fi - -if [ ! -d $targetdir ];then - log "[ERROR] $targetdir no exist" - exit 1 -fi - -upgrade() -{ - if [ ! -d ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 files" - return 0 - fi - - if [ ! -d ${targetdir}/$vendordir/$1 ];then - log "[INFO] create ${targetdir}/$vendordir/$1." - mkdir -p ${targetdir}/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/$1 failed" - return 1 - fi - else - has_same_file=-1 - for file_a in ${sourcedir}/$vendordir/$1/*; do - file_b=${file_a##*/}; - if [ "ls ${targetdir}/$vendordir/$1" = "" ]; then - log "[INFO] ${targetdir}/$vendordir/$1 is empty !!" - return 1 - fi - grep -q $file_b <<<`ls ${targetdir}/$vendordir/$1`; - if [[ $? -eq 0 ]]; then - echo -n "${file_b} " - has_same_file=0 - fi - done - if [ 0 -eq $has_same_file ]; then - if test $QUIET = "n"; then - echo "[INFO]: has old version in ${targetdir}/$vendordir/$1, \ - you want to Overlay Installation , please enter:[o]; \ - or replace directory installation , please enter: [r]; \ - or not install , please enter:[n]." - - while true - do - read orn - if [ "$orn" = n ]; then - return 0 - elif [ "$orn" = m ]; then - break; - elif [ "$0rn" = r ]; then - [ -n "${targetdir}/$vendordir/$1/" ] && rm -rf "${targetdir}/$vendordir/$1"/* - break; - else - echo "[ERROR] input error, please input again!" - fi - done - fi - fi - log "[INFO] replace or merge old ops $1 files .g....." - fi - - log "copy new ops $1 files ......" - if [ -d ${targetdir}/$vendordir/$1/ ]; then - chmod -R +w "$targetdir/$vendordir/$1/" >/dev/null 2>&1 - fi - cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 files failed" - return 1 - fi - - return 0 -} -upgrade_proto() -{ - if [ ! -f ${sourcedir}/$vendordir/custom.proto ]; then - log "[INFO] no need to upgrade custom.proto files" - return 0 - fi - if [ ! -d ${targetdir}/$vendordir/framework/caffe ];then - log "[INFO] create ${targetdir}/$vendordir/framework/caffe." - mkdir -p ${targetdir}/$vendordir/framework/caffe - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/framework/caffe failed" - return 1 - fi - else - if [ -f ${targetdir}/$vendordir/framework/caffe/custom.proto ]; then - # 有老版本,判断是否要覆盖式安装 - if test $QUIET = "n"; then - echo "[INFO] ${targetdir}/$vendordir/framework/caffe has old version"\ - "custom.proto file. Do you want to replace? [y/n] " - - while true - do - read yn - if [ "$yn" = n ]; then - return 0 - elif [ "$yn" = y ]; then - break; - else - echo "[ERROR] input error, please input again!" - fi - done - fi - fi - log "[INFO] replace old caffe.proto files ......" - fi - chmod -R +w "$targetdir/$vendordir/framework/caffe/" >/dev/null 2>&1 - cp -rf ${sourcedir}/$vendordir/custom.proto ${targetdir}/$vendordir/framework/caffe/ - if [ $? -ne 0 ];then - log "[ERROR] copy new custom.proto failed" - return 1 - fi - log "[INFO] copy custom.proto success" - - return 0 -} - -upgrade_file() -{ - if [ ! -e ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 file" - return 0 - fi - - log "copy new $1 files ......" - cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 file failed" - return 1 - fi - - return 0 -} - -delete_optiling_file() -{ - if [ ! -d ${targetdir}/vendors ];then - log "[INFO] $1 not exist, no need to uninstall" - return 0 - fi - sys_info=$(uname -m) - if [ ! -d ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/lib/linux/${sys_info} ];then - rm -rf ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/liboptiling.so - fi - return 0 -} - -log "[INFO] copy uninstall sh success" - -if [ ! -d ${targetdir}/vendors ];then - log "[INFO] create ${targetdir}/vendors." - mkdir -p ${targetdir}/vendors - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/vendors failed" - return 1 - fi -fi -chmod u+w ${targetdir}/vendors - -echo "[ops_custom]upgrade framework" -upgrade framework -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op proto" -upgrade op_proto -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade version.info" -upgrade_file version.info -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op impl" -delete_optiling_file op_impl -upgrade op_impl -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op api" -upgrade op_api -if [ $? -ne 0 ];then - exit 1 -fi - -upgrade_proto -if [ $? -ne 0 ];then - exit 1 -fi - -# set the set_env.bash -if [ -n "${INSTALL_PATH}" ] && [ -d ${INSTALL_PATH} ]; then - _ASCEND_CUSTOM_OPP_PATH=${targetdir}/${vendordir} - bin_path="${_ASCEND_CUSTOM_OPP_PATH}/bin" - set_env_variable="#!/bin/bash\nexport ASCEND_CUSTOM_OPP_PATH=${_ASCEND_CUSTOM_OPP_PATH}:\${ASCEND_CUSTOM_OPP_PATH}" - if [ ! -d ${bin_path} ]; then - mkdir -p ${bin_path} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${bin_path} failed" - exit 1 - fi - fi - echo -e ${set_env_variable} > ${bin_path}/set_env.bash - if [ $? -ne 0 ]; then - log "[ERROR] write ASCEND_CUSTOM_OPP_PATH to set_env.bash failed" - exit 1 - else - log "[INFO] using requirements: when custom module install finished or before you run the custom module, \ - execute the command [ source ${bin_path}/set_env.bash ] to set the environment path" - fi -else - config_file=${targetdir}/vendors/config.ini - if [ ! -f ${config_file} ]; then - touch ${config_file} - chmod 640 ${config_file} - echo "load_priority=$vendor_name" > ${config_file} - if [ $? -ne 0 ];then - echo "echo load_priority failed" - exit 1 - fi - else - found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" - found_vendor=$(echo $found_vendors | sed "s/\<$vendor_name\>//g" | tr ',' ' ') - vendor=$(echo $found_vendor | tr -s ' ' ',') - if [ "$vendor" != "" ]; then - sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" - fi - fi -fi - -chmod u-w ${targetdir}/vendors - -if [ -d ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/ ]; then - chmod -R 440 ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/* >/dev/null 2>&1 -fi -if [ -f ${targetdir}/ascend_install.info ]; then - chmod -R 440 ${targetdir}/ascend_install.info -fi -if [ -f ${targetdir}/scene.info ]; then - chmod -R 440 ${targetdir}/scene.info -fi -if [ -f ${targetdir}/version.info ]; then - chmod -R 440 ${targetdir}/version.info -fi - -echo "SUCCESS" -exit 0 - diff --git a/atb_operator/AtbAddSample/AddOp/scripts/upgrade.sh b/atb_operator/AtbAddSample/AddOp/scripts/upgrade.sh deleted file mode 100644 index 34801efc4..000000000 --- a/atb_operator/AtbAddSample/AddOp/scripts/upgrade.sh +++ /dev/null @@ -1,151 +0,0 @@ -#!/bin/bash -vendor_name=customize -targetdir=/usr/local/Ascend/opp -target_custom=0 - -sourcedir=$PWD/packages -vendordir=vendors/$vendor_name - -log() { - cur_date=`date +"%Y-%m-%d %H:%M:%S"` - echo "[runtime] [$cur_date] "$1 -} - -if [[ "x${ASCEND_OPP_PATH}" == "x" ]];then - log "[ERROR] env ASCEND_OPP_PATH no exist" - exit 1 -fi - -targetdir=${ASCEND_OPP_PATH} - -if [ ! -d $targetdir ];then - log "[ERROR] $targetdir no exist" - exit 1 -fi - -upgrade() -{ - if [ ! -d ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 files" - return 0 - fi - - if [ ! -d ${targetdir}/$vendordir/$1 ];then - log "[INFO] create ${targetdir}/$vendordir/$1." - mkdir -p ${targetdir}/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/$1 failed" - return 1 - fi - else - vendor_installed_dir=$(ls "$targetdir/vendors" 2> /dev/null) - for i in $vendor_installed_dir;do - vendor_installed_file=$(ls "$vendor_installed_dir/$vendor_name/$i" 2> /dev/null) - if [ "$i" = "$vendor_name" ] && [ "$vendor_installed_file" != "" ]; then - echo "[INFO]: $vendor_name custom opp package has been installed on the path $vendor_installed_dir, \ - you want to Overlay Installation , please enter:[o]; \ - or replace directory installation , please enter: [r]; \ - or not install , please enter:[n]." - fi - while true - do - read mrn - if [ "$mrn" = m ]; then - break - elif [ "$mrn" = r ]; then - [ -n "$vendor_installed_file"] && rm -rf "$vendor_installed_file" - break - elif [ "$mrn" = n ]; then - return 0 - else - echo "[WARNING]: Input error, please input m or r or n to choose!" - fi - done - done - log "[INFO] replace old ops $1 files ......" - fi - - log "copy new ops $1 files ......" - cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 files failed" - return 1 - fi - - return 0 -} - -upgrade_file() -{ - if [ ! -e ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 file" - return 0 - fi - - log "copy new $1 files ......" - cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 file failed" - return 1 - fi - - return 0 -} - -log "[INFO] copy uninstall sh success" - -echo "[ops_custom]upgrade framework" -upgrade framework -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op proto" -upgrade op_proto -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op impl" -upgrade op_impl -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op api" -upgrade op_api -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade version.info" -upgrade_file version.info -if [ $? -ne 0 ];then - exit 1 -fi - -config_file=${targetdir}/vendors/config.ini -found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" -found_vendor=$(echo $found_vendors | sed "s/\<$vendor_name\>//g" | tr ',' ' ') -vendor=$(echo $found_vendor | tr -s ' ' ',') -if [ "$vendor" != "" ]; then - sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" -fi - -changemode() -{ - if [ -d ${targetdir} ];then - chmod -R 550 ${targetdir}>/dev/null 2>&1 - fi - - return 0 -} -echo "[ops_custom]changemode..." -#changemode -if [ $? -ne 0 ];then - exit 1 -fi - -echo "SUCCESS" -exit 0 - -- Gitee From d830852e0952ac5d795cecd865644b506da31ee1 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Mon, 31 Mar 2025 06:22:46 +0000 Subject: [PATCH 05/26] =?UTF-8?q?=E5=88=A0=E9=99=A4=E6=96=87=E4=BB=B6=20at?= =?UTF-8?q?b=5Foperator/AtbAddSample/AtbAddSample?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .../AtbAddSample/AddAscendC/CMakeLists.txt | 69 -- .../AtbAddSample/AddAscendC/CMakePresets.json | 63 -- .../AtbAddSample/AddAscendC/README.md | 240 ----- .../AtbAddSample/AddAscendC/build.sh | 37 - .../AddAscendC/cmake/config.cmake | 25 - .../AtbAddSample/AddAscendC/cmake/func.cmake | 228 ----- .../AtbAddSample/AddAscendC/cmake/intf.cmake | 26 - .../AddAscendC/cmake/makeself.cmake | 17 - .../AddAscendC/cmake/util/__init__.py | 8 - .../cmake/util/ascendc_bin_param_build.py | 129 --- .../cmake/util/ascendc_impl_build.py | 446 ---------- .../cmake/util/ascendc_ops_config.py | 114 --- .../cmake/util/ascendc_replay_build.py | 65 -- .../cmake/util/batch_replay_impl.temp | 117 --- .../cmake/util/code_channel_infer.py | 115 --- .../AddAscendC/cmake/util/const_var.py | 33 - .../cmake/util/gen_impl_and_mrege_json.sh | 57 -- .../AddAscendC/cmake/util/gen_ops_filter.sh | 62 -- .../AddAscendC/cmake/util/gen_version_info.sh | 6 - .../AddAscendC/cmake/util/insert_op_info.py | 36 - .../cmake/util/insert_simplified_keys.py | 248 ------ .../AddAscendC/cmake/util/kernel_entry.py | 115 --- .../AddAscendC/cmake/util/kernel_impl.temp | 10 - .../AddAscendC/cmake/util/makeself/COPYING | 339 -------- .../AddAscendC/cmake/util/makeself/README.md | 246 ------ .../AddAscendC/cmake/util/makeself/VERSION | 1 - .../cmake/util/makeself/make-release.sh | 9 - .../cmake/util/makeself/makeself-header.sh | 660 -------------- .../AddAscendC/cmake/util/makeself/makeself.1 | 110 --- .../cmake/util/makeself/makeself.lsm | 16 - .../cmake/util/makeself/makeself.sh | 822 ------------------ .../cmake/util/makeself/run-tests.sh | 8 - .../cmake/util/merge_aicpu_info_json.sh | 31 - .../AddAscendC/cmake/util/opdesc_parser.py | 260 ------ .../cmake/util/parse_ini_to_json.py | 338 ------- .../AddAscendC/cmake/util/preset_parse.py | 23 - .../AddAscendC/cmake/util/replay_codegen.py | 105 --- .../AddAscendC/cmake/util/replay_impl.temp | 120 --- .../cmake/util/tiling_data_def_build.py | 84 -- .../AddAscendC/framework/CMakeLists.txt | 11 - .../framework/tf_plugin/CMakeLists.txt | 14 - .../tf_plugin/tensorflow_add_custom_plugin.cc | 23 - .../AddAscendC/op_host/CMakeLists.txt | 82 -- .../AddAscendC/op_host/add_custom.cpp | 117 --- .../AddAscendC/op_host/add_custom_tiling.h | 22 - .../AddAscendC/op_kernel/CMakeLists.txt | 61 -- .../AddAscendC/op_kernel/add_custom.cpp | 134 --- .../AddAscendC/scripts/install.sh | 318 ------- .../AddAscendC/scripts/upgrade.sh | 151 ---- .../AddOperationATBPlugin/CMakeLists.txt | 21 - .../aclnn_add_operation.cpp | 142 --- .../aclnn_add_operation.h | 56 -- .../AddOperationATBPlugin/build.sh | 33 - .../AddOperationTest/CMakeLists.txt | 40 - .../AddOperationTest/script/gen_data.py | 20 - .../AddOperationTest/script/run.sh | 52 -- .../AddOperationTest/script/verify_result.py | 28 - .../AddOperationTest/src/main.cpp | 217 ----- .../AtbAddSample/AddOperationTest/src/main.h | 55 -- .../AtbAddSample/AtbAddSample/readme.md | 81 -- 60 files changed, 7116 deletions(-) delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakePresets.json delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/README.md delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/build.sh delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/config.cmake delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/func.cmake delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/intf.cmake delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/makeself.cmake delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/__init__.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/const_var.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/README.md delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/preset_parse.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom.cpp delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/install.sh delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/upgrade.sh delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h delete mode 100755 atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/build.sh delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/gen_data.py delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/run.sh delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/verify_result.py delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.cpp delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.h delete mode 100644 atb_operator/AtbAddSample/AtbAddSample/readme.md diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakeLists.txt deleted file mode 100644 index 584132d80..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakeLists.txt +++ /dev/null @@ -1,69 +0,0 @@ -cmake_minimum_required(VERSION 3.16.0) -project(opp) -if(ENABLE_CROSS_COMPILE) - if(${CMAKE_SYSTEM_PROCESSOR} STREQUAL x86_64) - set(CROSS_COMPILE_PLATFORM aarch64) - else() - set(CROSS_COMPILE_PLATFORM x86_64) - endif() - set(PLATFORM ${CMAKE_SYSTEM_PROCESSOR}) - set(CMAKE_COMPILE_COMPILER_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/linux/${CROSS_COMPILE_PLATFORM}/) - set(CMAKE_COMPILE_RUNTIME_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/${CROSS_COMPILE_PLATFORM}/) - set(CMAKE_SYSTEM_PROCESSOR ${CROSS_COMPILE_PLATFORM}) - set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) - set(CMAKE_CXX_COMPILER ${CMAKE_CROSS_PLATFORM_COMPILER}) -else() - set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) -endif() - -include(cmake/config.cmake) -include(cmake/func.cmake) -include(cmake/intf.cmake) - -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/framework) - add_subdirectory(framework) -endif() -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_host) - add_subdirectory(op_host) -endif() -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_kernel) - add_subdirectory(op_kernel) -endif() -if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) - add_subdirectory(testcases) -endif() - -# modify vendor_name in install.sh and upgrade.sh -add_custom_command(OUTPUT ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh - COMMAND mkdir -p ${CMAKE_BINARY_DIR}/scripts - COMMAND cp -r ${CMAKE_SOURCE_DIR}/scripts/* ${CMAKE_BINARY_DIR}/scripts/ - COMMAND sed -i "s/vendor_name=customize/vendor_name=${vendor_name}/g" ${CMAKE_BINARY_DIR}/scripts/* -) -add_custom_target(modify_vendor ALL DEPENDS ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh) -install(DIRECTORY ${CMAKE_BINARY_DIR}/scripts/ DESTINATION . FILE_PERMISSIONS OWNER_EXECUTE OWNER_READ GROUP_READ) - -install(FILES ${CMAKE_SOURCE_DIR}/custom.proto DESTINATION packages OPTIONAL) - -get_system_info(SYSTEM_INFO) - -# gen version.info -add_custom_target(gen_version_info ALL - COMMAND bash ${CMAKE_CURRENT_SOURCE_DIR}/cmake/util/gen_version_info.sh ${ASCEND_CANN_PACKAGE_PATH} ${CMAKE_CURRENT_BINARY_DIR} -) - -install(FILES ${CMAKE_CURRENT_BINARY_DIR}/version.info - DESTINATION packages/vendors/${vendor_name}/) - -# CPack config -set(CPACK_PACKAGE_NAME ${CMAKE_PROJECT_NAME}) -set(CPACK_PACKAGE_VERSION ${CMAKE_PROJECT_VERSION}) -set(CPACK_PACKAGE_DESCRIPTION "CPack opp project") -set(CPACK_PACKAGE_DESCRIPTION_SUMMARY "CPack opp project") -set(CPACK_PACKAGE_DIRECTORY ${CMAKE_INSTALL_PREFIX}) -set(CPACK_PACKAGE_FILE_NAME "custom_opp_${SYSTEM_INFO}.run") -set(CPACK_GENERATOR External) -set(CPACK_CMAKE_GENERATOR "Unix Makefiles") -set(CPACK_EXTERNAL_ENABLE_STAGING TRUE) -set(CPACK_EXTERNAL_PACKAGE_SCRIPT ${CMAKE_SOURCE_DIR}/cmake/makeself.cmake) -set(CPACK_EXTERNAL_BUILT_PACKAGES ${CPACK_PACKAGE_DIRECTORY}/_CPack_Packages/Linux/External/${CPACK_PACKAGE_FILE_NAME}/${CPACK_PACKAGE_FILE_NAME}) -include(CPack) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakePresets.json b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakePresets.json deleted file mode 100644 index ddec9431b..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/CMakePresets.json +++ /dev/null @@ -1,63 +0,0 @@ -{ - "version": 1, - "cmakeMinimumRequired": { - "major": 3, - "minor": 19, - "patch": 0 - }, - "configurePresets": [ - { - "name": "default", - "displayName": "Default Config", - "description": "Default build using Unix Makefiles generator", - "generator": "Unix Makefiles", - "binaryDir": "${sourceDir}/build_out", - "cacheVariables": { - "CMAKE_BUILD_TYPE": { - "type": "STRING", - "value": "Release" - }, - "ENABLE_SOURCE_PACKAGE": { - "type": "BOOL", - "value": "True" - }, - "ENABLE_BINARY_PACKAGE": { - "type": "BOOL", - "value": "True" - }, - "ASCEND_COMPUTE_UNIT": { - "type": "STRING", - "value": "ascend310b;ascend910b" - }, - "ENABLE_TEST": { - "type": "BOOL", - "value": "True" - }, - "vendor_name": { - "type": "STRING", - "value": "customize" - }, - "ASCEND_CANN_PACKAGE_PATH": { - "type": "PATH", - "value": "/usr/local/Ascend/ascend-toolkit/latest" - }, - "ASCEND_PYTHON_EXECUTABLE": { - "type": "STRING", - "value": "python3" - }, - "CMAKE_INSTALL_PREFIX": { - "type": "PATH", - "value": "${sourceDir}/build_out" - }, - "ENABLE_CROSS_COMPILE": { - "type": "BOOL", - "value": "False" - }, - "CMAKE_CROSS_PLATFORM_COMPILER": { - "type": "PATH", - "value": "/usr/bin/aarch64-linux-gnu-g++" - } - } - } - ] -} diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/README.md b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/README.md deleted file mode 100644 index da59a83e7..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/README.md +++ /dev/null @@ -1,240 +0,0 @@ -<<<<<<< HEAD -## 概述 -本样例基于AddCustom算子工程,介绍了单算子工程及单算子调用。 - -## 算子描述 -======= -# 概述 -本样例基于AddCustom算子工程,介绍了单算子工程及aclnn接口说明。 - -## 算子工程介绍 -### 算子描述 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd -Add算子实现了两个数据相加,返回相加结果的功能,该算子实现了非32B对齐shape下的算子功能,完成了算子的泛化实现。 -对应的数学表达式为: -``` -z = x + y -``` -<<<<<<< HEAD -## 算子规格描述 -======= -### 算子规格描述 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd - - - - - - - - - - - -
算子类型(OpType)Add
算子输入nameshapedata typeformat
x-float32,float16,int32,int8ND
y-float32,float16,int32,int8ND
算子输出z-float32,float16,int32,int8ND
核函数名add_custom
- - -<<<<<<< HEAD -## 算子工程介绍 -======= -### 算子工程文件介绍 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd -其中,算子工程目录AddCustom包含算子实现的模板文件、编译脚本等,如下所示: -``` -├── AddCustom //Add自定义算子工程 -│ ├── cmake -│ ├── framework // 算子插件实现文件目录,单算子模型文件的生成不依赖算子适配插件,无需关注 -│ ├── op_host // host侧实现文件 -│ ├── op_kernel // kernel侧实现文件 -│ ├── scripts // 自定义算子工程打包相关脚本所在目录 -│ ├── build.sh // 编译入口脚本 -│ ├── CMakeLists.txt // 算子工程的CMakeLists.txt -│ └── CMakePresets.json // 编译配置项 -``` -<<<<<<< HEAD -CANN软件包中提供了工程创建工具msopgen,AddCustom算子工程可通过AddCustom.json自动创建,具体请参考[Ascend C算子开发](https://hiascend.com/document/redirect/CannCommunityOpdevAscendC)>算子开发>算子开发工程>基于自定义算子工程的算子开发>创建算子工程 章节。 -## 编译运行样例算子 -======= - -### 编译运行样例算子 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd -针对自定义算子工程,编译运行包含如下步骤: -- 编译自定义算子工程生成算子安装包; -- 安装自定义算子到算子库中; -- 调用执行自定义算子; - -详细操作如下所示。 -<<<<<<< HEAD -### 1. 获取源码包 -编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 -### 2. 编译算子工程 -======= -#### 1. 获取源码包 -编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 -#### 2. 编译算子工程 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd - 编译自定义算子工程,构建生成自定义算子包。 - - - 执行如下命令,切换到算子工程AddCustom目录。 - - ```bash - cd ${git_clone_path}/samples/atb_operator/AtbAddSample/AddOp - ``` - - - 修改CMakePresets.json中ASCEND_CANN_PACKAGE_PATH为CANN软件包安装后的实际路径。 - - - ```json - { - …… - "configurePresets": [ - { - …… - "ASCEND_CANN_PACKAGE_PATH": { - "type": "PATH", - "value": "/usr/local/Ascend/ascend-toolkit/latest" //请替换为CANN软件包安装后的实际路径。eg:/home/HwHiAiUser/Ascend/ascend-toolkit/latest - }, - …… - } - ] - } - ``` - - 在算子工程AddOp目录下执行如下命令,进行算子工程编译。 - - ```bash - ./build.sh - ``` -编译成功后,会在当前目录下创建build_out目录,并在build_out目录下生成自定义算子安装包custom_opp_\_\.run,例如“custom_opp_ubuntu_x86_64.run”。 - -备注:如果要使用dump调试功能,需要移除op_host内和CMakeLists.txt内的Atlas 200/500 A2 推理产品的配置。 - -<<<<<<< HEAD -### 3. 部署算子包 -======= -#### 3. 部署算子包 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd - -执行如下命令,在自定义算子安装包所在路径下,安装自定义算子包。 - ```bash - cd build_out - ./custom_opp__.run - ``` -命令执行成功后,自定义算子包中的相关文件将部署至当前环境的OPP算子库的vendors/customize目录中。 - -<<<<<<< HEAD -### 4. 配置环境变量 -======= -#### 4. 配置环境变量 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd - - 请根据当前环境上CANN开发套件包的[安装方式](https://hiascend.com/document/redirect/CannCommunityInstSoftware),选择对应配置环境变量的命令。 - - 默认路径,root用户安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=/usr/local/Ascend/ascend-toolkit/latest - ``` - - 默认路径,非root用户安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=$HOME/Ascend/ascend-toolkit/latest - ``` - - 指定路径install_path,安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=${install_path}/ascend-toolkit/latest - ``` - -## aclnn接口说明 - -自定义算子编译部署后,会自动生成单算子API,可以直接在应用程序中调用。算子API的形式一般定义为“两段式接口”,形如: - ```cpp - aclnnStatus aclnnAddCustomGetWorkspaceSize(const aclTensor *x, const aclTensor *y, const alcTensor *out, uint64_t workspaceSize, aclOpExecutor **executor); - aclnnStatus aclnnAddCustom(void *workspace, int64_t workspaceSize, aclOpExecutor **executor, aclrtStream stream); - ``` -其中aclnnAddCustomGetWorkspaceSize为第一段接口,主要用于计算本次API调用计算过程中需要多少的workspace内存。获取到本次API计算需要的workspace大小之后,按照workspaceSize大小申请Device侧内存,然后调用第二段接口aclnnAddCustom执行计算。具体参考[AscendCL单算子调用](https://hiascend.com/document/redirect/CannCommunityAscendCInVorkSingleOp)>单算子API执行 章节。 - -<<<<<<< HEAD -======= -### 功能描述 -* 算子功能:完成加法计算。 -* 计算公式: -``` -z = x + y -``` - -#### 参数说明 -##### aclnnAddCustomGetWorkspaceSize: - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
接口参数描述
aclnnAddCustomGetWorkspaceSizex输入x的Tensor,支持flaot/half/int8/int32类型,ND排布格式
y输入y的Tensor,支持flaot/half/int8/int32类型,ND排布格式
out输出z的Tensor,支持flaot/half/int8/int32类型,ND排布格式
workspaceSize返回需要在Device侧申请的workspace大小
executor返回op执行器,包含了算子计算流程
- - -| 返回值 | 描述 | -| ---- | ---- | -| ACL_SUCCESS | 成功 | -| ACL_ERROR | 失败 | - -##### aclnnAddCustom - - - - - - - - - - - - - - - - - - - - - - - -
接口参数描述
aclnnAddCustomworkspace输入Device侧申请的workspace内存
workspaceSizeworkspace大小
executor输入op执行器,包含了算子计算流程
stream输入执行流
- -| 返回值 | 描述 | -| ---- | ---- | -| ACL_SUCCESS | 成功 | -| ACL_ERROR | 失败 | - ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd - -## 更新说明 -| 时间 | 更新事项 | -|----|------| -<<<<<<< HEAD -| 2025/03/27 | 修改不同用户环境配置 | -======= -| 2025/03/27 | 新增README | ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/build.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/build.sh deleted file mode 100755 index 4be96d7d8..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/build.sh +++ /dev/null @@ -1,37 +0,0 @@ -#!/bin/bash -script_path=$(realpath $(dirname $0)) - - -mkdir -p build_out -rm -rf build_out/* -cd build_out - -cmake_version=$(cmake --version | grep "cmake version" | awk '{print $3}') -if [ "$cmake_version" \< "3.19.0" ] ; then - opts=$(python3 $script_path/cmake/util/preset_parse.py $script_path/CMakePresets.json) - echo $opts - cmake .. $opts -else - cmake .. --preset=default -fi -target=package -if [ "$1"x != ""x ]; then target=$1; fi - -cmake --build . --target $target -j16 -if [ $? -ne 0 ]; then exit 1; fi - -if [ $target = "package" ]; then - if test -d ./op_kernel/binary ; then - ./cust*.run - if [ $? -ne 0 ]; then exit 1; fi - cmake --build . --target binary -j16 - if [ $? -ne 0 ]; then exit 1; fi - cmake --build . --target $target -j16 - fi -fi - -# for debug -# cd build_out -# make -# cpack -# verbose append -v diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/config.cmake b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/config.cmake deleted file mode 100755 index 886119daa..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/config.cmake +++ /dev/null @@ -1,25 +0,0 @@ - -set(CMAKE_CXX_FLAGS_DEBUG "") -set(CMAKE_CXX_FLAGS_RELEASE "") - -if (NOT DEFINED vendor_name) - set(vendor_name customize CACHE STRING "") -endif() -if (NOT DEFINED ASCEND_CANN_PACKAGE_PATH) - set(ASCEND_CANN_PACKAGE_PATH /usr/local/Ascend/latest CACHE PATH "") -endif() -if (NOT DEFINED ASCEND_PYTHON_EXECUTABLE) - set(ASCEND_PYTHON_EXECUTABLE python3 CACHE STRING "") -endif() -if (NOT DEFINED ASCEND_COMPUTE_UNIT) - message(FATAL_ERROR "ASCEND_COMPUTE_UNIT not set in CMakePreset.json ! -") -endif() -set(ASCEND_TENSOR_COMPILER_PATH ${ASCEND_CANN_PACKAGE_PATH}/compiler) -set(ASCEND_CCEC_COMPILER_PATH ${ASCEND_TENSOR_COMPILER_PATH}/ccec_compiler/bin) -set(ASCEND_AUTOGEN_PATH ${CMAKE_BINARY_DIR}/autogen) -set(ASCEND_FRAMEWORK_TYPE tensorflow) -file(MAKE_DIRECTORY ${ASCEND_AUTOGEN_PATH}) -set(CUSTOM_COMPILE_OPTIONS "custom_compile_options.ini") -execute_process(COMMAND rm -rf ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} - COMMAND touch ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS}) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/func.cmake b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/func.cmake deleted file mode 100755 index ad187e7d6..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/func.cmake +++ /dev/null @@ -1,228 +0,0 @@ - -function(get_system_info SYSTEM_INFO) - if (UNIX) - execute_process(COMMAND grep -i ^id= /etc/os-release OUTPUT_VARIABLE TEMP) - string(REGEX REPLACE "\n|id=|ID=|\"" "" SYSTEM_NAME ${TEMP}) - set(${SYSTEM_INFO} ${SYSTEM_NAME}_${CMAKE_SYSTEM_PROCESSOR} PARENT_SCOPE) - elseif (WIN32) - message(STATUS "System is Windows. Only for pre-build.") - else () - message(FATAL_ERROR "${CMAKE_SYSTEM_NAME} not support.") - endif () -endfunction() - -function(opbuild) - message(STATUS "Opbuild generating sources") - cmake_parse_arguments(OPBUILD "" "OUT_DIR;PROJECT_NAME;ACCESS_PREFIX" "OPS_SRC" ${ARGN}) - execute_process(COMMAND ${CMAKE_COMPILE} -g -fPIC -shared -std=c++11 ${OPBUILD_OPS_SRC} -D_GLIBCXX_USE_CXX11_ABI=0 - -I ${ASCEND_CANN_PACKAGE_PATH}/include -L ${ASCEND_CANN_PACKAGE_PATH}/lib64 -lexe_graph -lregister -ltiling_api - -o ${OPBUILD_OUT_DIR}/libascend_all_ops.so - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("build ops lib info: ${EXEC_INFO}") - message("build ops lib error: ${EXEC_ERROR}") - message(FATAL_ERROR "opbuild run failed!") - endif() - set(proj_env "") - set(prefix_env "") - if (NOT "${OPBUILD_PROJECT_NAME}x" STREQUAL "x") - set(proj_env "OPS_PROJECT_NAME=${OPBUILD_PROJECT_NAME}") - endif() - if (NOT "${OPBUILD_ACCESS_PREFIX}x" STREQUAL "x") - set(prefix_env "OPS_DIRECT_ACCESS_PREFIX=${OPBUILD_ACCESS_PREFIX}") - endif() - execute_process(COMMAND ${proj_env} ${prefix_env} ${ASCEND_CANN_PACKAGE_PATH}/toolkit/tools/opbuild/op_build - ${OPBUILD_OUT_DIR}/libascend_all_ops.so ${OPBUILD_OUT_DIR} - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("opbuild ops info: ${EXEC_INFO}") - message("opbuild ops error: ${EXEC_ERROR}") - endif() - message(STATUS "Opbuild generating sources - done") -endfunction() - -function(add_ops_info_target) - cmake_parse_arguments(OPINFO "" "TARGET;OPS_INFO;OUTPUT;INSTALL_DIR" "" ${ARGN}) - get_filename_component(opinfo_file_path "${OPINFO_OUTPUT}" DIRECTORY) - add_custom_command(OUTPUT ${OPINFO_OUTPUT} - COMMAND mkdir -p ${opinfo_file_path} - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/parse_ini_to_json.py - ${OPINFO_OPS_INFO} ${OPINFO_OUTPUT} - ) - add_custom_target(${OPINFO_TARGET} ALL - DEPENDS ${OPINFO_OUTPUT} - ) - install(FILES ${OPINFO_OUTPUT} - DESTINATION ${OPINFO_INSTALL_DIR} - ) -endfunction() - -function(add_ops_compile_options OP_TYPE) - cmake_parse_arguments(OP_COMPILE "" "OP_TYPE" "COMPUTE_UNIT;OPTIONS" ${ARGN}) - file(APPEND ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} - "${OP_TYPE},${OP_COMPILE_COMPUTE_UNIT},${OP_COMPILE_OPTIONS}\n") -endfunction() - -function(add_ops_impl_target) - cmake_parse_arguments(OPIMPL "" "TARGET;OPS_INFO;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) - add_custom_command(OUTPUT ${OPIMPL_OUT_DIR}/.impl_timestamp - COMMAND mkdir -m 700 -p ${OPIMPL_OUT_DIR}/dynamic - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py - ${OPIMPL_OPS_INFO} - \"${OPIMPL_OPS_BATCH}\" \"${OPIMPL_OPS_ITERATE}\" - ${OPIMPL_IMPL_DIR} - ${OPIMPL_OUT_DIR}/dynamic - ${ASCEND_AUTOGEN_PATH} - - COMMAND rm -rf ${OPIMPL_OUT_DIR}/.impl_timestamp - COMMAND touch ${OPIMPL_OUT_DIR}/.impl_timestamp - DEPENDS ${OPIMPL_OPS_INFO} - ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py - ) - add_custom_target(${OPIMPL_TARGET} ALL - DEPENDS ${OPIMPL_OUT_DIR}/.impl_timestamp) - if (${ENABLE_SOURCE_PACKAGE}) - install(DIRECTORY ${OPIMPL_OUT_DIR}/dynamic - DESTINATION ${OPIMPL_INSTALL_DIR} - ) - endif() -endfunction() - -function(add_ops_replay_targets) - cmake_parse_arguments(OPREPLAY "" "OPS_INFO;COMPUTE_UNIT;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) - # ccec compile options - set(ccec_base_opts -c -O2 --cce-aicore-only -mllvm -cce-aicore-function-stack-size=16000 - -mllvm -cce-aicore-record-overflow=false -std=c++17) - set(ccec_extopts_ascend310p --cce-aicore-arch=dav-m200 -mllvm -cce-aicore-fp-ceiling=2) - set(ccec_extopts_ascend910 --cce-aicore-arch=dav-c100) - set(ccec_extopts_ascend910b --cce-aicore-arch=dav-c220-cube) - file(MAKE_DIRECTORY ${OPREPLAY_OUT_DIR}) - execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_replay_build.py - ${OPREPLAY_OPS_INFO} - "${OPREPLAY_OPS_BATCH}" "${OPREPLAY_OPS_ITERATE}" - ${OPREPLAY_IMPL_DIR} - ${OPREPLAY_OUT_DIR} - ${OPREPLAY_COMPUTE_UNIT} - ) - file(GLOB replay_kernel_entries ${OPREPLAY_OUT_DIR}/*.cce) - if (NOT "${replay_kernel_entries}x" STREQUAL "x") - foreach(replay_kernel_file ${replay_kernel_entries}) - get_filename_component(replay_kernel_file_name "${replay_kernel_file}" NAME) - string(REPLACE "_entry.cce" "" op_kerne_name ${replay_kernel_file_name}) - file(GLOB replay_lib_src ${OPREPLAY_OUT_DIR}/${op_kerne_name}*.cpp) - set(OP_TILING_DATA_H_PATH ${OPREPLAY_OUT_DIR}/${op_kerne_name}_tiling_data.h) - add_library(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} SHARED ${replay_lib_src}) - if(EXISTS ${OP_TILING_DATA_H_PATH}) - target_compile_options(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE - -include ${OP_TILING_DATA_H_PATH} - ) - endif() - target_compile_definitions(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE - ${op_kerne_name}=${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} - ) - target_compile_options(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE - -D__ASCENDC_REPLAY__ - ) - target_link_libraries(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE intf_pub - tikreplaylib::${OPREPLAY_COMPUTE_UNIT} - register - ) - add_custom_command(OUTPUT ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o - COMMAND ccec ${ccec_base_opts} ${ccec_extopts_${OPREPLAY_COMPUTE_UNIT}} ${replay_kernel_file} - -o ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o - DEPENDS ${replay_kernel_file} - ) - add_custom_target(replay_kernel_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} ALL - DEPENDS ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o - ) - install(TARGETS replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_replay - ) - install(FILES ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o - DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_replay - ) - endforeach() - endif() -endfunction() - -function(add_npu_support_target) - cmake_parse_arguments(NPUSUP "" "TARGET;OPS_INFO_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) - get_filename_component(npu_sup_file_path "${NPUSUP_OUT_DIR}" DIRECTORY) - add_custom_command(OUTPUT ${NPUSUP_OUT_DIR}/npu_supported_ops.json - COMMAND mkdir -p ${NPUSUP_OUT_DIR} - COMMAND ${CMAKE_SOURCE_DIR}/cmake/util/gen_ops_filter.sh - ${NPUSUP_OPS_INFO_DIR} - ${NPUSUP_OUT_DIR} - ) - add_custom_target(npu_supported_ops ALL - DEPENDS ${NPUSUP_OUT_DIR}/npu_supported_ops.json - ) - install(FILES ${NPUSUP_OUT_DIR}/npu_supported_ops.json - DESTINATION ${NPUSUP_INSTALL_DIR} - ) -endfunction() - -function(add_bin_compile_target) - cmake_parse_arguments(BINCMP "" "TARGET;OPS_INFO;COMPUTE_UNIT;IMPL_DIR;ADP_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/src) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/gen) - execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_bin_param_build.py - ${BINCMP_OPS_INFO} ${BINCMP_OUT_DIR}/gen ${BINCMP_COMPUTE_UNIT} - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("ops binary compile scripts gen info: ${EXEC_INFO}") - message("ops binary compile scripts gen error: ${EXEC_ERROR}") - message(FATAL_ERROR "ops binary compile scripts gen failed!") - endif() - if (NOT TARGET binary) - add_custom_target(binary) - endif() - add_custom_target(${BINCMP_TARGET} - COMMAND cp -r ${BINCMP_IMPL_DIR}/*.* ${BINCMP_OUT_DIR}/src - ) - add_custom_target(${BINCMP_TARGET}_gen_ops_config - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/insert_simplified_keys.py -p ${BINCMP_OUT_DIR}/bin - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_ops_config.py -p ${BINCMP_OUT_DIR}/bin - -s ${BINCMP_COMPUTE_UNIT} - ) - add_dependencies(binary ${BINCMP_TARGET}_gen_ops_config) - file(GLOB bin_scripts ${BINCMP_OUT_DIR}/gen/*.sh) - foreach(bin_script ${bin_scripts}) - get_filename_component(bin_file ${bin_script} NAME_WE) - string(REPLACE "-" ";" bin_sep ${bin_file}) - list(GET bin_sep 0 op_type) - list(GET bin_sep 1 op_file) - list(GET bin_sep 2 op_index) - if (NOT TARGET ${BINCMP_TARGET}_${op_file}_copy) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file}) - add_custom_target(${BINCMP_TARGET}_${op_file}_copy - COMMAND cp ${BINCMP_ADP_DIR}/${op_file}.py ${BINCMP_OUT_DIR}/src/${op_type}.py - ) - install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} - DESTINATION ${BINCMP_INSTALL_DIR}/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) - install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json - DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT}/ OPTIONAL - ) - endif() - add_custom_target(${BINCMP_TARGET}_${op_file}_${op_index} - COMMAND export HI_PYTHON=${ASCEND_PYTHON_EXECUTABLE} && bash ${bin_script} ${BINCMP_OUT_DIR}/src/${op_type}.py ${BINCMP_OUT_DIR}/bin/${op_file} - WORKING_DIRECTORY ${BINCMP_OUT_DIR} - ) - add_dependencies(${BINCMP_TARGET}_${op_file}_${op_index} ${BINCMP_TARGET} ${BINCMP_TARGET}_${op_file}_copy) - add_dependencies(${BINCMP_TARGET}_gen_ops_config ${BINCMP_TARGET}_${op_file}_${op_index}) - endforeach() - install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json - DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) -endfunction() diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/intf.cmake b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/intf.cmake deleted file mode 100755 index 2f362c396..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/intf.cmake +++ /dev/null @@ -1,26 +0,0 @@ - -add_library(intf_pub INTERFACE) -target_compile_options(intf_pub INTERFACE - -fPIC - -fvisibility=hidden - -fvisibility-inlines-hidden - $<$:-O2> - $<$:-O0 -g> - $<$:-std=c++11> - $<$,$>:-ftrapv -fstack-check> - $<$:-pthread -Wfloat-equal -Wshadow -Wformat=2 -Wno-deprecated -Wextra> - $,-fstack-protector-strong,-fstack-protector-all> -) -target_compile_definitions(intf_pub INTERFACE - _GLIBCXX_USE_CXX11_ABI=0 - $<$:_FORTIFY_SOURCE=2> -) -target_include_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/include) -target_link_options(intf_pub INTERFACE - $<$,EXECUTABLE>:-pie> - $<$:-s> - -Wl,-z,relro - -Wl,-z,now - -Wl,-z,noexecstack -) -target_link_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/lib64) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/makeself.cmake b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/makeself.cmake deleted file mode 100755 index 48c565bfb..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/makeself.cmake +++ /dev/null @@ -1,17 +0,0 @@ -execute_process(COMMAND chmod +x ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh) -execute_process(COMMAND ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh - --header ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself-header.sh - --help-header ./help.info - --gzip --complevel 4 --nomd5 --sha256 - ./ ${CPACK_PACKAGE_FILE_NAME} "version:1.0" ./install.sh - WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} - RESULT_VARIABLE EXEC_RESULT - ERROR_VARIABLE EXEC_ERROR -) -if (NOT "${EXEC_RESULT}x" STREQUAL "0x") - message(FATAL_ERROR "CPack Command error: ${EXEC_RESULT}\n${EXEC_ERROR}") -endif() -execute_process(COMMAND cp ${CPACK_EXTERNAL_BUILT_PACKAGES} ${CPACK_PACKAGE_DIRECTORY}/ - COMMAND echo "Copy ${CPACK_EXTERNAL_BUILT_PACKAGES} to ${CPACK_PACKAGE_DIRECTORY}/" - WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} -) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/__init__.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/__init__.py deleted file mode 100755 index c4ddc893a..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/__init__.py +++ /dev/null @@ -1,8 +0,0 @@ -#!/usr/bin/env python -# -*- coding: utf-8 -*- - -import sys -import os - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) -sys.path.append(PYF_PATH) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py deleted file mode 100755 index decf34544..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py +++ /dev/null @@ -1,129 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import json -import hashlib -import const_var -import opdesc_parser - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - - -class BinParamBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - super().__init__(op_type) - self.soc = '' - self.out_path = '' - - def set_soc_version(self: any, soc: str): - self.soc = soc - - def set_out_path(self: any, out_path: str): - self.out_path = out_path - - def gen_input_json(self: any): - key_map = {} - count = len(self.input_dtype[0].split(',')) - for i in range(0, count): - inputs = [] - outputs = [] - attrs = [] - op_node = {} - for idx in range(0, len(self.input_name)): - idtypes = self.input_dtype[idx].split(',') - ifmts = self.input_fmt[idx].split(',') - itype = self.input_type[idx] - para = {} - para['name'] = self.input_name[idx] - para['index'] = idx - para['dtype'] = idtypes[i] - para['format'] = ifmts[i] - para['paramType'] = itype - para['shape'] = [-2] - if itype == 'dynamic': - inputs.append([para]) - else: - inputs.append(para) - for idx in range(0, len(self.output_name)): - odtypes = self.output_dtype[idx].split(',') - ofmts = self.output_fmt[idx].split(',') - otype = self.output_type[idx] - para = {} - para['name'] = self.output_name[idx] - para['index'] = idx - para['dtype'] = odtypes[i] - para['format'] = ofmts[i] - para['paramType'] = otype - para['shape'] = [-2] - if otype == 'dynamic': - outputs.append([para]) - else: - outputs.append(para) - for attr in self.attr_list: - att = {} - att['name'] = attr - atype = self.attr_val.get(attr).get('type').lower() - atype = atype.replace('list', 'list_') - att['dtype'] = atype - att['value'] = const_var.ATTR_DEF_VAL.get(atype) - attrs.append(att) - op_node['bin_filename'] = '' - op_node['inputs'] = inputs - op_node['outputs'] = outputs - if len(attrs) > 0: - op_node['attrs'] = attrs - param = {} - param['op_type'] = self.op_type - param['op_list'] = [op_node] - objstr = json.dumps(param, indent=' ') - md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() - while key_map.get(md5sum) is not None: - objstr += '1' - md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() - key_map[md5sum] = md5sum - bin_file = self.op_type + '_' + md5sum - op_node['bin_filename'] = bin_file - param_file = os.path.join(self.out_path, bin_file + '_param.json') - param_file = os.path.realpath(param_file) - with os.fdopen(os.open(param_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - json.dump(param, fd, indent=' ') - self._write_buld_cmd(param_file, bin_file, i) - - - def _write_buld_cmd(self: any, param_file: str, bin_file: str, index: int): - hard_soc = const_var.SOC_MAP_EXT.get(self.soc) - if not hard_soc: - hard_soc = soc.capitalize() - name_com = [self.op_type, self.op_file, str(index)] - compile_file = os.path.join(self.out_path, '-'.join(name_com) + '.sh') - compile_file = os.path.realpath(compile_file) - with os.fdopen(os.open(compile_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - fd.write('#!/bin/bash\n') - fd.write('echo "[{}] Generating {} ..."\n'.format(hard_soc, bin_file)) - cmd = const_var.BIN_CMD.format(fun=self.op_intf, soc=hard_soc, param=param_file, impl='""') - fd.write(cmd) - chk = const_var.CHK_CMD.format(res_file=bin_file + '.json') - fd.write(chk) - chk = const_var.CHK_CMD.format(res_file=bin_file + '.o') - fd.write(chk) - fd.write('echo "[{}] Generating {} Done"\n'.format(hard_soc, bin_file)) - - -def gen_bin_param_file(cfgfile: str, out_dir: str, soc: str): - op_descs = opdesc_parser.get_op_desc(cfgfile, [], [], BinParamBuilder, None) - for op_desc in op_descs: - op_desc.set_soc_version(soc) - op_desc.set_out_path(out_dir) - op_desc.gen_input_json() - - -if __name__ == '__main__': - if len(sys.argv) <= 3: - raise RuntimeError('arguments must greater than 3') - gen_bin_param_file(sys.argv[1], sys.argv[2], sys.argv[3]) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py deleted file mode 100755 index 7fe177da1..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py +++ /dev/null @@ -1,446 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import re -import stat -import opdesc_parser -import const_var - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - -IMPL_HEAD = ''' -import os, sys -import ctypes -import json -import shutil -from tbe.common.platform import get_soc_spec -from tbe.common.utils import para_check -from tbe.tikcpp import compile_op, replay_op, check_op_cap, generalize_op_params, get_code_channel, OpInfo -from tbe.common.buildcfg import get_default_build_config -from impl.util.platform_adapter import tbe_register -from tbe.common.buildcfg import get_current_build_config -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - -DTYPE_MAP = {"float32": ["DT_FLOAT", "float"], - "float16": ["DT_FLOAT16", "half"], - "int8": ["DT_INT8", "int8_t"], - "int16": ["DT_INT16", "int16_t"], - "int32": ["DT_INT32", "int32_t"], - "int64": ["DT_INT64", "int64_t"], - "uint1": ["DT_UINT1", "uint8_t"], - "uint8": ["DT_UINT8", "uint8_t"], - "uint16": ["DT_UINT16", "uint16_t"], - "uint32": ["DT_UINT32", "uint32_t"], - "uint64": ["DT_UINT64", "uint64_t"], - "bool": ["DT_BOOL", "bool"], - "double": ["DT_DOUBLE", "double"], - "dual": ["DT_DUAL", "unknown"], - "dual_sub_int8": ["DT_DUAL_SUB_INT8", "unknown"], - "dual_sub_uint8": ["DT_DUAL_SUB_UINT8", "unknown"], - "string": ["DT_STRING", "unknown"], - "complex64": ["DT_COMPLEX64", "unknown"], - "complex128": ["DT_COMPLEX128", "unknown"], - "qint8": ["DT_QINT8", "unknown"], - "qint16": ["DT_QINT16", "unknown"], - "qint32": ["DT_QINT32", "unknown"], - "quint8": ["DT_QUINT8", "unknown"], - "quint16": ["DT_QUINT16", "unknown"], - "resource": ["DT_RESOURCE", "unknown"], - "string_ref": ["DT_STRING_REF", "unknown"], - "int4": ["DT_INT4", "int8_t"], - "bfloat16": ["DT_BF16", "bfloat16_t"]} - -def get_dtype_fmt_options(__inputs__, __outputs__): - options = [] - for x in __inputs__ + __outputs__: - x_n = x.get("param_name").upper() - x_fmt = x.get("format") - x_dtype = x.get("dtype") - options.append("-DDTYPE_{n}={t}".format(n=x_n, t=DTYPE_MAP.get(x_dtype)[1])) - options.append("-DORIG_DTYPE_{n}={ot}".format(n=x_n, ot=DTYPE_MAP.get(x_dtype)[0])) - options.append("-DFORMAT_{n}=FORMAT_{f}".format(n=x_n, f=x_fmt)) - return options - -def load_dso(so_path): - try: - ctypes.CDLL(so_path) - except OSError as error : - print(error) - raise RuntimeError("cannot open %s" %(so_path)) - else: - print("load so succ ", so_path) - -def get_shortsoc_compile_option(compile_option_list: list, shortsoc:str): - compile_options = [] - if shortsoc in compile_option_list: - compile_options = compile_option_list[shortsoc] - elif '__ALLSOC__' in compile_option_list: - compile_options = compile_option_list['__ALLSOC__'] - return compile_options - -''' - -IMPL_API = ''' -@tbe_register.register_operator("{}") -@para_check.check_op_params({}) -def {}({}, kernel_name="{}", impl_mode=""): - if get_current_build_config("enable_op_prebuild"): - return - __inputs__, __outputs__, __attrs__ = _build_args({}) - options = get_dtype_fmt_options(__inputs__, __outputs__) - options += ["-x", "cce"] - ccec = os.environ.get('CCEC_REAL_PATH') - if ccec is None: - ccec = shutil.which("ccec") - if ccec != None: - ccec_path = os.path.dirname(ccec) - tikcpp_path = os.path.realpath(os.path.join(ccec_path, "..", "..", "tikcpp")) - else: - tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") - options.append("-I" + tikcpp_path) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) - options.append("-I" + os.path.join(PYF_PATH, "..", "ascendc", "common")) - if impl_mode == "high_performance": - options.append("-DHIGH_PERFORMANCE=1") - elif impl_mode == "high_precision": - options.append("-DHIGH_PRECISION=1") - if get_default_build_config("enable_deterministic_mode") == 1: - options.append("-DDETEMINISTIC_MODE=1") - - custom_compile_options = {}, - custom_all_compile_options = {}, - soc_version = get_soc_spec("SOC_VERSION") - soc_short = get_soc_spec("SHORT_SOC_VERSION").lower() - custom_compile_options_soc = get_shortsoc_compile_option(custom_compile_options[0], soc_short) - custom_all_compile_options_soc = get_shortsoc_compile_option(custom_all_compile_options[0], soc_short) - options += custom_all_compile_options_soc - options += custom_compile_options_soc - - origin_func_name = "{}" - ascendc_src_dir = "{}" - ascendc_src_file = "{}" - src = os.path.join(PYF_PATH, "..", "ascendc", ascendc_src_dir, ascendc_src_file) - if not os.path.exists(src): - src = os.path.join(PYF_PATH, ascendc_src_file) -''' - -REPLAY_OP_API = ''' - print("start replay Acend C Operator {}, kernel name is {}") - tikreplay_codegen_path = tikcpp_path + "/tikreplaylib/lib" - tikreplay_stub_path = tikcpp_path + "/tikreplaylib/lib/" + soc_version - print("start load libtikreplaylib_codegen.so and libtikreplaylib_stub.so") - codegen_so_path = tikreplay_codegen_path + "/libtikreplaylib_codegen.so" - replaystub_so_path = tikreplay_stub_path + "/libtikreplaylib_stub.so" - if PYF_PATH.endswith("dynamic"): - op_replay_path = os.path.join(PYF_PATH, "..", "..", "op_replay") - else: - op_replay_path = os.path.join(PYF_PATH, "..", "op_replay") - replayapi_so_path = os.path.join(op_replay_path, "libreplay_{}_" + soc_short + ".so") - load_dso(codegen_so_path) - load_dso(replaystub_so_path) - load_dso(replayapi_so_path) - op_type = "{}" - entry_obj = os.path.join(op_replay_path, "{}_entry_" + soc_short + ".o") - code_channel = get_code_channel(src, kernel_name, op_type, options) - op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ - attrs = __attrs__, impl_mode = impl_mode) - res, msg = replay_op(op_info, entry_obj, code_channel, src, options) - if not res: - print("call replay op failed for %s and get into call compile op" %(msg)) - compile_op(src, origin_func_name, op_info, options, code_channel, '{}') -''' - -COMPILE_OP_API = ''' - print("start compile Ascend C operator {}. kernel name is {}") - op_type = "{}" - code_channel = get_code_channel(src, kernel_name, op_type, options) - op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ - attrs = __attrs__, impl_mode = impl_mode, origin_inputs=[{}], origin_outputs = [{}]) - compile_op(src, origin_func_name, op_info, options, code_channel, '{}') -''' - -SUP_API = ''' -def {}({}, impl_mode=""): - __inputs__, __outputs__, __attrs__ = _build_args({}) - ret_str = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) - ret_dict = json.loads(ret_str) - err_code = ret_dict.get("ret_code") - sup = "Unknown" - reason = "Unknown reason" - if err_code is not None: - if err_code is 0: - sup = "True" - reason = "" - elif err_code is 1: - sup = "False" - reason = ret_dict.get("reason") - else: - sup = "Unknown" - reason = ret_dict.get("reason") - return sup, reason -''' -CAP_API = ''' -def {}({}, impl_mode=""): - __inputs__, __outputs__, __attrs__ = _build_args({}) - result = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) - return result.decode("utf-8") -''' -GLZ_API = ''' -@tbe_register.register_param_generalization("{}") -def {}_generalization({}, generalize_config=None): - __inputs__, __outputs__, __attrs__ = _build_args({}) - ret_str = generalize_op_params("{}", __inputs__, __outputs__, __attrs__, generalize_config) - return [json.loads(ret_str)] -''' - -ATTR_DEFAULT = {'bool': 'False', 'int': '0', 'float': '0.0', 'listInt': '[]', - 'listFloat': '[]', 'listBool': '[]', 'listListInt': '[[]]', 'str': ''} - - -def optype_snake(origin_str): - temp_str = origin_str[0].lower() + origin_str[1:] - new_str = re.sub(r'([A-Z])', r'_\1', temp_str).lower() - return new_str - - -class AdpBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - self.argsname = [] - self.argsdefv = [] - self.op_compile_option:str = '{}' - super().__init__(op_type) - - - def write_adapt(self: any, impl_path, path: str, op_compile_option_all:list = None): - self._build_paradefault() - if impl_path != "": - src_file = os.path.join(impl_path, self.op_file + '.cpp') - if not os.path.exists(src_file): - return - out_path = os.path.abspath(path) - if self.dynamic_shape and not out_path.endswith('dynamic'): - out_path = os.path.join(path, 'dynamic') - os.makedirs(out_path, mode=0o700, exist_ok=True) - adpfile = os.path.join(out_path, self.op_file + '.py') - self._gen_op_compile_option(op_compile_option_all) - with os.fdopen(os.open(adpfile, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - self._write_head(fd) - self._write_argparse(fd) - self._write_impl(fd) - if self.op_chk_support: - self._write_cap('check_supported', fd) - self._write_cap('get_op_support_info', fd) - if self.op_fmt_sel: - self._write_cap('op_select_format', fd) - self._write_cap('get_op_specific_info', fd) - if self.op_range_limit == 'limited' or self.op_range_limit == 'dynamic': - self._write_glz(fd) - - - def _gen_op_compile_option(self:any, op_compile_option_all:list =None): - if op_compile_option_all is not None: - if self.op_type in op_compile_option_all: - self.op_compile_option = op_compile_option_all[self.op_type] - elif "__all__" in op_compile_option_all: - self.op_compile_option = op_compile_option_all["__all__"] - - - def _ip_argpack(self: any, default: bool = True) -> list: - args = [] - for i in range(len(self.input_name)): - arg = self.input_name[i] - if default and self.argsdefv[i] is not None: - arg += '=' + self.argsdefv[i] - args.append(arg) - return args - - def _op_argpack(self: any, default: bool = True) -> list: - args = [] - argidx = len(self.input_name) - for i in range(len(self.output_name)): - arg = self.output_name[i] - if default and self.argsdefv[i + argidx] is not None: - arg += '=' + self.argsdefv[i + argidx] - args.append(arg) - return args - - def _attr_argpack(self: any, default: bool = True) -> list: - args = [] - argidx = len(self.input_name) + len(self.output_name) - for i in range(len(self.attr_list)): - att = self.attr_list[i] - arg = att - if default and self.argsdefv[i + argidx] is not None: - if self.attr_val.get(att).get('type') == 'str': - arg += '="' + self.argsdefv[i + argidx] + '"' - elif self.attr_val.get(att).get('type') == 'bool': - arg += '=' + self.argsdefv[i + argidx].capitalize() - else: - arg += '=' + self.argsdefv[i + argidx] - args.append(arg) - return args - - def _build_paralist(self: any, default: bool = True) -> str: - args = [] - args.extend(self._ip_argpack(default)) - args.extend(self._op_argpack(default)) - args.extend(self._attr_argpack(default)) - return ', '.join(args) - - def _io_parachk(self: any, types: list, type_name: str) -> list: - chk = [] - for iot in types: - if iot == 'optional': - ptype = 'OPTION' - else: - ptype = iot.upper() - chk.append('para_check.{}_{}'.format(ptype, type_name)) - return chk - - def _attr_parachk(self: any) -> list: - chk = [] - for att in self.attr_list: - if self.attr_val.get(att).get('paramType') == 'optional': - pt = 'OPTION' - else: - pt = self.attr_val.get(att).get('paramType').upper() - att_type = self.attr_val.get(att).get('type').upper() - att_type = att_type.replace('LIST', 'LIST_') - chk.append('para_check.{}_ATTR_{}'.format(pt, att_type)) - return chk - - def _build_parachk(self: any) -> str: - chk = [] - chk.extend(self._io_parachk(self.input_type, 'INPUT')) - chk.extend(self._io_parachk(self.output_type, 'OUTPUT')) - chk.extend(self._attr_parachk()) - chk.append('para_check.KERNEL_NAME') - return ', '.join(chk) - - def _build_paradefault(self: any): - optional = False - argtypes = [] - argtypes.extend(self.input_type) - argtypes.extend(self.output_type) - for atype in argtypes: - if atype == 'optional': - optional = True - if optional: - self.argsdefv.append('None') - else: - self.argsdefv.append(None) - for attr in self.attr_list: - atype = self.attr_val.get(attr).get('paramType') - if atype == 'optional': - optional = True - attrval = self.attr_val.get(attr).get('defaultValue') - if attrval is not None: - optional = True - if type == "bool": - attrval = attrval.capitalize() - elif type == "str": - attrval = "\"" + attrval + "\"" - self.argsdefv.append(attrval) - continue - if optional: - self.argsdefv.append(ATTR_DEFAULT.get(self.attr_val.get(attr).get('type'))) - else: - self.argsdefv.append(None) - - def _write_head(self: any, fd: object): - fd.write(IMPL_HEAD) - - def _write_argparse(self: any, fd: object): - args = self._build_paralist(False) - fd.write('def _build_args({}):\n'.format(args)) - fd.write(' __inputs__ = []\n') - fd.write(' for arg in [{}]:\n'.format(', '.join(self.input_name))) - fd.write(' if arg != None:\n') - fd.write(' if isinstance(arg, (list, tuple)):\n') - fd.write(' if len(arg) == 0:\n') - fd.write(' continue\n') - fd.write(' __inputs__.append(arg[0])\n') - fd.write(' else:\n') - fd.write(' __inputs__.append(arg)\n') - fd.write(' __outputs__ = []\n') - fd.write(' for arg in [{}]:\n'.format(', '.join(self.output_name))) - fd.write(' if arg != None:\n') - fd.write(' if isinstance(arg, (list, tuple)):\n') - fd.write(' if len(arg) == 0:\n') - fd.write(' continue\n') - fd.write(' __outputs__.append(arg[0])\n') - fd.write(' else:\n') - fd.write(' __outputs__.append(arg)\n') - fd.write(' __attrs__ = []\n') - for attr in self.attr_list: - fd.write(' if {} != None:\n'.format(attr)) - fd.write(' attr = {}\n') - fd.write(' attr["name"] = "{}"\n'.format(attr)) - fd.write(' attr["dtype"] = "{}"\n'.format(self.attr_val.get(attr).get('type'))) - fd.write(' attr["value"] = {}\n'.format(attr)) - fd.write(' __attrs__.append(attr)\n') - fd.write(' return __inputs__, __outputs__, __attrs__\n') - - def _write_impl(self: any, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - pchk = self._build_parachk() - if len(self.kern_name) > 0: - kern_name = self.kern_name - else: - kern_name = self.op_intf - src = self.op_file + '.cpp' - fd.write(IMPL_API.format(self.op_type, pchk, self.op_intf, argsdef, kern_name, argsval,\ - self.custom_compile_options, self.custom_all_compile_options, self.op_intf,\ - optype_snake(self.op_type), src)) - if self.op_replay_flag: - fd.write(REPLAY_OP_API.format(self.op_type, kern_name, self.op_file, self.op_type, self.op_file,\ - self.op_compile_option)) - else: - fd.write(COMPILE_OP_API.format(self.op_type, kern_name, self.op_type, ', '.join(self.input_name),\ - ', '.join(self.output_name), self.op_compile_option)) - - def _write_cap(self: any, cap_name: str, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - if cap_name == 'check_supported': - fd.write(SUP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) - else: - fd.write(CAP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) - - def _write_glz(self: any, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - fd.write(GLZ_API.format(self.op_type, self.op_intf, argsdef, argsval, self.op_type)) - - -def write_scripts(cfgfile: str, cfgs: dict, dirs: dict, ops: list = None, op_compile_option:list = None): - batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') - iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') - file_map = {} - op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, AdpBuilder,\ - ops, dirs.get(const_var.AUTO_GEN_DIR)) - for op_desc in op_descs: - op_desc.write_adapt(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), op_compile_option) - file_map[op_desc.op_type] = op_desc.op_file - return file_map - -if __name__ == '__main__': - if len(sys.argv) <= 6: - raise RuntimeError('arguments must greater equal than 6') - rep_cfg = {} - rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] - rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] - cfg_dir = {} - cfg_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] - cfg_dir[const_var.CFG_OUT_DIR] = sys.argv[5] - cfg_dir[const_var.AUTO_GEN_DIR] = sys.argv[6] - write_scripts(cfgfile=sys.argv[1], cfgs=rep_cfg, dirs=cfg_dir) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py deleted file mode 100755 index 7a97180be..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py +++ /dev/null @@ -1,114 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import glob -import json -import argparse -import const_var - - -def load_json(json_file: str): - with open(json_file, encoding='utf-8') as file: - json_content = json.load(file) - return json_content - - -def get_specified_suffix_file(root_dir, suffix): - specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) - all_suffix_files = glob.glob(specified_suffix, recursive=True) - return all_suffix_files - - -def add_simplified_config(op_type, key, core_type, objfile, config): - simple_cfg = config.get('binary_info_config.json') - op_cfg = simple_cfg.get(op_type) - if not op_cfg: - op_cfg = {} - op_cfg['dynamicRankSupport'] = True - op_cfg['simplifiedKeyMode'] = 0 - op_cfg['binaryList'] = [] - simple_cfg[op_type] = op_cfg - bin_list = op_cfg.get('binaryList') - bin_list.append({'coreType': core_type, 'simplifiedKey': key, 'binPath': objfile}) - - -def add_op_config(op_file, bin_info, config): - op_cfg = config.get(op_file) - if not op_cfg: - op_cfg = {} - op_cfg['binList'] = [] - config[op_file] = op_cfg - op_cfg.get('binList').append(bin_info) - - -def gen_ops_config(json_file, soc, config): - core_type_map = {"MIX": 0, "AiCore": 1, "VectorCore": 2} - contents = load_json(json_file) - if ('binFileName' not in contents) or ('supportInfo' not in contents): - return - json_base_name = os.path.basename(json_file) - op_dir = os.path.basename(os.path.dirname(json_file)) - support_info = contents.get('supportInfo') - bin_name = contents.get('binFileName') - bin_suffix = contents.get('binFileSuffix') - core_type = core_type_map.get(contents.get("coreType")) - bin_file_name = bin_name + bin_suffix - op_type = bin_name.split('_')[0] - op_file = op_dir + '.json' - bin_info = {} - keys = support_info.get('simplifiedKey') - if keys: - bin_info['simplifiedKey'] = keys - for key in keys: - add_simplified_config(op_type, key, core_type, os.path.join(soc, op_dir, bin_file_name), config) - bin_info['staticKey'] = support_info.get('staticKey') - bin_info['int64Mode'] = support_info.get('int64Mode') - bin_info['inputs'] = support_info.get('inputs') - bin_info['outputs'] = support_info.get('outputs') - if support_info.get('attrs'): - bin_info['attrs'] = support_info.get('attrs') - bin_info['binInfo'] = {'jsonFilePath': os.path.join(soc, op_dir, json_base_name)} - add_op_config(op_file, bin_info, config) - - -def gen_all_config(root_dir, soc): - suffix = 'json' - config = {} - config['binary_info_config.json'] = {} - all_json_files = get_specified_suffix_file(root_dir, suffix) - for _json in all_json_files: - gen_ops_config(_json, soc, config) - for cfg_key in config.keys(): - cfg_file = os.path.join(root_dir, cfg_key) - with os.fdopen(os.open(cfg_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - json.dump(config.get(cfg_key), fd, indent=' ') - - -def args_prase(): - parser = argparse.ArgumentParser() - parser.add_argument('-p', - '--path', - nargs='?', - required=True, - help='Parse the path of the json file.') - parser.add_argument('-s', - '--soc', - nargs='?', - required=True, - help='Parse the soc_version of ops.') - return parser.parse_args() - - -def main(): - args = args_prase() - gen_all_config(args.path, args.soc) - - -if __name__ == '__main__': - main() diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py deleted file mode 100755 index 1cac7d911..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py +++ /dev/null @@ -1,65 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import opdesc_parser -import replay_codegen -import const_var -from replay_codegen import ReplayCodeGenParams - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - - -class ReplayBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - super().__init__(op_type) - - def gen_replay_source(self: any, impl_path: str, out_path: str, ops_product: str): - if not self.op_replay_flag: - print('{} replay not enabled'.format(self.op_type)) - return - argn = len(self.input_name) + len(self.output_name) + 1 - if self.op_replay_batch: - print('{} replay in batch mode'.format(self.op_type)) - else: - print('{} replay in normal mode'.format(self.op_type)) - if impl_path.endswith('op_kernel'): - implf = os.path.join(impl_path, self.op_file + '.cpp') - tiling_file = os.path.join(impl_path, "../op_host", self.op_file + '_tiling.h') - else: - if self.dynamic_shape: - dyn_path = 'dynamic' - else: - dyn_path = '' - implf = os.path.join(impl_path, dyn_path, self.op_file + '.cpp') - tiling_file = os.path.join(impl_path, "../../op_tiling", self.op_file + '_tiling.h') - rep_conf = replay_codegen.ReplayCodeGen(ReplayCodeGenParams(self.op_type, implf, tiling_file, self.op_file, \ - self.op_intf, argn, self.op_replay_batch, self.max_block_dim, self.max_shape_size)) - rep_conf.set_batch(self.op_replay_batch) - rep_conf.set_outdir(out_path) - rep_conf.gen_replay(ops_product) - - -def gen_replay(cfgfile: str, cfgs: dict, dirs: dict, ops_product: str, ops: list = None): - batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') - iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') - op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, ReplayBuilder, ops) - for op_desc in op_descs: - op_desc.gen_replay_source(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), ops_product) - - -if __name__ == '__main__': - if len(sys.argv) <= 6: - raise RuntimeError('arguments must greater than 6') - rep_cfg = {} - rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] - rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] - rep_dir = {} - rep_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] - rep_dir[const_var.CFG_OUT_DIR] = sys.argv[5] - gen_replay(sys.argv[1], rep_cfg, rep_dir, sys.argv[6]) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp deleted file mode 100755 index 0e8834664..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp +++ /dev/null @@ -1,117 +0,0 @@ -#include -#include -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#include "register/op_check.h" -#define __ASCENDC_REPLAY_CODE__ -#include - -using namespace std; -using namespace optiling; -using namespace AscendCReplay; - -extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); -extern "C" int elf_batch_append(char *elf, uint32_t elfSize, char *jit, int kernum, char *atext[], int alen[], - int atlen, const char* kernelname[]); - -#define KERNEL_N 1 -#define ARG_N (__ARG_NUM__) -#define MAX_L (1024 * 1024 * 100) -#define MAX_E (1024 * 1024) - -int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) -{ - // gen type 1 : direct call codes 0: load .o file - if (param.gentype < 0 || param.gentype > 1) { - printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); - return 0; - } else if (param.gentype == 1 && param.objptr == nullptr) { - printf("Error: call replay with direct call mode, but code obj addr is null\n"); - return 0; - } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { - printf("Error: call replay with object file mode, but object file path is null\n"); - return 0; - } - // core_type 0:MIX 1:CUBE 2:VEC - if (core_type < 0 || core_type > 2) { - printf("Error: call replay core type is %d !\n", core_type); - return 0; - } - g_coreType = __CORE_TYPE__; - g_taskRation = param.task_ration; - g_tilingKey = param.tiling_key; - - unsigned char *buf, *jit; - char *kernel[KERNEL_N]; - int len[KERNEL_N]; - block_idx = 0; - block_num = param.block_dim; - g_ubBase = block_num; - uint8_t *code = (uint8_t *)malloc(MAX_L); - uint8_t *pos = code; - struct timespec tp1, tp2; - - clock_gettime(CLOCK_MONOTONIC, &tp1); - if (block_num > 32) { - printf("Error: block_num > 32\n"); - return 0; - } - //__OP_FOPEN__ - for (int i = 0; i < KERNEL_N; i++) { - //__OP_SET_KERNEL__ - for (int j = 0; j < ARG_N; j++) - AddArg(j, ARG_STEP * (j + 1)); -#ifdef FP_CEILING - SetCtrlFloatEnable(); -#else - SetCtrlFloatDisable(); -#endif - CodeInit(pos, true); - __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); - CodeEnd(); - kernel[i] = (char *)pos; - len[i] = CodeLen(); - pos += len[i]; - } - //__OP_FCLOSE__ - clock_gettime(CLOCK_MONOTONIC, &tp2); - buf = (unsigned char *)malloc(MAX_E); - int fd = open(param.entry_file, O_RDONLY); - if (fd < 0) { - printf("[error]: cannot find entry.o : %s\n", param.entry_file); - return 0; - } - uint32_t bufSize = read(fd, buf, MAX_E); - if (bufSize <= 0) { - printf("[error]: entry.o : %s is too small ! \n", param.entry_file); - } - close(fd); - jit = (unsigned char *)malloc(MAX_L); - printf("total code generated %ld\n", pos - code); - int sz = elf_batch_append((char *)buf, bufSize, (char *)jit, KERNEL_N, kernel, len, pos - code, ¶m.kernel_name); - if (tp1.tv_sec != tp2.tv_sec) { - printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); - } else { - printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); - } - printf("new elf size %d\n", sz); - if (param.gentype == 0) { - fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); - (void)write(fd, jit, sz); - close(fd); - free(jit); - } else if (param.gentype == 1) { - *param.objptr = (char*)jit; - } - free(buf); - free(code); - return sz; -} - -REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py deleted file mode 100755 index 137c9f39f..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py +++ /dev/null @@ -1,115 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" -import os -import stat -import ctypes -import collections -import shutil -import subprocess -import copy - -"""CODE_* is used to cube/vector api is called in operator code -CODE_MIX means both cube and vector api is called -CODE_CUBE means only cube api is called -CODE_VEC means only vector api is called -""" -CODE_MIX = 0 -CODE_CUBE = 1 -CODE_VEC = 2 - - -def _is_v220(op_product: str): - """return if current soc version is V220 - - Returns: - res: True means V220 - """ - if op_product in ["ascend910b"]: - return True - return False - - -InfoCodeChanelParams = collections.namedtuple('InfoCodeChanelParams',\ -['src_file', 'tiling_header', 'kernel_name', 'outdir', 'op_product', 'compile_options']) - - -def infer_code_channel(params: InfoCodeChanelParams): - """get code channel for v220, return CODE_MIX if soc version is not V220 - - Args: - src_file (str): AscendC operator code file - src_file (str): AscendC operator tiling header file - kernel_name (str): kernel function name - optype (str): operator type - compile_options (list): compile options for ccec cmd - - Raises: - Exception: if not exist L1/L0/UB if code, it's not a aicore code - - Returns: - res (int): CODE_MIX/CODE_CUBE/CODE_VEC - """ - if not _is_v220(params.op_product): - return CODE_MIX - return CODE_VEC - if params.compile_options is None: - compile_options = [] - else: - compile_options = params.compile_options - ccec = shutil.which("ccec") - if ccec is not None: - ccec_path = os.path.dirname(ccec) - tikcpp_path = os.path.realpath(os.path.join(ccec_path, "..", "..", "tikcpp")) - else: - tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") - compile_options.append("-I" + tikcpp_path) - compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) - compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) - compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) - compile_options += ["-include", params.tiling_header] - arch = "dav-c220-cube" - sub_core_type = "AIC" - optional_core = "AiCore" - compile_cmd = [shutil.which("ccec"), '-c', '-O3'] - compile_cmd += compile_options - temp_file_name_tag = "_" + str(os.getpid()) + "_temp.o" - dst_file = os.path.join(kernel_meta_dir, kernel_name + temp_file_name_tag) - compile_cmd += [params.src_file, "--cce-aicore-arch={}".format(arch), - "--cce-aicore-only", "-o", dst_file, - "-mllvm", "-cce-aicore-function-stack-size=16000", - "-mllvm", "-cce-aicore-record-overflow=true", - "-mllvm", "-cce-aicore-addr-transform"] - compile_cmd += ["-std=c++17"] - print('get_code_channel: ', ' '.join(compile_cmd)) - proc = subprocess.Popen( - compile_cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) - (out, _) = proc.communicate() - if proc.returncode != 0: - print('get_code_channel coretype compile error: ', out.decode()) - msg = "compile %s error :%s\n" % (params.src_file, out.decode()) - raise Exception(f"get_code_channel coretype error, msg is{msg}") - objdump_cmd = ['objdump', '-s', '-j', '.text', '{}'.format(dst_file)] - - proc = subprocess.Popen( - objdump_cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) - (out, _) = proc.communicate() - if proc.returncode != 0: - print('get_code_channel objdump error: ', out.decode()) - msg = "get_code_channel objdump %s error :%s\n" % (src_file, out.decode()) - raise Exception(f"get_code_channel objdump error, msg is{msg}") - os.remove(dst_file) - lines = out.decode('utf-8').split('\n') - for line in lines: - insts = line.strip().split() - if len(insts) < 5: - continue - for inst in insts[1:]: - if len(inst) != 8: - continue - if inst[6] == 'f': - return CODE_MIX - return CODE_VEC diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/const_var.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/const_var.py deleted file mode 100755 index 8b32c3b91..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/const_var.py +++ /dev/null @@ -1,33 +0,0 @@ - -#!/usr/bin/env python -# coding=utf-8 -""" -Function: -The replay funtion entry -Copyright Information: -Huawei Technologies Co., Ltd. All Rights Reserved © 2020 -""" - -import os -import stat - - -REPLAY_BATCH = 'batch' -REPLAY_ITERATE = 'iterate' -CFG_IMPL_DIR = 'impl_dir' -CFG_OUT_DIR = 'out_dir' -AUTO_GEN_DIR = 'auto_gen_dir' -WFLAGS = os.O_WRONLY | os.O_CREAT | os.O_TRUNC -WMODES = stat.S_IWUSR | stat.S_IRUSR -SOC_MAP_EXT = {'ascend310p': 'Ascend310P3', 'ascend310b': 'Ascend310B1', - 'ascend910': 'Ascend910A', 'ascend910b': 'Ascend910B1'} -BIN_CMD = 'opc $1 --main_func={fun} --input_param={param} --soc_version={soc} \ ---output=$2 --impl_mode={impl} --simplified_key_mode=0 --op_mode=dynamic\n' -CHK_CMD = ''' -if ! test -f $2/{res_file} ; then - echo "$2/{res_file} not generated!" - exit 1 -fi -''' -ATTR_DEF_VAL = {'str' : '', 'int': 0, 'float': 0.0, 'bool': False, 'list_bool': [], - 'list_int': [], 'list_float': [], 'list_list_int': [[]]} diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh deleted file mode 100755 index 55e12e5ed..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh +++ /dev/null @@ -1,57 +0,0 @@ -#!/usr/bin/bash - -project_path=$1 -build_path=$2 -vendor_name=customize -if [[ ! -d "$project_path" ]]; then - echo "[ERROR] No projcet path is provided" - exit 1 -fi - -if [[ ! -d "$build_path" ]]; then - echo "[ERROR] No build path is provided" - exit 1 -fi - -# copy ai_core operators implements -# tbe_impl_files_num=$(ls $project_path/tbe/impl/* 2> /dev/null | wc -l) -# if [[ "$tbe_impl_files_num" -gt 0 ]];then -# cp -r ${project_path}/tbe/impl/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/ai_core/tbe/customize_impl -# cp -r ${project_path}/tbe/impl/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/vector_core/tbe/customize_impl -# fi - -# copy aicpu kernel so operators -if [[ -d "${project_path}/cpukernel/aicpu_kernel_lib" ]]; then - cp -f ${project_path}/cpukernel/aicpu_kernel_lib/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/cpu/aicpu_kernel/impl - rm -rf ${project_path}/cpukernel/aicpu_kernel_lib -fi - -# merge aicpu.ini and aicore.ini to generate npu_supported_ops.json -# mkdir -p ${build_path}/framework/op_info_cfg -# mkdir -p ${build_path}/framework/op_info_cfg/aicpu_kernel -# mkdir -p ${build_path}/framework/op_info_cfg/ai_core - -# if [[ -d "${project_path}/tbe/op_info_cfg/ai_core" ]]; then -# bash ${project_path}/cmake/util/gen_ops_filter.sh ${project_path}/tbe/op_info_cfg/ai_core ${build_path}/framework/op_info_cfg/ai_core -# fi - -# if [[ -d "${project_path}/cpukernel/op_info_cfg/aicpu_kernel" ]]; then -# bash ${project_path}/cmake/util/gen_ops_filter.sh ${project_path}/cpukernel/op_info_cfg/aicpu_kernel ${build_path}/framework/op_info_cfg/aicpu_kernel -# fi - -# aicpu_filter_file=${build_path}/framework/op_info_cfg/aicpu_kernel/npu_supported_ops.json -# aicore_filter_file=${build_path}/framework/op_info_cfg/ai_core/npu_supported_ops.json -# if [[ -f "${aicpu_filter_file}" ]] && [[ ! -f "${aicore_filter_file}" ]]; then -# cp $aicpu_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow -# fi -# if [[ -f "${aicore_filter_file}" ]] && [[ ! -f "${aicpu_filter_file}" ]]; then -# cp $aicore_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow -# fi - -# if [[ -f "${aicore_filter_file}" ]] && [[ -f "${aicpu_filter_file}" ]]; then -# chmod u+w ${aicpu_filter_file} -# python3 ${project_path}/cmake/util/insert_op_info.py ${aicore_filter_file} ${aicpu_filter_file} -# chmod u-w ${aicpu_filter_file} -# cp $aicpu_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow -# fi - diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh deleted file mode 100755 index d4c27d17f..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh +++ /dev/null @@ -1,62 +0,0 @@ -#!/bin/bash -# Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -# Description: Generate npu_supported_ops.json -# ============================================================================== - -if [[ -z "$1" ]]; then - echo -e "[ERROR] No source dir provided" - exit 1 -fi - -if [[ -z "$2" ]]; then - echo -e "[ERROR] No destination dir provided" - exit 1 -fi - -src=$1 -dest_file=$2/npu_supported_ops.json - -if [ -f "$dest_file" ];then - chmod u+w $dest_file -fi - -echo $* - -add_ops() { - name=$1 - isHeavy=$2 - file=$3 - grep -w "\"$name\"" ${file} >/dev/null - if [ $? == 0 ];then - return - fi - echo " \"${name}\": {" >> ${file} - echo " \"isGray\": false," >> ${file} - echo " \"isHeavy\": ${isHeavy}" >> ${file} - echo " }," >> ${file} -} - -echo "{" > ${dest_file} -ini_files=$(find ${src} -name "*.ini") -for file in ${ini_files} ; do - name=$(grep '^\[' ${file} | sed 's/\[//g' | sed 's/]//g' | sed 's/\r//g') - grep 'heavyOp.flag' ${file} >/dev/null - if [ $? == 0 ];then - isHeavy=$(grep 'heavyOp.flag' ${file} | awk -F= '{print $2}') - else - isHeavy="false" - fi - for op in ${name} ; do - add_ops ${op} "false" ${dest_file} - done -done -echo "}" >> ${dest_file} -file_count=$(cat ${dest_file} | wc -l) -line=$(($file_count-1)) -sed -i "${line}{s/,//g}" ${dest_file} - -chmod 640 "${dest_file}" -echo -e "[INFO] Succed generated ${dest_file}" - -exit 0 - diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh deleted file mode 100755 index a06cfc78d..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh +++ /dev/null @@ -1,6 +0,0 @@ -ascend_install_dir=$1 -gen_file_dir=$2 - -# create version.info -compiler_version=$(grep "Version" -w ${ascend_install_dir}/compiler/version.info | awk -F = '{print $2}') -echo "custom_opp_compiler_version=${compiler_version}" > ${gen_file_dir}/version.info \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py deleted file mode 100755 index 28ba08757..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py +++ /dev/null @@ -1,36 +0,0 @@ -# -*- coding: utf-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" -import json -import os -import sys -import stat -import const_var - - -if __name__ == '__main__': - if len(sys.argv) != 3: - print(sys.argv) - print('argv error, inert_op_info.py your_op_file lib_op_file') - sys.exit(2) - - with open(sys.argv[1], 'r') as load_f: - insert_operator = json.load(load_f) - - all_operators = {} - if os.path.exists(sys.argv[2]): - if os.path.getsize(sys.argv[2]) != 0: - with open(sys.argv[2], 'r') as load_f: - all_operators = json.load(load_f) - - for k in insert_operator.keys(): - if k in all_operators.keys(): - print('replace op:[', k, '] success') - else: - print('insert op:[', k, '] success') - all_operators[k] = insert_operator[k] - - with os.fdopen(os.open(sys.argv[2], const_var.WFLAGS, const_var.WMODES), 'w') as json_file: - json_file.write(json.dumps(all_operators, indent=4)) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py deleted file mode 100755 index ace727b90..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py +++ /dev/null @@ -1,248 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import re -import glob -import json -import argparse -import const_var - - -DATA_TPYE_DICT = { - 'float32': 0, - 'float16': 1, - 'int8': 2, - 'int16': 6, - 'uint16': 7, - 'uint8': 4, - 'int32': 3, - 'int64': 9, - 'uint32': 8, - 'uint64': 10, - 'bool': 12, - 'double': 11, - 'string': 13, - 'dual': 14, - 'dual': 15, - 'complex64': 16, - 'complex128': 17, - 'qint8': 18, - 'qint16': 19, - 'qint32': 20, - 'quint8': 21, - 'quint16': 22, - 'resource': 23, - 'string': 24, - 'dual': 25, - 'variant': 26, - 'bf16': 27, - 'bfloat16': 27, - 'undefined': 28, - 'int4': 29, - 'uint1': 30, - 'int2': 31 -} - -FORMAT_DICT = { - 'NCHW': 0, - 'NHWC': 1, - 'ND': 2, - 'NC1HWC0': 3, - 'FRACTAL_Z': 4, - 'NC1C0HWPAD': 5, - 'NHWC1C0': 6, - 'FSR_NCHW': 7, - 'FRACTAL_DECONV': 8, - 'C1HWNC0': 9, - 'FRACTAL_DECONV_TRANSPOSE': 10, - 'FRACTAL_DECONV_SP_STRIDE_TRANS': 11, - 'NC1HWC0_C04': 12, - 'FRACTAL_Z_C04': 13, - 'CHWN': 14, - 'FRACTAL_DECONV_SP_STRIDE8_TRANS': 15, - 'HWCN': 16, - 'NC1KHKWHWC0': 17, - 'BN_WEIGHT': 18, - 'FILTER_HWCK': 19, - 'HASHTABLE_LOOKUP_LOOKUPS': 20, - 'HASHTABLE_LOOKUP_KEYS': 21, - 'HASHTABLE_LOOKUP_VALUE': 22, - 'HASHTABLE_LOOKUP_OUTPUT': 23, - 'HASHTABLE_LOOKUP_HITS': 24, - 'C1HWNCoC0': 25, - 'MD': 26, - 'NDHWC': 27, - 'FRACTAL_ZZ': 28, - 'FRACTAL_NZ': 29, - 'NCDHW': 30, - 'DHWCN': 31, - 'NDC1HWC0': 32, - 'FRACTAL_Z_3D': 33, - 'CN': 34, - 'NC': 35, - 'DHWNC': 36, - 'FRACTAL_Z_3D_TRANSPOSE': 37, - 'FRACTAL_ZN_LSTM': 38, - 'FRACTAL_Z_G': 39, - 'RESERVED': 40, - 'ALL': 41, - 'NULL': 42, - 'ND_RNN_BIAS': 43, - 'FRACTAL_ZN_RNN': 44, - 'NYUV': 45, - 'NYUV_A': 46 -} - - -def load_json(json_file: str): - with open(json_file, encoding='utf-8') as file: - json_content = json.load(file) - return json_content - - -def get_specified_suffix_file(root_dir, suffix): - specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) - all_suffix_files = glob.glob(specified_suffix, recursive=True) - return all_suffix_files - - -def get_deterministic_value(support_info): - deterministic_key = 'deterministic' - if deterministic_key not in support_info: - return 0 - deterministic_value = support_info.get(deterministic_key) - if deterministic_value == 'true': - return 1 - else: - return 0 - - -def get_precision_value(support_info): - precision_key = 'implMode' - precision_value = support_info.get(precision_key) - if precision_value == 'high_performance': - _value = 1 - elif precision_value == 'high_precision': - _value = 2 - else: - _value = 0 - return _value - - -def get_overflow_value(support_info): - return 0 - - -def get_parameters(info): - if info: - if 'dtype' in info: - data_type = info['dtype'] - data_type_value = DATA_TPYE_DICT.get(data_type) - else: - data_type_value = 0 - if 'format' in info: - _format = info['format'] - _format_value = FORMAT_DICT.get(_format) - else: - _format_value = 0 - else: - data_type_value = 0 - _format_value = 0 - return str(data_type_value), str(_format_value) - - -def get_dynamic_parameters(info): - # 动态输入时只需获取第一个参数 - return get_parameters(info[0]) - - -def get_all_parameters(support_info, _type): - result_list = list() - info_lists = support_info.get(_type) - if info_lists: - for _info in info_lists: - # 输入为列表时是动态输入 - if isinstance(_info, (list, tuple)): - data_type_value, _format_value = get_dynamic_parameters(_info) - else: - data_type_value, _format_value = get_parameters(_info) - result_list.append("{},{}".format(data_type_value, _format_value)) - return result_list - - -def get_all_input_parameters(support_info): - result = get_all_parameters(support_info, 'inputs') - return '/'.join(result) - - -def insert_content_into_file(input_file, content): - with open(input_file, 'r+') as file: - lines = file.readlines() - for index, line in enumerate(lines): - match_result = re.search(r'"staticKey":', line) - if match_result: - count = len(line) - len(line.lstrip()) - new_content = "{}{}".format(' ' * count, content) - # 插入到前一行,防止插入最后时还需要考虑是否添加逗号 - lines.insert(index, new_content) - break - file.seek(0) - file.write(''.join(lines)) - - -def insert_simplified_keys(json_file): - contents = load_json(json_file) - # 不存在'binFileName'或者'supportInfo'字段时,非需要替换的解析json文件 - if ('binFileName' not in contents) or ('supportInfo' not in contents): - return - support_info = contents.get('supportInfo') - bin_file_name = contents.get('binFileName') - bin_suffix = contents.get('binFileSuffix') - # 'simplifiedKey'字段已经存在时,直接返回,不重复生成 - if 'simplifiedKey' in support_info: - return - op_type = bin_file_name.split('_')[0] - deterministic = str(get_deterministic_value(support_info)) - precision = str(get_precision_value(support_info)) - overflow = str(get_overflow_value(support_info)) - input_parameters = get_all_input_parameters(support_info) - key = '{}/d={},p={},o={}/{}/'.format( - op_type, - deterministic, - precision, - overflow, - input_parameters) - result = '"simplifiedKey": "' + key + '",\n' - insert_content_into_file(json_file, result) - - -def insert_all_simplified_keys(root_dir): - suffix = 'json' - all_json_files = get_specified_suffix_file(root_dir, suffix) - for _json in all_json_files: - insert_simplified_keys(_json) - - -def args_prase(): - parser = argparse.ArgumentParser() - parser.add_argument('-p', - '--path', - nargs='?', - required=True, - help='Parse the path of the json file.') - return parser.parse_args() - - -def main(): - args = args_prase() - insert_all_simplified_keys(args.path) - - -if __name__ == '__main__': - main() diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py deleted file mode 100755 index 2b77c970d..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py +++ /dev/null @@ -1,115 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - - -def gen_fun_def(title, kernel, argn, arg_type, arg_name): - entry = [] - entry.append(title) - entry.append(kernel) - entry.append('(') - args = [] - for i in range(0, argn): - args.append(arg_type + ' ' + arg_name + str(i)) - entry.append(', '.join(args)) - entry.append(')') - return ' '.join(entry) - - -def gen_batch_kernel_body(fname, argn, arg_name): - body = [] - body.append('{') - fun = [] - fun.append(fname) - fun.append('(') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - fun.append(', '.join(args)) - fun.append(');') - body.append(' '.join(fun)) - body.append('}') - return '\n'.join(body) - - -def gen_mc_kernel_body(kn, argn, arg_name, blknum): - body = [] - body.append('{') - body.append(' switch(block_idx) {') - for blk in range(0, blknum): - fun = [] - fun.append('{}_blk{:02d}'.format(kn, blk)) - fun.append('(') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - fun.append(', '.join(args)) - fun.append(')') - body.append(' case {}: {}; break;'.format(blk, ' '.join(fun))) - body.append(' default: break;') - body.append(' }') - body.append('}') - return '\n'.join(body) - - -def gen_proc_body(argn, arg_name): - body = [] - body.append('{') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - body.append('uint64_t __x = (uint64_t)' + ' + (uint64_t)'.join(args) + ';') - body.append('__asm__ ("NOP");') - body.append('__asm__ ("NOP");') - body.append('__asm__ ("NOP");') - body.append('}') - return '\n'.join(body) - - -def batch_code_gen(kn, argn, argt): - codes = [] - kernel_name = kn - proc_name = kernel_name + '_percore' - arg_num = int(argn) - data_type = argt - arg_type = '__gm__ ' + data_type + '* __restrict__' - arg_name = 'arg' - kernel_title = 'extern \"C\" __global__ __aicore__ void' - proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' - codes.append('#ifndef __aicore__') - codes.append('#define __aicore__ [aicore]') - codes.append('#endif') - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') - codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) - codes.append(gen_batch_kernel_body(proc_name, arg_num, arg_name)) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) - codes.append(gen_proc_body(arg_num, arg_name)) - return '\n'.join(codes) + '\n' - - -def mc_code_gen(kn, argn, argt, blknum): - codes = [] - kernel_name = kn - core_num = int(blknum) - arg_num = int(argn) - data_type = argt - arg_type = '__gm__ ' + data_type + '* __restrict__' - arg_name = 'arg' - kernel_title = 'extern \"C\" __global__ __aicore__ void' - proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' - codes.append('#ifndef __aicore__') - codes.append('#define __aicore__ [aicore]') - codes.append('#endif') - for i in range(0, core_num): - proc_name = '{}_blk{:02d}'.format(kernel_name, i) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') - codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) - codes.append(gen_mc_kernel_body(kernel_name, arg_num, arg_name, core_num)) - for i in range(0, core_num): - proc_name = '{}_blk{:02d}'.format(kernel_name, i) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) - codes.append(gen_proc_body(arg_num, arg_name)) - return '\n'.join(codes) + '\n' diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp deleted file mode 100755 index 5079a1043..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp +++ /dev/null @@ -1,10 +0,0 @@ -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#define __ASCENDC_REPLAY_CODE__ -#include "__CCE_FILE__" diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING deleted file mode 100644 index d159169d1..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING +++ /dev/null @@ -1,339 +0,0 @@ - GNU GENERAL PUBLIC LICENSE - Version 2, June 1991 - - Copyright (C) 1989, 1991 Free Software Foundation, Inc., - 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA - Everyone is permitted to copy and distribute verbatim copies - of this license document, but changing it is not allowed. - - Preamble - - The licenses for most software are designed to take away your -freedom to share and change it. By contrast, the GNU General Public -License is intended to guarantee your freedom to share and change free -software--to make sure the software is free for all its users. This -General Public License applies to most of the Free Software -Foundation's software and to any other program whose authors commit to -using it. (Some other Free Software Foundation software is covered by -the GNU Lesser General Public License instead.) You can apply it to -your programs, too. - - When we speak of free software, we are referring to freedom, not -price. Our General Public Licenses are designed to make sure that you -have the freedom to distribute copies of free software (and charge for -this service if you wish), that you receive source code or can get it -if you want it, that you can change the software or use pieces of it -in new free programs; and that you know you can do these things. - - To protect your rights, we need to make restrictions that forbid -anyone to deny you these rights or to ask you to surrender the rights. -These restrictions translate to certain responsibilities for you if you -distribute copies of the software, or if you modify it. - - For example, if you distribute copies of such a program, whether -gratis or for a fee, you must give the recipients all the rights that -you have. You must make sure that they, too, receive or can get the -source code. And you must show them these terms so they know their -rights. - - We protect your rights with two steps: (1) copyright the software, and -(2) offer you this license which gives you legal permission to copy, -distribute and/or modify the software. - - Also, for each author's protection and ours, we want to make certain -that everyone understands that there is no warranty for this free -software. If the software is modified by someone else and passed on, we -want its recipients to know that what they have is not the original, so -that any problems introduced by others will not reflect on the original -authors' reputations. - - Finally, any free program is threatened constantly by software -patents. We wish to avoid the danger that redistributors of a free -program will individually obtain patent licenses, in effect making the -program proprietary. To prevent this, we have made it clear that any -patent must be licensed for everyone's free use or not licensed at all. - - The precise terms and conditions for copying, distribution and -modification follow. - - GNU GENERAL PUBLIC LICENSE - TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION - - 0. This License applies to any program or other work which contains -a notice placed by the copyright holder saying it may be distributed -under the terms of this General Public License. The "Program", below, -refers to any such program or work, and a "work based on the Program" -means either the Program or any derivative work under copyright law: -that is to say, a work containing the Program or a portion of it, -either verbatim or with modifications and/or translated into another -language. (Hereinafter, translation is included without limitation in -the term "modification".) Each licensee is addressed as "you". - -Activities other than copying, distribution and modification are not -covered by this License; they are outside its scope. The act of -running the Program is not restricted, and the output from the Program -is covered only if its contents constitute a work based on the -Program (independent of having been made by running the Program). -Whether that is true depends on what the Program does. - - 1. You may copy and distribute verbatim copies of the Program's -source code as you receive it, in any medium, provided that you -conspicuously and appropriately publish on each copy an appropriate -copyright notice and disclaimer of warranty; keep intact all the -notices that refer to this License and to the absence of any warranty; -and give any other recipients of the Program a copy of this License -along with the Program. - -You may charge a fee for the physical act of transferring a copy, and -you may at your option offer warranty protection in exchange for a fee. - - 2. You may modify your copy or copies of the Program or any portion -of it, thus forming a work based on the Program, and copy and -distribute such modifications or work under the terms of Section 1 -above, provided that you also meet all of these conditions: - - a) You must cause the modified files to carry prominent notices - stating that you changed the files and the date of any change. - - b) You must cause any work that you distribute or publish, that in - whole or in part contains or is derived from the Program or any - part thereof, to be licensed as a whole at no charge to all third - parties under the terms of this License. - - c) If the modified program normally reads commands interactively - when run, you must cause it, when started running for such - interactive use in the most ordinary way, to print or display an - announcement including an appropriate copyright notice and a - notice that there is no warranty (or else, saying that you provide - a warranty) and that users may redistribute the program under - these conditions, and telling the user how to view a copy of this - License. (Exception: if the Program itself is interactive but - does not normally print such an announcement, your work based on - the Program is not required to print an announcement.) - -These requirements apply to the modified work as a whole. If -identifiable sections of that work are not derived from the Program, -and can be reasonably considered independent and separate works in -themselves, then this License, and its terms, do not apply to those -sections when you distribute them as separate works. But when you -distribute the same sections as part of a whole which is a work based -on the Program, the distribution of the whole must be on the terms of -this License, whose permissions for other licensees extend to the -entire whole, and thus to each and every part regardless of who wrote it. - -Thus, it is not the intent of this section to claim rights or contest -your rights to work written entirely by you; rather, the intent is to -exercise the right to control the distribution of derivative or -collective works based on the Program. - -In addition, mere aggregation of another work not based on the Program -with the Program (or with a work based on the Program) on a volume of -a storage or distribution medium does not bring the other work under -the scope of this License. - - 3. You may copy and distribute the Program (or a work based on it, -under Section 2) in object code or executable form under the terms of -Sections 1 and 2 above provided that you also do one of the following: - - a) Accompany it with the complete corresponding machine-readable - source code, which must be distributed under the terms of Sections - 1 and 2 above on a medium customarily used for software interchange; or, - - b) Accompany it with a written offer, valid for at least three - years, to give any third party, for a charge no more than your - cost of physically performing source distribution, a complete - machine-readable copy of the corresponding source code, to be - distributed under the terms of Sections 1 and 2 above on a medium - customarily used for software interchange; or, - - c) Accompany it with the information you received as to the offer - to distribute corresponding source code. (This alternative is - allowed only for noncommercial distribution and only if you - received the program in object code or executable form with such - an offer, in accord with Subsection b above.) - -The source code for a work means the preferred form of the work for -making modifications to it. For an executable work, complete source -code means all the source code for all modules it contains, plus any -associated interface definition files, plus the scripts used to -control compilation and installation of the executable. However, as a -special exception, the source code distributed need not include -anything that is normally distributed (in either source or binary -form) with the major components (compiler, kernel, and so on) of the -operating system on which the executable runs, unless that component -itself accompanies the executable. - -If distribution of executable or object code is made by offering -access to copy from a designated place, then offering equivalent -access to copy the source code from the same place counts as -distribution of the source code, even though third parties are not -compelled to copy the source along with the object code. - - 4. You may not copy, modify, sublicense, or distribute the Program -except as expressly provided under this License. Any attempt -otherwise to copy, modify, sublicense or distribute the Program is -void, and will automatically terminate your rights under this License. -However, parties who have received copies, or rights, from you under -this License will not have their licenses terminated so long as such -parties remain in full compliance. - - 5. You are not required to accept this License, since you have not -signed it. However, nothing else grants you permission to modify or -distribute the Program or its derivative works. These actions are -prohibited by law if you do not accept this License. Therefore, by -modifying or distributing the Program (or any work based on the -Program), you indicate your acceptance of this License to do so, and -all its terms and conditions for copying, distributing or modifying -the Program or works based on it. - - 6. Each time you redistribute the Program (or any work based on the -Program), the recipient automatically receives a license from the -original licensor to copy, distribute or modify the Program subject to -these terms and conditions. You may not impose any further -restrictions on the recipients' exercise of the rights granted herein. -You are not responsible for enforcing compliance by third parties to -this License. - - 7. If, as a consequence of a court judgment or allegation of patent -infringement or for any other reason (not limited to patent issues), -conditions are imposed on you (whether by court order, agreement or -otherwise) that contradict the conditions of this License, they do not -excuse you from the conditions of this License. If you cannot -distribute so as to satisfy simultaneously your obligations under this -License and any other pertinent obligations, then as a consequence you -may not distribute the Program at all. For example, if a patent -license would not permit royalty-free redistribution of the Program by -all those who receive copies directly or indirectly through you, then -the only way you could satisfy both it and this License would be to -refrain entirely from distribution of the Program. - -If any portion of this section is held invalid or unenforceable under -any particular circumstance, the balance of the section is intended to -apply and the section as a whole is intended to apply in other -circumstances. - -It is not the purpose of this section to induce you to infringe any -patents or other property right claims or to contest validity of any -such claims; this section has the sole purpose of protecting the -integrity of the free software distribution system, which is -implemented by public license practices. Many people have made -generous contributions to the wide range of software distributed -through that system in reliance on consistent application of that -system; it is up to the author/donor to decide if he or she is willing -to distribute software through any other system and a licensee cannot -impose that choice. - -This section is intended to make thoroughly clear what is believed to -be a consequence of the rest of this License. - - 8. If the distribution and/or use of the Program is restricted in -certain countries either by patents or by copyrighted interfaces, the -original copyright holder who places the Program under this License -may add an explicit geographical distribution limitation excluding -those countries, so that distribution is permitted only in or among -countries not thus excluded. In such case, this License incorporates -the limitation as if written in the body of this License. - - 9. The Free Software Foundation may publish revised and/or new versions -of the General Public License from time to time. Such new versions will -be similar in spirit to the present version, but may differ in detail to -address new problems or concerns. - -Each version is given a distinguishing version number. If the Program -specifies a version number of this License which applies to it and "any -later version", you have the option of following the terms and conditions -either of that version or of any later version published by the Free -Software Foundation. If the Program does not specify a version number of -this License, you may choose any version ever published by the Free Software -Foundation. - - 10. If you wish to incorporate parts of the Program into other free -programs whose distribution conditions are different, write to the author -to ask for permission. For software which is copyrighted by the Free -Software Foundation, write to the Free Software Foundation; we sometimes -make exceptions for this. Our decision will be guided by the two goals -of preserving the free status of all derivatives of our free software and -of promoting the sharing and reuse of software generally. - - NO WARRANTY - - 11. BECAUSE THE PROGRAM IS LICENSED FREE OF CHARGE, THERE IS NO WARRANTY -FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN -OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES -PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED -OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF -MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS -TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE -PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, -REPAIR OR CORRECTION. - - 12. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING -WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MAY MODIFY AND/OR -REDISTRIBUTE THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, -INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING -OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED -TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY -YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER -PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE -POSSIBILITY OF SUCH DAMAGES. - - END OF TERMS AND CONDITIONS - - How to Apply These Terms to Your New Programs - - If you develop a new program, and you want it to be of the greatest -possible use to the public, the best way to achieve this is to make it -free software which everyone can redistribute and change under these terms. - - To do so, attach the following notices to the program. It is safest -to attach them to the start of each source file to most effectively -convey the exclusion of warranty; and each file should have at least -the "copyright" line and a pointer to where the full notice is found. - - - Copyright (C) - - This program is free software; you can redistribute it and/or modify - it under the terms of the GNU General Public License as published by - the Free Software Foundation; either version 2 of the License, or - (at your option) any later version. - - This program is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the - GNU General Public License for more details. - - You should have received a copy of the GNU General Public License along - with this program; if not, write to the Free Software Foundation, Inc., - 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. - -Also add information on how to contact you by electronic and paper mail. - -If the program is interactive, make it output a short notice like this -when it starts in an interactive mode: - - Gnomovision version 69, Copyright (C) year name of author - Gnomovision comes with ABSOLUTELY NO WARRANTY; for details type `show w'. - This is free software, and you are welcome to redistribute it - under certain conditions; type `show c' for details. - -The hypothetical commands `show w' and `show c' should show the appropriate -parts of the General Public License. Of course, the commands you use may -be called something other than `show w' and `show c'; they could even be -mouse-clicks or menu items--whatever suits your program. - -You should also get your employer (if you work as a programmer) or your -school, if any, to sign a "copyright disclaimer" for the program, if -necessary. Here is a sample; alter the names: - - Yoyodyne, Inc., hereby disclaims all copyright interest in the program - `Gnomovision' (which makes passes at compilers) written by James Hacker. - - , 1 April 1989 - Ty Coon, President of Vice - -This General Public License does not permit incorporating your program into -proprietary programs. If your program is a subroutine library, you may -consider it more useful to permit linking proprietary applications with the -library. If this is what you want to do, use the GNU Lesser General -Public License instead of this License. diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/README.md b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/README.md deleted file mode 100644 index b41f01682..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/README.md +++ /dev/null @@ -1,246 +0,0 @@ -[![License: GPL v2](https://img.shields.io/badge/License-GPL%20v2-blue.svg)](https://www.gnu.org/licenses/old-licenses/gpl-2.0.en.html) -![Build Status](https://github.com/megastep/makeself/workflows/CI/badge.svg) - -# makeself - Make self-extractable archives on Unix - -[makeself.sh][1] is a small shell script that generates a self-extractable -compressed tar archive from a directory. The resulting file appears as a shell script -(many of those have a **.run** suffix), and can be launched as is. The archive -will then uncompress itself to a temporary directory and an optional arbitrary -command will be executed (for example an installation script). This is pretty -similar to archives generated with WinZip Self-Extractor in the Windows world. -Makeself archives also include checksums for integrity self-validation (CRC -and/or MD5/SHA256 checksums). - -The makeself.sh script itself is used only to create the archives from a -directory of files. The resultant archive is actually a compressed (using -gzip, bzip2, or compress) TAR archive, with a small shell script stub at the -beginning. This small stub performs all the steps of extracting the files, -running the embedded command, and removing the temporary files when done. -All the user has to do to install the software contained in such an -archive is to "run" the archive, i.e **sh nice-software.run**. I recommend -using the ".run" (which was introduced by some Makeself archives released by -Loki Software) or ".sh" suffix for such archives not to confuse the users, -so that they will know they are actually shell scripts (with quite a lot of binary data -attached to them though!). - -I am trying to keep the code of this script as portable as possible, i.e it is -not relying on any bash-specific features and only calls commands that are -installed on any functioning UNIX-compatible system. This script as well as -the archives it generates should run on any Unix flavor, with any compatible -Bourne shell, provided of course that the compression programs are available. - -As of version 2.1, Makeself has been rewritten and tested on the following -platforms : - - * Linux (all distributions) - * Sun Solaris (8 and above) - * HP-UX (tested on 11.0 and 11i on HPPA RISC) - * SCO OpenUnix and OpenServer - * IBM AIX 5.1L - * macOS (Darwin) - * SGI IRIX 6.5 - * FreeBSD - * UnicOS / Cray - * Cygwin (Windows) - -If you successfully run Makeself and/or archives created with it on another -system, then please [let me know][2]! - -Examples of publicly available archives made using makeself are : - - * Game patches and installers for [Id Software][3] games like Quake 3 for Linux or Return To Castle Wolfenstein ; - * All game patches released by [Loki Software][4] for the Linux version of popular games ; - * The [nVidia drivers][5] for Linux - * The installer for the Linux version of [Google Earth][6] - * The [VirtualBox][7] installers for Linux - * The [Makeself][1] distribution itself ;-) - * and countless others... - -**Important note for Apache users:** By default, most Web servers will think that Makeself archives are regular text files and thus they may show up as text in a Web browser. The correct way to prevent this is to add a MIME type for this file format, like so (in httpd.conf) : - -`AddType application/x-makeself .run` - -**Important note for certain GNU/Linux distributions:** Archives created with Makeself prior to v2.1.2 were using an old syntax for the _head_ and _tail_ Unix commands that is being progressively obsoleted in their GNU forms. Therefore you may have problems uncompressing some of these archives. A workaround for this is to set the environment variable $_POSIX2_VERSION to enable the old syntax, i.e. : - -`export _POSIX2_VERSION=199209` - -## Usage - -The syntax of makeself is the following: - -``` -makeself.sh [args] archive_dir file_name label startup_script [script_args] -``` - - * _args_ are optional options for Makeself. The available ones are : - - * **`--version`** : Prints the version number on stdout, then exits immediately - * **`--gzip`** : Use gzip for compression (the default on platforms on which gzip is commonly available, like Linux) - * **`--bzip2`** : Use bzip2 instead of gzip for better compression. The bzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. - * **`--pbzip2`** : Use pbzip2 instead of gzip for better and faster compression on machines having multiple CPUs. The pbzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. - * **`--xz`** : Use xz instead of gzip for better compression. The xz command must be available in the command path. It is recommended that the archive prefix be set to something like '.xz.run' for the archive, so that potential users know that they'll need xz to extract it. - * **`--lzo`** : Use lzop instead of gzip for better compression. The lzop command must be available in the command path. It is recommended that the archive prefix be set to something like `.lzo.run` for the archive, so that potential users know that they'll need lzop to extract it. - * **`--lz4`** : Use lz4 instead of gzip for better compression. The lz4 command must be available in the command path. It is recommended that the archive prefix be set to something like '.lz4.run' for the archive, so that potential users know that they'll need lz4 to extract it. - * **`--zstd`** : Use zstd instead of gzip for better compression. The zstd command must be available in the command path. It is recommended that the archive prefix be set to something like '.zstd.run' for the archive, so that potential users know that they'll need zstd to extract it. - * **`--pigz`** : Use pigz for compression. - * **`--base64`** : Encode the archive to ASCII in Base64 format instead of compressing (base64 command required). - * **`--gpg-encrypt`** : Encrypt the archive using `gpg -ac -z $COMPRESS_LEVEL`. This will prompt for a password to encrypt with. Assumes that potential users have `gpg` installed. - * **`--ssl-encrypt`** : Encrypt the archive using `openssl aes-256-cbc -a -salt`. This will prompt for a password to encrypt with. Assumes that the potential users have the OpenSSL tools installed. - * **`--compress`** : Use the UNIX `compress` command to compress the data. This should be the default on all platforms that don't have gzip available. - * **`--nocomp`** : Do not use any compression for the archive, which will then be an uncompressed TAR. - * **`--complevel`** : Specify the compression level for gzip, bzip2, pbzip2, zstd, xz, lzo or lz4. (defaults to 9) - * **`--threads`** : Specify the number of threads to be used by compressors that support parallelization. Omit to use compressor's default. Most useful (and required) for opting into xz's threading, usually with `--threads=0` for all available cores. pbzip2 and pigz are parallel by default, and setting this value allows limiting the number of threads they use. - * **`--notemp`** : The generated archive will not extract the files to a temporary directory, but in a new directory created in the current directory. This is better to distribute software packages that may extract and compile by themselves (i.e. launch the compilation through the embedded script). - * **`--current`** : Files will be extracted to the current directory, instead of in a subdirectory. This option implies `--notemp` above. - * **`--follow`** : Follow the symbolic links inside of the archive directory, i.e. store the files that are being pointed to instead of the links themselves. - * **`--append`** _(new in 2.1.x)_: Append data to an existing archive, instead of creating a new one. In this mode, the settings from the original archive are reused (compression type, label, embedded script), and thus don't need to be specified again on the command line. - * **`--header`** : Makeself uses a separate file to store the header stub, called `makeself-header.sh`. By default, it is assumed that it is stored in the same location as makeself.sh. This option can be used to specify its actual location if it is stored someplace else. - * **`--cleanup`** : Specify a script that is run when execution is interrupted or finishes successfully. The script is executed with the same environment and initial `script_args` as `startup_script`. - * **`--copy`** : Upon extraction, the archive will first extract itself to a temporary directory. The main application of this is to allow self-contained installers stored in a Makeself archive on a CD, when the installer program will later need to unmount the CD and allow a new one to be inserted. This prevents "Filesystem busy" errors for installers that span multiple CDs. - * **`--nox11`** : Disable the automatic spawning of a new terminal in X11. - * **`--nowait`** : When executed from a new X11 terminal, disable the user prompt at the end of the script execution. - * **`--nomd5`** and **`--nocrc`** : Disable the creation of a MD5 / CRC checksum for the archive. This speeds up the extraction process if integrity checking is not necessary. - * **`--sha256`** : Adds a SHA256 checksum for the archive. This is in addition to the MD5 / CRC checksums unless `--nomd5` is also used. - * **`--lsm` _file_** : Provide and LSM file to makeself, that will be embedded in the generated archive. LSM files are describing a software package in a way that is easily parseable. The LSM entry can then be later retrieved using the `--lsm` argument to the archive. An example of a LSM file is provided with Makeself. - * **`--tar-format opt`** : Specify the tar archive format (default is ustar); you may use any value accepted by your tar command (such as posix, v7, etc). - * **`--tar-extra opt`** : Append more options to the tar command line. - - For instance, in order to exclude the `.git` directory from the packaged archive directory using the GNU `tar`, one can use `makeself.sh --tar-extra "--exclude=.git" ...` - - * **`--keep-umask`** : Keep the umask set to shell default, rather than overriding when executing self-extracting archive. - * **`--packaging-date date`** : Use provided string as the packaging date instead of the current date. - * **`--license`** : Append a license file. - * **`--nooverwrite`** : Do not extract the archive if the specified target directory already exists. - * **`--help-header file`** : Add a header to the archive's `--help` output. - * `archive_dir` is the name of the directory that contains the files to be archived - * `file_name` is the name of the archive to be created - * `label` is an arbitrary text string describing the package. It will be displayed while extracting the files. - * `startup_script` is the command to be executed _from within_ the directory of extracted files. Thus, if you wish to execute a program contained in this directory, you must prefix your command with `./`. For example, `./program` will be fine. The `script_args` are additional arguments for this command. - -Here is an example, assuming the user has a package image stored in a **/home/joe/mysoft**, and he wants to generate a self-extracting package named -**mysoft.sh**, which will launch the "setup" script initially stored in /home/joe/mysoft : - -`makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup -` - -Here is also how I created the [makeself.run][9] archive which contains the Makeself distribution : - -`makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" ` - -Archives generated with Makeself can be passed the following arguments: - - * **`--keep`** : Prevent the files to be extracted in a temporary directory that will be removed after the embedded script's execution. The files will then be extracted in the current working directory and will stay here until you remove them. - * **`--verbose`** : Will prompt the user before executing the embedded command - * **`--target dir`** : Allows to extract the archive in an arbitrary place. - * **`--nox11`** : Do not spawn a X11 terminal. - * **`--confirm`** : Prompt the user for confirmation before running the embedded command. - * **`--info`** : Print out general information about the archive (does not extract). - * **`--lsm`** : Print out the LSM entry, if it is present. - * **`--list`** : List the files in the archive. - * **`--check`** : Check the archive for integrity using the embedded checksums. Does not extract the archive. - * **`--nochown`** : By default, a `chown -R` command is run on the target directory after extraction, so that all files belong to the current user. This is mostly needed if you are running as root, as tar will then try to recreate the initial user ownerships. You may disable this behavior with this flag. - * **`--tar`** : Run the tar command on the contents of the archive, using the following arguments as parameter for the command. - * **`--noexec`** : Do not run the embedded script after extraction. - * **`--noexec-cleanup`** : Do not run the embedded cleanup script. - * **`--nodiskspace`** : Do not check for available disk space before attempting to extract. - * **`--cleanup-args`** : Specify arguments to be passed to the cleanup script. Wrap value in quotes to specify multiple arguments. - -Any subsequent arguments to the archive will be passed as additional arguments to the embedded command. You must explicitly use the `--` special command-line construct before any such options to make sure that Makeself will not try to interpret them. - -## Startup Script - -The startup script must be a regular Shell script. - -Within the startup script, you can use the `$USER_PWD` variable to get the path of the folder from which the self-extracting script is executed. This is especially useful to access files that are located in the same folder as the script, as shown in the example below. - -`my-self-extracting-script.sh --fooBarFileParameter foo.bar` - -## Building and Testing - -Clone the git repo and execute `git submodule update --init --recursive` to obtain all submodules. - -* To make a release: `make` -* To run all tests: `make test` - -## Maven Usage - -Makeself is now supported by the following maven plugin [makeself-maven-plugin](https://github.com/hazendaz/makeself-maven-plugin). Please refer to project for usage and report any bugs in regards to maven plugin on that project. - -## License - -Makeself itself is covered by the [GNU General Public License][8] (GPL) version 2 and above. Archives generated by Makeself don't have to be placed under this license (although I encourage it ;-)), since the archive itself is merely data for Makeself. - -## Contributing - -I will gladly consider merging your pull requests on the [GitHub][10] repository. However, please keep the following in mind: - - * One of the main purposes of Makeself is portability. Do not submit patches that will break supported platforms. The more platform-agnostic, the better. - * Please explain clearly what the purpose of the patch is, and how you achieved it. - -## Download - -Get the latest official distribution [here][9] (version 2.4.2). - -The latest development version can be grabbed from [GitHub][10]. Feel free to submit any patches there through the fork and pull request process. - -## Version history - - * **v1.0:** Initial public release - * **v1.1:** The archive can be passed parameters that will be passed on to the embedded script, thanks to John C. Quillan - * **v1.2:** Cosmetic updates, support for bzip2 compression and non-temporary archives. Many ideas thanks to Francois Petitjean. - * **v1.3:** More patches from Bjarni R. Einarsson and Francois Petitjean: Support for no compression (`--nocomp`), script is no longer mandatory, automatic launch in an xterm, optional verbose output, and -target archive option to indicate where to extract the files. - * **v1.4:** Many patches from Francois Petitjean: improved UNIX compatibility, automatic integrity checking, support of LSM files to get info on the package at run time.. - * **v1.5.x:** A lot of bugfixes, and many other patches, including automatic verification through the usage of checksums. Version 1.5.5 was the stable release for a long time, even though the Web page didn't get updated ;-). Makeself was also officially made a part of the [Loki Setup installer][11], and its source is being maintained as part of this package. - * **v2.0:** Complete internal rewrite of Makeself. The command-line parsing was vastly improved, the overall maintenance of the package was greatly improved by separating the stub from makeself.sh. Also Makeself was ported and tested to a variety of Unix platforms. - * **v2.0.1:** First public release of the new 2.0 branch. Prior versions are officially obsoleted. This release introduced the `--copy` argument that was introduced in response to a need for the [UT2K3][12] Linux installer. - * **v2.1.0:** Big change : Makeself can now support multiple embedded tarballs, each stored separately with their own checksums. An existing archive can be updated with the `--append` flag. Checksums are also better managed, and the `--nochown` option for archives appeared. - * **v2.1.1:** Fixes related to the Unix compression (compress command). Some Linux distributions made the insane choice to make it unavailable, even though gzip is capable of uncompressing these files, plus some more bugfixes in the extraction and checksum code. - * **v2.1.2:** Some bug fixes. Use head -n to avoid problems with POSIX conformance. - * **v2.1.3:** Bug fixes with the command line when spawning terminals. Added `--tar`, `--noexec` for archives. Added `--nomd5` and `--nocrc` to avoid creating checksums in archives. The embedded script is now run through "eval". The `--info` output now includes the command used to create the archive. A man page was contributed by Bartosz Fenski. - * **v2.1.4:** Fixed `--info` output. Generate random directory name when extracting files to . to avoid problems. Better handling of errors with wrong permissions for the directory containing the files. Avoid some race conditions, Unset the $CDPATH variable to avoid problems if it is set. Better handling of dot files in the archive directory. - * **v2.1.5:** Made the md5sum detection consistent with the header code. Check for the presence of the archive directory. Added `--encrypt` for symmetric encryption through gpg (Eric Windisch). Added support for the digest command on Solaris 10 for MD5 checksums. Check for available disk space before extracting to the target directory (Andreas Schweitzer). Allow extraction to run asynchronously (patch by Peter Hatch). Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo). - * **v2.1.6:** Replaced one dot per file progress with a realtime progress percentage and a spinning cursor. Added `--noprogress` to prevent showing the progress during the decompression. Added `--target` dir to allow extracting directly to a target directory. (Guy Baconniere) - * **v2.2.0:** First major new release in years! Includes many bugfixes and user contributions. Please look at the [project page on Github][10] for all the details. - * **v2.3.0:** Support for archive encryption via GPG or OpenSSL. Added LZO and LZ4 compression support. Options to set the packaging date and stop the umask from being overriden. Optionally ignore check for available disk space when extracting. New option to check for root permissions before extracting. - * **v2.3.1:** Various compatibility updates. Added unit tests for Travis CI in the GitHub repo. New `--tar-extra`, `--untar-extra`, `--gpg-extra`, `--gpg-asymmetric-encrypt-sign` options. - * **v2.4.0:** Added optional support for SHA256 archive integrity checksums. - * **v2.4.2:** New --cleanup and --cleanup-args arguments for cleanup scripts. Added threading support for supported compressors. Now supports zstd compression. - * **v2.4.3:** Make explicit POSIX tar archives for increased compatibility. - * **v2.4.4:** Fixed various compatibility issues (no longer use POSIX tar archives), Github Actions to check on Solaris and FreeBSD. - * **v2.4.5:** Added `--tar-format` option to set the tar archive format (default is ustar) - -## Links - - * Check out the ["Loki Setup"][11] installer, used to install many Linux games and other applications, and of which I am the co-author. Since the demise of Loki, I am now the official maintainer of the project, and it is now being hosted here on GitHub. - * Bjarni R. Einarsson also wrote the **setup.sh** installer script, inspired by Makeself. [Check it out !][14] - -## Contact - -This script was written by [Stéphane Peter][15] (megastep at megastep.org). Any enhancements and suggestions are welcome. - -Contributions were included from John C. Quillan, Bjarni R. Einarsson, -Francois Petitjean, Ryan C. Gordon, and many contributors on GitHub. If you think I forgot -your name, don't hesitate to contact me. - -This project is now hosted on GitHub. Feel free to submit patches and bug reports on the [project page][10]. - -* * * - -[Stephane Peter][2] - - [1]: http://makeself.io/ - [2]: mailto:megastep@megastep.org - [3]: http://www.idsoftware.com/ - [4]: http://www.lokigames.com/products/myth2/updates.php3 - [5]: http://www.nvidia.com/ - [6]: http://earth.google.com/ - [7]: http://www.virtualbox.org/ - [8]: http://www.gnu.org/copyleft/gpl.html - [9]: https://github.com/megastep/makeself/releases/download/release-2.4.5/makeself-2.4.5.run - [10]: https://github.com/megastep/makeself - [11]: https://github.com/megastep/loki_setup/ - [12]: http://www.unrealtournament2003.com/ - [13]: http://www.icculus.org/ - [14]: http://bre.klaki.net/programs/setup.sh/ - [15]: https://stephanepeter.com/ diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION deleted file mode 100644 index 59aa62c1f..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION +++ /dev/null @@ -1 +0,0 @@ -2.4.5 diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh deleted file mode 100644 index b5692d490..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh +++ /dev/null @@ -1,9 +0,0 @@ -#!/bin/sh -# -# Create a distributable archive of the current version of Makeself - -VER=`cat VERSION` -mkdir -p /tmp/makeself-$VER release -cp -pPR makeself* test README.md COPYING VERSION .gitmodules /tmp/makeself-$VER/ -./makeself.sh --notemp /tmp/makeself-$VER release/makeself-$VER.run "Makeself v$VER" echo "Makeself has extracted itself" - diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh deleted file mode 100644 index 940903148..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh +++ /dev/null @@ -1,660 +0,0 @@ -cat << EOF > "$archname" -#!/bin/bash -# This script was generated using Makeself $MS_VERSION -# The license covering this archive and its contents, if any, is wholly independent of the Makeself license (GPL) -# 2022.3.19-Modified the MS_Help function and some options -# Huawei Technologies Co., Ltd. - -ORIG_UMASK=\`umask\` - -CRCsum="$CRCsum" -MD5="$MD5sum" -SHA="$SHAsum" -SIGNATURE="$Signature" -TMPROOT=\${TMPDIR:="\$HOME"} -if ! test -d "\$TMPROOT"; then - TMPROOT="\$PWD" -fi -export TMPDIR="\$TMPROOT" -USER_PWD="\$PWD" -if ! test -d "\$USER_PWD"; then - exit 1 -fi -export USER_PWD -ARCHIVE_DIR=\`dirname "\$0"\` -export ARCHIVE_DIR - -name_of_file="\$0 " -pwd_of_file="\$PWD" -label="$LABEL" -script="$SCRIPT" -scriptargs="$SCRIPTARGS" -cleanup_script="${CLEANUP_SCRIPT}" -licensetxt="$LICENSE" -helpheader='$HELPHEADER' -targetdir="$archdirname" -filesizes="$filesizes" -totalsize="$totalsize" -keep="$KEEP" -nooverwrite="$NOOVERWRITE" -quiet="n" -accept="n" -nodiskspace="n" -export_conf="$EXPORT_CONF" -decrypt_cmd="$DECRYPT_CMD" -skip="$SKIP" - -print_cmd_arg="" -if type printf > /dev/null; then - print_cmd="printf" -elif test -x /usr/ucb/echo; then - print_cmd="/usr/ucb/echo" -else - print_cmd="echo" -fi - -if test -d /usr/xpg4/bin; then - PATH=/usr/xpg4/bin:\$PATH - export PATH -fi - -if test -d /usr/sfw/bin; then - PATH=\$PATH:/usr/sfw/bin - export PATH -fi - -unset CDPATH - -MS_Printf() -{ - \$print_cmd \$print_cmd_arg "\$1" -} - -MS_PrintLicense() -{ - PAGER=\${PAGER:=more} - if test x"\$licensetxt" != x; then - PAGER_PATH=\`exec <&- 2>&-; which \$PAGER || command -v \$PAGER || type \$PAGER\` - if test -x "\$PAGER_PATH"; then - echo "\$licensetxt" | \$PAGER - else - echo "\$licensetxt" - fi - if test x"\$accept" != xy; then - while true - do - MS_Printf "Please type y to accept, n otherwise: " - read yn - if test x"\$yn" = xn; then - keep=n - eval \$finish; exit 1 - break; - elif test x"\$yn" = xy; then - break; - fi - done - fi - fi -} - -MS_diskspace() -{ - ( - df -kP "\$1" | tail -1 | awk '{ if (\$4 ~ /%/) {print \$3} else {print \$4} }' - ) -} - -MS_dd() -{ - blocks=\`expr \$3 / 1024\` - bytes=\`expr \$3 % 1024\` - # Test for ibs, obs and conv feature - if dd if=/dev/zero of=/dev/null count=1 ibs=512 obs=512 conv=sync 2> /dev/null; then - dd if="\$1" ibs=\$2 skip=1 obs=1024 conv=sync 2> /dev/null | \\ - { test \$blocks -gt 0 && dd ibs=1024 obs=1024 count=\$blocks ; \\ - test \$bytes -gt 0 && dd ibs=1 obs=1024 count=\$bytes ; } 2> /dev/null - else - dd if="\$1" bs=\$2 skip=1 2> /dev/null - fi -} - -MS_dd_Progress() -{ - if test x"\$noprogress" = xy; then - MS_dd "\$@" - return \$? - fi - file="\$1" - offset=\$2 - length=\$3 - pos=0 - bsize=4194304 - while test \$bsize -gt \$length; do - bsize=\`expr \$bsize / 4\` - done - blocks=\`expr \$length / \$bsize\` - bytes=\`expr \$length % \$bsize\` - ( - dd ibs=\$offset skip=1 2>/dev/null - pos=\`expr \$pos \+ \$bsize\` - MS_Printf " 0%% " 1>&2 - if test \$blocks -gt 0; then - while test \$pos -le \$length; do - dd bs=\$bsize count=1 2>/dev/null - pcent=\`expr \$length / 100\` - pcent=\`expr \$pos / \$pcent\` - if test \$pcent -lt 100; then - MS_Printf "\b\b\b\b\b\b\b" 1>&2 - if test \$pcent -lt 10; then - MS_Printf " \$pcent%% " 1>&2 - else - MS_Printf " \$pcent%% " 1>&2 - fi - fi - pos=\`expr \$pos \+ \$bsize\` - done - fi - if test \$bytes -gt 0; then - dd bs=\$bytes count=1 2>/dev/null - fi - MS_Printf "\b\b\b\b\b\b\b" 1>&2 - MS_Printf " 100%% " 1>&2 - ) < "\$file" -} - -MS_Help() -{ - cat << EOH >&2 -Usage: \$0 [options] -Options: - --help | -h Print this message - --info Print embedded info : title, default target directory, embedded script ... - --list Print the list of files in the archive - --check Checks integrity and version dependency of the archive - --quiet Quiet install mode, skip human-computer interactions - --nox11 Do not spawn an xterm - --noexec Do not run embedded script - --extract= Extract directly to a target directory (absolute or relative) - Usually used with --noexec to just extract files without running - --tar arg1 [arg2 ...] Access the contents of the archive through the tar command -\${helpheader} -EOH -} - -MS_Verify_Sig() -{ - GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` - MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` - test -x "\$GPG_PATH" || GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` - test -x "\$MKTEMP_PATH" || MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` - offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` - temp_sig=\`mktemp -t XXXXX\` - echo \$SIGNATURE | base64 --decode > "\$temp_sig" - gpg_output=\`MS_dd "\$1" \$offset \$totalsize | LC_ALL=C "\$GPG_PATH" --verify "\$temp_sig" - 2>&1\` - gpg_res=\$? - rm -f "\$temp_sig" - if test \$gpg_res -eq 0 && test \`echo \$gpg_output | grep -c Good\` -eq 1; then - if test \`echo \$gpg_output | grep -c \$sig_key\` -eq 1; then - test x"\$quiet" = xn && echo "GPG signature is good" >&2 - else - echo "GPG Signature key does not match" >&2 - exit 2 - fi - else - test x"\$quiet" = xn && echo "GPG signature failed to verify" >&2 - exit 2 - fi -} - -MS_Check() -{ - OLD_PATH="\$PATH" - PATH=\${GUESS_MD5_PATH:-"\$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} - MD5_ARG="" - MD5_PATH=\`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum\` - test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which md5 || command -v md5 || type md5\` - test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which digest || command -v digest || type digest\` - PATH="\$OLD_PATH" - - SHA_PATH=\`exec <&- 2>&-; which shasum || command -v shasum || type shasum\` - test -x "\$SHA_PATH" || SHA_PATH=\`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum\` - - if test x"\$quiet" = xn; then - MS_Printf "Verifying archive integrity..." - fi - offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` - fsize=\`cat "\$1" | wc -c | tr -d " "\` - if test \$totalsize -ne \`expr \$fsize - \$offset\`; then - echo " Unexpected archive size." >&2 - exit 2 - fi - verb=\$2 - i=1 - for s in \$filesizes - do - crc=\`echo \$CRCsum | cut -d" " -f\$i\` - if test -x "\$SHA_PATH"; then - if test x"\`basename \$SHA_PATH\`" = xshasum; then - SHA_ARG="-a 256" - fi - sha=\`echo \$SHA | cut -d" " -f\$i\` - if test x"\$sha" = x0000000000000000000000000000000000000000000000000000000000000000; then - test x"\$verb" = xy && echo " \$1 does not contain an embedded SHA256 checksum." >&2 - else - shasum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$SHA_PATH \$SHA_ARG" | cut -b-64\`; - if test x"\$shasum" != x"\$sha"; then - echo "Error in SHA256 checksums: \$shasum is different from \$sha" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " SHA256 checksums are OK." >&2 - fi - crc="0000000000"; - fi - fi - if test -x "\$MD5_PATH"; then - if test x"\`basename \$MD5_PATH\`" = xdigest; then - MD5_ARG="-a md5" - fi - md5=\`echo \$MD5 | cut -d" " -f\$i\` - if test x"\$md5" = x00000000000000000000000000000000; then - test x"\$verb" = xy && echo " \$1 does not contain an embedded MD5 checksum." >&2 - else - md5sum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$MD5_PATH \$MD5_ARG" | cut -b-32\`; - if test x"\$md5sum" != x"\$md5"; then - echo "Error in MD5 checksums: \$md5sum is different from \$md5" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " MD5 checksums are OK." >&2 - fi - crc="0000000000"; verb=n - fi - fi - if test x"\$crc" = x0000000000; then - test x"\$verb" = xy && echo " \$1 does not contain a CRC checksum." >&2 - else - sum1=\`MS_dd_Progress "\$1" \$offset \$s | CMD_ENV=xpg4 cksum | awk '{print \$1}'\` - if test x"\$sum1" != x"\$crc"; then - echo "Error in checksums: \$sum1 is different from \$crc" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " CRC checksums are OK." >&2 - fi - fi - i=\`expr \$i + 1\` - offset=\`expr \$offset + \$s\` - done - if test x"\$quiet" = xn; then - echo " All good." - fi -} - -MS_Decompress() -{ - if test x"\$decrypt_cmd" != x""; then - { eval "\$decrypt_cmd" || echo " ... Decryption failed." >&2; } | eval "$GUNZIP_CMD" - else - eval "$GUNZIP_CMD" - fi - - if test \$? -ne 0; then - echo " ... Decompression failed." >&2 - fi -} - -UnTAR() -{ - if test x"\$quiet" = xn; then - tar \$1vf - $UNTAR_EXTRA 2>&1 || { echo " ... Extraction failed." >&2; kill -15 \$$; } - else - tar \$1f - $UNTAR_EXTRA 2>&1 || { echo Extraction failed. >&2; kill -15 \$$; } - fi -} - -MS_exec_cleanup() { - if test x"\$cleanup" = xy && test x"\$cleanup_script" != x""; then - cleanup=n - cd "\$tmpdir" - eval "\"\$cleanup_script\" \$scriptargs \$cleanupargs" - fi -} - -MS_cleanup() -{ - echo 'Signal caught, cleaning up' >&2 - MS_exec_cleanup - cd "\$TMPROOT" - rm -rf "\$tmpdir" - eval \$finish; exit 15 -} - -Script_Args_Check() -{ - script_supported_args=\$(echo \${helpheader} | grep -o -E "\-\-[^ ]+" | awk -F"=" {'print \$1'}) - arg_to_test=\$(echo \$1|awk -F"=" {'print \$1'}) - - for arg in \${script_supported_args}; - do - if test x"\$arg_to_test" = x"\$arg" ;then - return - fi - done - - MS_Help - exit 1 -} - -finish=true -xterm_loop= -noprogress=$NOPROGRESS -nox11=$NOX11 -copy=$COPY -ownership=$OWNERSHIP -verbose=n -cleanup=y -cleanupargs= -sig_key= - -initargs="\$@" - -while [ -n "\$*" ] -do - case "\$1" in - -h | --help) - MS_Help - exit 0 - ;; - -q | --quiet) - quiet=y - noprogress=y - shift - ;; - --info) - echo Identification: "\$label" - echo Target directory: "\$targetdir" - echo Uncompressed size: $USIZE KB - echo Compression: $COMPRESS - if test x"$ENCRYPT" != x""; then - echo Encryption: $ENCRYPT - fi - echo Date of packaging: $DATE - echo Built with Makeself version $MS_VERSION - echo Build command was: "$MS_COMMAND" - if test x"\$script" != x; then - echo Script run after extraction: - echo " " \$script \$scriptargs - fi - if test x"$copy" = xcopy; then - echo "Archive will copy itself to a temporary location" - fi - if test x"$NEED_ROOT" = xy; then - echo "Root permissions required for extraction" - fi - if test x"$KEEP" = xy; then - echo "directory \$targetdir is permanent" - else - echo "\$targetdir will be removed after extraction" - fi - exit 0 - ;; - --list) - echo Target directory: \$targetdir - offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - for s in \$filesizes - do - MS_dd "\$0" \$offset \$s | MS_Decompress | UnTAR t - offset=\`expr \$offset + \$s\` - done - exit 0 - ;; - --tar) - offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - arg1="\$2" - shift 2 || { MS_Help; exit 1; } - for s in \$filesizes - do - MS_dd "\$0" \$offset \$s | MS_Decompress | tar "\$arg1" - "\$@" - offset=\`expr \$offset + \$s\` - done - exit 0 - ;; - --check) - MS_Check "\$0" y - scriptargs="\$scriptargs \$1" - shift - ;; - --noexec) - script="" - cleanup_script="" - shift - ;; - --extract=*) - keep=y - targetdir=\`echo \$1 | cut -d"=" -f2 \` - if ! shift; then MS_Help; exit 1; fi - ;; - --nox11) - nox11=y - shift - ;; - --xwin) - if test "$NOWAIT" = n; then - finish="echo Press Return to close this window...; read junk" - fi - xterm_loop=1 - shift - ;; - --phase2) - copy=phase2 - shift - ;; - --repack | --repack-path=*) - Script_Args_Check \$1 - scriptargs="\$scriptargs '\$1'" - shift - if [[ ! "\$1" =~ ^-.* ]]; then - scriptargs="\$scriptargs '\$1'" - shift - fi - ;; - *) - Script_Args_Check \$1 - scriptargs="\$scriptargs '\$1'" - shift - ;; - esac -done - -quiet_para="" -if test x"\$quiet" = xy; then - quiet_para="--quiet " -fi -scriptargs="--\$name_of_file""--\"\$pwd_of_file\""" \$quiet_para""\$scriptargs" - -if test x"\$quiet" = xy -a x"\$verbose" = xy; then - echo Cannot be verbose and quiet at the same time. >&2 - exit 1 -fi - -if test x"$NEED_ROOT" = xy -a \`id -u\` -ne 0; then - echo "Administrative privileges required for this archive (use su or sudo)" >&2 - exit 1 -fi - -if test x"\$copy" \!= xphase2; then - MS_PrintLicense -fi - -case "\$copy" in -copy) - tmpdir="\$TMPROOT"/makeself.\$RANDOM.\`date +"%y%m%d%H%M%S"\`.\$\$ - mkdir "\$tmpdir" || { - echo "Could not create temporary directory \$tmpdir" >&2 - exit 1 - } - SCRIPT_COPY="\$tmpdir/makeself" - echo "Copying to a temporary location..." >&2 - cp "\$0" "\$SCRIPT_COPY" - chmod +x "\$SCRIPT_COPY" - cd "\$TMPROOT" - exec "\$SCRIPT_COPY" --phase2 -- \$initargs - ;; -phase2) - finish="\$finish ; rm -rf \`dirname \$0\`" - ;; -esac - -if test x"\$nox11" = xn; then - if tty -s; then # Do we have a terminal? - : - else - if test x"\$DISPLAY" != x -a x"\$xterm_loop" = x; then # No, but do we have X? - if xset q > /dev/null 2>&1; then # Check for valid DISPLAY variable - GUESS_XTERMS="xterm gnome-terminal rxvt dtterm eterm Eterm xfce4-terminal lxterminal kvt konsole aterm terminology" - for a in \$GUESS_XTERMS; do - if type \$a >/dev/null 2>&1; then - XTERM=\$a - break - fi - done - chmod a+x \$0 || echo Please add execution rights on \$0 - if test \`echo "\$0" | cut -c1\` = "/"; then # Spawn a terminal! - exec \$XTERM -e "\$0 --xwin \$initargs" - else - exec \$XTERM -e "./\$0 --xwin \$initargs" - fi - fi - fi - fi -fi - -if test x"\$targetdir" = x.; then - tmpdir="." -else - if test x"\$keep" = xy; then - if test x"\$nooverwrite" = xy && test -d "\$targetdir"; then - echo "Target directory \$targetdir already exists, aborting." >&2 - exit 1 - fi - if test x"\$quiet" = xn; then - echo "Creating directory \$targetdir" >&2 - fi - tmpdir="\$targetdir" - dashp="-p" - else - tmpdir="\$TMPROOT/selfgz\$\$\$RANDOM" - dashp="" - fi - mkdir \$dashp "\$tmpdir" || { - echo 'Cannot create target directory' \$tmpdir >&2 - echo 'You should try option --extract=' >&2 - eval \$finish - exit 1 - } -fi - -location="\`pwd\`" -if test x"\$SETUP_NOCHECK" != x1; then - MS_Check "\$0" -fi -offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - -if test x"\$verbose" = xy; then - MS_Printf "About to extract $USIZE KB in \$tmpdir ... Proceed ? [Y/n] " - read yn - if test x"\$yn" = xn; then - eval \$finish; exit 1 - fi -fi - -if test x"\$quiet" = xn; then - # Decrypting with openssl will ask for password, - # the prompt needs to start on new line - if test x"$ENCRYPT" = x"openssl"; then - echo "Decrypting and uncompressing \$label..." - else - MS_Printf "Uncompressing \$label" - fi -fi -res=3 -if test x"\$keep" = xn; then - trap MS_cleanup 1 2 3 15 -fi - -if test x"\$nodiskspace" = xn; then - leftspace=\`MS_diskspace "\$tmpdir"\` - if test -n "\$leftspace"; then - if test "\$leftspace" -lt $USIZE; then - echo - echo "Not enough space left in "\`dirname \$tmpdir\`" (\$leftspace KB) to decompress \$0 ($USIZE KB)" >&2 - if test x"\$keep" = xn; then - echo "Consider setting TMPDIR to a directory with more free space." - fi - eval \$finish; exit 1 - fi - fi -fi - -for s in \$filesizes -do - if MS_dd_Progress "\$0" \$offset \$s | MS_Decompress | ( cd "\$tmpdir"; umask \$ORIG_UMASK ; UnTAR xp ) 1>/dev/null; then - if test x"\$ownership" = xy; then - (cd "\$tmpdir"; chown -R \`id -u\` .; chgrp -R \`id -g\` .) - fi - else - echo >&2 - echo "Unable to decompress \$0" >&2 - eval \$finish; exit 1 - fi - offset=\`expr \$offset + \$s\` -done -if test x"\$quiet" = xn; then - echo -fi - -cd "\$tmpdir" -res=0 -if test x"\$script" != x; then - if test x"\$export_conf" = x"y"; then - MS_BUNDLE="\$0" - MS_LABEL="\$label" - MS_SCRIPT="\$script" - MS_SCRIPTARGS="\$scriptargs" - MS_ARCHDIRNAME="\$archdirname" - MS_KEEP="\$KEEP" - MS_NOOVERWRITE="\$NOOVERWRITE" - MS_COMPRESS="\$COMPRESS" - MS_CLEANUP="\$cleanup" - export MS_BUNDLE MS_LABEL MS_SCRIPT MS_SCRIPTARGS - export MS_ARCHDIRNAME MS_KEEP MS_NOOVERWRITE MS_COMPRESS - fi - - if test x"\$verbose" = x"y"; then - yn="x" - while test x"\$yn" != x -a x"\$yn" != xy -a x"\$yn" != xY -a x"\$yn" != xn -a x"\$yn" != xN - do - MS_Printf "OK to execute: \$script \$scriptargs \$* ? [Y/n] " - read yn - if test x"\$yn" = x -o x"\$yn" = xy -o x"\$yn" = xY; then - eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$?; - elif test x"\$yn" = xn -o x"\$yn" = xN; then - echo "Unable to decompress \$script ,because of aborting! ";res=\$? - else - echo "Input value is unacceptable,please try again." - fi - done - else - eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$? - fi - if test "\$res" -ne 0; then - test x"\$verbose" = xy && echo "The program '\$script' returned an error code (\$res)" >&2 - fi -fi - -MS_exec_cleanup - -if test x"\$keep" = xn; then - cd "\$TMPROOT" - rm -rf "\$tmpdir" -fi -eval \$finish; exit \$res -EOF diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 deleted file mode 100644 index 81bf6e4ff..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 +++ /dev/null @@ -1,110 +0,0 @@ -.TH "MAKESELF" "1" "2.4.5" -.SH "NAME" -makeself \- An utility to generate self-extractable archives. -.SH "SYNTAX" -.B makeself [\fIoptions\fP] archive_dir file_name label -.B [\fIstartup_script\fP] [\fIargs\fP] -.SH "DESCRIPTION" -This program is a free (GPL) utility designed to create self-extractable -archives from a directory. -.SH "OPTIONS" -The following options are supported. -.TP 15 -.B -v, --version -Prints out the makeself version number and exits. -.TP -.B -h, --help -Print out help information. -.TP -.B --tar-quietly -Suppress verbose output from the tar command -.TP -.B --quiet -Do not print any messages other than errors -.TP -.B --gzip -Compress using gzip (default if detected). -.TP -.B --bzip2 -Compress using bzip2. -.TP -.B --pbzip2 -Compress using pbzip2. -.TP -.B --xz -Compress using xz. -.TP -.B --lzo -Compress using lzop. -.TP -.B --lz4 -Compress using lz4. -.TP -.B --compress -Compress using the UNIX 'compress' command. -.TP -.B --nocomp -Do not compress the data. -.TP -.B --complevel lvl -Specify the compression level for gzip,bzip2,pbzui2,xz,lzo or lz4 -.TP -.B --notemp -The archive will create archive_dir in the current directory and -uncompress in ./archive_dir. -.TP -.B --copy -Upon extraction, the archive will first copy itself to a temporary directory. -.TP -.B --append -Append more files to an existing makeself archive. The label and startup scripts will then be ignored. -.TP -.B --current -Files will be extracted to the current directory. Both --current and --target dir imply --notemp. -.TP -.B --target dir -Extract directly to a target directory. Directory path can be either absolute or relative. -.TP -.B --header file -Specify location of the header script. -.TP -.B --cleanup file -Specify a cleanup script that executes on interrupt and when finished successfully. -.TP -.B --follow -Follow the symlinks in the archive. -.TP -.B --noprogress -Do not show the progress during the decompression. -.TP -.B --nox11 -Disable automatic spawn of an xterm if running in X11. -.TP -.B --nowait -Do not wait for user input after executing embedded program from an xterm. -.TP -.B --nomd5 -Do not create a MD5 checksum for the archive. -.TP -.B --nocrc -Do not create a CRC32 checksum for the archive. -.TP -.B --lsm file -LSM file describing the package. -.B --packaging-date date -Use provided string as the packaging date instead of the current date. -.SH "EXAMPLES" -Here is an example, assuming the user has a package image stored in a /home/joe/mysoft, -and he wants to generate a self-extracting package named mysoft.sh, which will launch -the "setup" script initially stored in /home/joe/mysoft: -.TP -makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup -.TP -Here is also how I created the makeself.run archive which contains the Makeself distribution: -.TP -makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" -.SH "AUTHORS" -Makeself has been written by Stéphane Peter . -.BR -This man page was originally written by Bartosz Fenski for the -Debian GNU/Linux distribution (but it may be used by others). diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm deleted file mode 100644 index 3c4cea8c1..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm +++ /dev/null @@ -1,16 +0,0 @@ -Begin3 -Title: makeself.sh -Version: 2.4.5 -Description: makeself.sh is a shell script that generates a self-extractable - tar.gz archive from a directory. The resulting file appears as a shell - script, and can be launched as is. The archive will then uncompress - itself to a temporary directory and an arbitrary command will be - executed (for example an installation script). This is pretty similar - to archives generated with WinZip Self-Extractor in the Windows world. -Keywords: Installation archive tar winzip -Author: Stephane Peter (megastep@megastep.org) -Maintained-by: Stephane Peter (megastep@megastep.org) -Original-site: https://makeself.io/ -Platform: Unix -Copying-policy: GPL -End diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh deleted file mode 100755 index c8ea56597..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh +++ /dev/null @@ -1,822 +0,0 @@ -#!/bin/sh -# -# Makeself version 2.4.x -# by Stephane Peter -# -# Utility to create self-extracting tar.gz archives. -# The resulting archive is a file holding the tar.gz archive with -# a small Shell script stub that uncompresses the archive to a temporary -# directory and then executes a given script from withing that directory. -# -# Makeself home page: https://makeself.io/ -# -# Version 2.0 is a rewrite of version 1.0 to make the code easier to read and maintain. -# -# Version history : -# - 1.0 : Initial public release -# - 1.1 : The archive can be passed parameters that will be passed on to -# the embedded script, thanks to John C. Quillan -# - 1.2 : Package distribution, bzip2 compression, more command line options, -# support for non-temporary archives. Ideas thanks to Francois Petitjean -# - 1.3 : More patches from Bjarni R. Einarsson and Francois Petitjean: -# Support for no compression (--nocomp), script is no longer mandatory, -# automatic launch in an xterm, optional verbose output, and -target -# archive option to indicate where to extract the files. -# - 1.4 : Improved UNIX compatibility (Francois Petitjean) -# Automatic integrity checking, support of LSM files (Francois Petitjean) -# - 1.5 : Many bugfixes. Optionally disable xterm spawning. -# - 1.5.1 : More bugfixes, added archive options -list and -check. -# - 1.5.2 : Cosmetic changes to inform the user of what's going on with big -# archives (Quake III demo) -# - 1.5.3 : Check for validity of the DISPLAY variable before launching an xterm. -# More verbosity in xterms and check for embedded command's return value. -# Bugfix for Debian 2.0 systems that have a different "print" command. -# - 1.5.4 : Many bugfixes. Print out a message if the extraction failed. -# - 1.5.5 : More bugfixes. Added support for SETUP_NOCHECK environment variable to -# bypass checksum verification of archives. -# - 1.6.0 : Compute MD5 checksums with the md5sum command (patch from Ryan Gordon) -# - 2.0 : Brand new rewrite, cleaner architecture, separated header and UNIX ports. -# - 2.0.1 : Added --copy -# - 2.1.0 : Allow multiple tarballs to be stored in one archive, and incremental updates. -# Added --nochown for archives -# Stopped doing redundant checksums when not necesary -# - 2.1.1 : Work around insane behavior from certain Linux distros with no 'uncompress' command -# Cleaned up the code to handle error codes from compress. Simplified the extraction code. -# - 2.1.2 : Some bug fixes. Use head -n to avoid problems. -# - 2.1.3 : Bug fixes with command line when spawning terminals. -# Added --tar for archives, allowing to give arbitrary arguments to tar on the contents of the archive. -# Added --noexec to prevent execution of embedded scripts. -# Added --nomd5 and --nocrc to avoid creating checksums in archives. -# Added command used to create the archive in --info output. -# Run the embedded script through eval. -# - 2.1.4 : Fixed --info output. -# Generate random directory name when extracting files to . to avoid problems. (Jason Trent) -# Better handling of errors with wrong permissions for the directory containing the files. (Jason Trent) -# Avoid some race conditions (Ludwig Nussel) -# Unset the $CDPATH variable to avoid problems if it is set. (Debian) -# Better handling of dot files in the archive directory. -# - 2.1.5 : Made the md5sum detection consistent with the header code. -# Check for the presence of the archive directory -# Added --encrypt for symmetric encryption through gpg (Eric Windisch) -# Added support for the digest command on Solaris 10 for MD5 checksums -# Check for available disk space before extracting to the target directory (Andreas Schweitzer) -# Allow extraction to run asynchronously (patch by Peter Hatch) -# Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo) -# - 2.1.6 : Replaced one dot per file progress with a realtime progress percentage and a spining cursor (Guy Baconniere) -# Added --noprogress to prevent showing the progress during the decompression (Guy Baconniere) -# Added --target dir to allow extracting directly to a target directory (Guy Baconniere) -# - 2.2.0 : Many bugfixes, updates and contributions from users. Check out the project page on Github for the details. -# - 2.3.0 : Option to specify packaging date to enable byte-for-byte reproducibility. (Marc Pawlowsky) -# - 2.4.0 : Optional support for SHA256 checksums in archives. -# - 2.4.2 : Add support for threads for several compressors. (M. Limber) -# Added zstd support. -# - 2.4.3 : Make explicit POSIX tar archives for increased compatibility. -# - 2.4.5 : Added --tar-format to override ustar tar archive format -# -# (C) 1998-2021 by Stephane Peter -# -# This software is released under the terms of the GNU GPL version 2 and above -# Please read the license at http://www.gnu.org/copyleft/gpl.html -# Self-extracting archives created with this script are explictly NOT released under the term of the GPL -# - -MS_VERSION=2.4.5 -MS_COMMAND="$0" -unset CDPATH - -for f in ${1+"$@"}; do - MS_COMMAND="$MS_COMMAND \\\\ - \\\"$f\\\"" -done - -# For Solaris systems -if test -d /usr/xpg4/bin; then - PATH=/usr/xpg4/bin:$PATH - export PATH -fi - -# Procedures - -MS_Usage() -{ - echo "Usage: $0 [args] archive_dir file_name label startup_script [script_args]" - echo "args can be one or more of the following :" - echo " --version | -v : Print out Makeself version number and exit" - echo " --help | -h : Print out this help message" - echo " --tar-quietly : Suppress verbose output from the tar command" - echo " --quiet | -q : Do not print any messages other than errors." - echo " --gzip : Compress using gzip (default if detected)" - echo " --pigz : Compress with pigz" - echo " --zstd : Compress with zstd" - echo " --bzip2 : Compress using bzip2 instead of gzip" - echo " --pbzip2 : Compress using pbzip2 instead of gzip" - echo " --xz : Compress using xz instead of gzip" - echo " --lzo : Compress using lzop instead of gzip" - echo " --lz4 : Compress using lz4 instead of gzip" - echo " --compress : Compress using the UNIX 'compress' command" - echo " --complevel lvl : Compression level for gzip pigz zstd xz lzo lz4 bzip2 and pbzip2 (default 9)" - echo " --threads thds : Number of threads to be used by compressors that support parallelization." - echo " Omit to use compressor's default. Most useful (and required) for opting" - echo " into xz's threading, usually with '--threads=0' for all available cores." - echo " pbzip2 and pigz are parallel by default, and setting this value allows" - echo " limiting the number of threads they use." - echo " --base64 : Instead of compressing, encode the data using base64" - echo " --gpg-encrypt : Instead of compressing, encrypt the data using GPG" - echo " --gpg-asymmetric-encrypt-sign" - echo " : Instead of compressing, asymmetrically encrypt and sign the data using GPG" - echo " --gpg-extra opt : Append more options to the gpg command line" - echo " --ssl-encrypt : Instead of compressing, encrypt the data using OpenSSL" - echo " --ssl-passwd pass : Use the given password to encrypt the data using OpenSSL" - echo " --ssl-pass-src src : Use the given src as the source of password to encrypt the data" - echo " using OpenSSL. See \"PASS PHRASE ARGUMENTS\" in man openssl." - echo " If this option is not supplied, the user will be asked to enter" - echo " encryption password on the current terminal." - echo " --ssl-no-md : Do not use \"-md\" option not supported by older OpenSSL." - echo " --nochown : Do not give the target folder to the current user (default)" - echo " --chown : Give the target folder to the current user recursively" - echo " --nocomp : Do not compress the data" - echo " --notemp : The archive will create archive_dir in the" - echo " current directory and uncompress in ./archive_dir" - echo " --needroot : Check that the root user is extracting the archive before proceeding" - echo " --copy : Upon extraction, the archive will first copy itself to" - echo " a temporary directory" - echo " --append : Append more files to an existing Makeself archive" - echo " The label and startup scripts will then be ignored" - echo " --target dir : Extract directly to a target directory" - echo " directory path can be either absolute or relative" - echo " --nooverwrite : Do not extract the archive if the specified target directory exists" - echo " --current : Files will be extracted to the current directory" - echo " Both --current and --target imply --notemp" - echo " --tar-format opt : Specify a tar archive format (default is ustar)" - echo " --tar-extra opt : Append more options to the tar command line" - echo " --untar-extra opt : Append more options to the during the extraction of the tar archive" - echo " --nomd5 : Don't calculate an MD5 for archive" - echo " --nocrc : Don't calculate a CRC for archive" - echo " --sha256 : Compute a SHA256 checksum for the archive" - echo " --header file : Specify location of the header script" - echo " --cleanup file : Specify a cleanup script that executes on interrupt and when finished successfully." - echo " --follow : Follow the symlinks in the archive" - echo " --noprogress : Do not show the progress during the decompression" - echo " --nox11 : Disable automatic spawn of a xterm" - echo " --nowait : Do not wait for user input after executing embedded" - echo " program from an xterm" - echo " --sign passphrase : Signature private key to sign the package with" - echo " --lsm file : LSM file describing the package" - echo " --license file : Append a license file" - echo " --help-header file : Add a header to the archive's --help output" - echo " --packaging-date date" - echo " : Use provided string as the packaging date" - echo " instead of the current date." - echo - echo " --keep-umask : Keep the umask set to shell default, rather than overriding when executing self-extracting archive." - echo " --export-conf : Export configuration variables to startup_script" - echo - echo "Do not forget to give a fully qualified startup script name" - echo "(i.e. with a ./ prefix if inside the archive)." - exit 1 -} - -# Default settings -if type gzip >/dev/null 2>&1; then - COMPRESS=gzip -elif type compress >/dev/null 2>&1; then - COMPRESS=compress -else - echo "ERROR: missing commands: gzip, compress" >&2 - MS_Usage -fi -ENCRYPT=n -PASSWD="" -PASSWD_SRC="" -OPENSSL_NO_MD=n -COMPRESS_LEVEL=9 -DEFAULT_THREADS=123456 # Sentinel value -THREADS=$DEFAULT_THREADS -KEEP=n -CURRENT=n -NOX11=n -NOWAIT=n -APPEND=n -TAR_QUIETLY=n -KEEP_UMASK=n -QUIET=n -NOPROGRESS=n -COPY=none -NEED_ROOT=n -TAR_ARGS=rvf -TAR_FORMAT=ustar -TAR_EXTRA="" -GPG_EXTRA="" -DU_ARGS=-ks -HEADER=`dirname "$0"`/makeself-header.sh -SIGNATURE="" -TARGETDIR="" -NOOVERWRITE=n -DATE=`LC_ALL=C date` -EXPORT_CONF=n -SHA256=n -OWNERSHIP=n -SIGN=n -GPG_PASSPHRASE="" - -# LSM file stuff -LSM_CMD="echo No LSM. >> \"\$archname\"" - -while true -do - case "$1" in - --version | -v) - echo Makeself version $MS_VERSION - exit 0 - ;; - --pbzip2) - COMPRESS=pbzip2 - shift - ;; - --bzip2) - COMPRESS=bzip2 - shift - ;; - --gzip) - COMPRESS=gzip - shift - ;; - --pigz) - COMPRESS=pigz - shift - ;; - --zstd) - COMPRESS=zstd - shift - ;; - --xz) - COMPRESS=xz - shift - ;; - --lzo) - COMPRESS=lzo - shift - ;; - --lz4) - COMPRESS=lz4 - shift - ;; - --compress) - COMPRESS=compress - shift - ;; - --base64) - COMPRESS=base64 - shift - ;; - --gpg-encrypt) - COMPRESS=gpg - shift - ;; - --gpg-asymmetric-encrypt-sign) - COMPRESS=gpg-asymmetric - shift - ;; - --gpg-extra) - GPG_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-encrypt) - ENCRYPT=openssl - shift - ;; - --ssl-passwd) - PASSWD=$2 - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-pass-src) - PASSWD_SRC=$2 - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-no-md) - OPENSSL_NO_MD=y - shift - ;; - --nocomp) - COMPRESS=none - shift - ;; - --complevel) - COMPRESS_LEVEL="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --threads) - THREADS="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --nochown) - OWNERSHIP=n - shift - ;; - --chown) - OWNERSHIP=y - shift - ;; - --notemp) - KEEP=y - shift - ;; - --copy) - COPY=copy - shift - ;; - --current) - CURRENT=y - KEEP=y - shift - ;; - --tar-format) - TAR_FORMAT="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --tar-extra) - TAR_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --untar-extra) - UNTAR_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --target) - TARGETDIR="$2" - KEEP=y - shift 2 || { MS_Usage; exit 1; } - ;; - --sign) - SIGN=y - GPG_PASSPHRASE="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --nooverwrite) - NOOVERWRITE=y - shift - ;; - --needroot) - NEED_ROOT=y - shift - ;; - --header) - HEADER="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --cleanup) - CLEANUP_SCRIPT="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --license) - # We need to escape all characters having a special meaning in double quotes - LICENSE=$(sed 's/\\/\\\\/g; s/"/\\\"/g; s/`/\\\`/g; s/\$/\\\$/g' "$2") - shift 2 || { MS_Usage; exit 1; } - ;; - --follow) - TAR_ARGS=rvhf - DU_ARGS=-ksL - shift - ;; - --noprogress) - NOPROGRESS=y - shift - ;; - --nox11) - NOX11=y - shift - ;; - --nowait) - NOWAIT=y - shift - ;; - --nomd5) - NOMD5=y - shift - ;; - --sha256) - SHA256=y - shift - ;; - --nocrc) - NOCRC=y - shift - ;; - --append) - APPEND=y - shift - ;; - --lsm) - LSM_CMD="cat \"$2\" >> \"\$archname\"" - shift 2 || { MS_Usage; exit 1; } - ;; - --packaging-date) - DATE="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --help-header) - HELPHEADER=`sed -e "s/'/'\\\\\''/g" $2` - shift 2 || { MS_Usage; exit 1; } - [ -n "$HELPHEADER" ] && HELPHEADER="$HELPHEADER -" - ;; - --tar-quietly) - TAR_QUIETLY=y - shift - ;; - --keep-umask) - KEEP_UMASK=y - shift - ;; - --export-conf) - EXPORT_CONF=y - shift - ;; - -q | --quiet) - QUIET=y - shift - ;; - -h | --help) - MS_Usage - ;; - -*) - echo Unrecognized flag : "$1" - MS_Usage - ;; - *) - break - ;; - esac -done - -if test $# -lt 1; then - MS_Usage -else - if test -d "$1"; then - archdir="$1" - else - echo "Directory $1 does not exist." >&2 - exit 1 - fi -fi -archname="$2" - -if test "$QUIET" = "y" || test "$TAR_QUIETLY" = "y"; then - if test "$TAR_ARGS" = "rvf"; then - TAR_ARGS="rf" - elif test "$TAR_ARGS" = "rvhf"; then - TAR_ARGS="rhf" - fi -fi - -if test "$APPEND" = y; then - if test $# -lt 2; then - MS_Usage - fi - - # Gather the info from the original archive - OLDENV=`sh "$archname" --dumpconf` - if test $? -ne 0; then - echo "Unable to update archive: $archname" >&2 - exit 1 - else - eval "$OLDENV" - OLDSKIP=`expr $SKIP + 1` - fi -else - if test "$KEEP" = n -a $# = 3; then - echo "ERROR: Making a temporary archive with no embedded command does not make sense!" >&2 - echo >&2 - MS_Usage - fi - # We don't want to create an absolute directory unless a target directory is defined - if test "$CURRENT" = y; then - archdirname="." - elif test x"$TARGETDIR" != x; then - archdirname="$TARGETDIR" - else - archdirname=`basename "$1"` - fi - - if test $# -lt 3; then - MS_Usage - fi - - LABEL="$3" - SCRIPT="$4" - test "x$SCRIPT" = x || shift 1 - shift 3 - SCRIPTARGS="$*" -fi - -if test "$KEEP" = n -a "$CURRENT" = y; then - echo "ERROR: It is A VERY DANGEROUS IDEA to try to combine --notemp and --current." >&2 - exit 1 -fi - -case $COMPRESS in -gzip) - GZIP_CMD="gzip -c$COMPRESS_LEVEL" - GUNZIP_CMD="gzip -cd" - ;; -pigz) - GZIP_CMD="pigz -$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD --processes $THREADS" - fi - GUNZIP_CMD="gzip -cd" - ;; -zstd) - GZIP_CMD="zstd -$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD --threads=$THREADS" - fi - GUNZIP_CMD="zstd -cd" - ;; -pbzip2) - GZIP_CMD="pbzip2 -c$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD -p$THREADS" - fi - GUNZIP_CMD="bzip2 -d" - ;; -bzip2) - GZIP_CMD="bzip2 -$COMPRESS_LEVEL" - GUNZIP_CMD="bzip2 -d" - ;; -xz) - GZIP_CMD="xz -c$COMPRESS_LEVEL" - # Must opt-in by specifying a value since not all versions of xz support threads - if test $THREADS -ne $DEFAULT_THREADS; then - GZIP_CMD="$GZIP_CMD --threads=$THREADS" - fi - GUNZIP_CMD="xz -d" - ;; -lzo) - GZIP_CMD="lzop -c$COMPRESS_LEVEL" - GUNZIP_CMD="lzop -d" - ;; -lz4) - GZIP_CMD="lz4 -c$COMPRESS_LEVEL" - GUNZIP_CMD="lz4 -d" - ;; -base64) - GZIP_CMD="base64" - GUNZIP_CMD="base64 --decode -i -" - ;; -gpg) - GZIP_CMD="gpg $GPG_EXTRA -ac -z$COMPRESS_LEVEL" - GUNZIP_CMD="gpg -d" - ENCRYPT="gpg" - ;; -gpg-asymmetric) - GZIP_CMD="gpg $GPG_EXTRA -z$COMPRESS_LEVEL -es" - GUNZIP_CMD="gpg --yes -d" - ENCRYPT="gpg" - ;; -compress) - GZIP_CMD="compress -fc" - GUNZIP_CMD="(type compress >/dev/null 2>&1 && compress -fcd || gzip -cd)" - ;; -none) - GZIP_CMD="cat" - GUNZIP_CMD="cat" - ;; -esac - -if test x"$ENCRYPT" = x"openssl"; then - if test x"$APPEND" = x"y"; then - echo "Appending to existing archive is not compatible with OpenSSL encryption." >&2 - fi - - ENCRYPT_CMD="openssl enc -aes-256-cbc -salt" - DECRYPT_CMD="openssl enc -aes-256-cbc -d" - - if test x"$OPENSSL_NO_MD" != x"y"; then - ENCRYPT_CMD="$ENCRYPT_CMD -md sha256" - DECRYPT_CMD="$DECRYPT_CMD -md sha256" - fi - - if test -n "$PASSWD_SRC"; then - ENCRYPT_CMD="$ENCRYPT_CMD -pass $PASSWD_SRC" - elif test -n "$PASSWD"; then - ENCRYPT_CMD="$ENCRYPT_CMD -pass pass:$PASSWD" - fi -fi - -tmpfile="${TMPDIR:-/tmp}/mkself$$" - -if test -f "$HEADER"; then - oldarchname="$archname" - archname="$tmpfile" - # Generate a fake header to count its lines - SKIP=0 - . "$HEADER" - SKIP=`cat "$tmpfile" |wc -l` - # Get rid of any spaces - SKIP=`expr $SKIP` - rm -f "$tmpfile" - if test "$QUIET" = "n"; then - echo "Header is $SKIP lines long" >&2 - fi - archname="$oldarchname" -else - echo "Unable to open header file: $HEADER" >&2 - exit 1 -fi - -if test "$QUIET" = "n"; then - echo -fi - -if test "$APPEND" = n; then - if test -f "$archname"; then - echo "WARNING: Overwriting existing file: $archname" >&2 - fi -fi - -USIZE=`du $DU_ARGS "$archdir" | awk '{print $1}'` - -if test "." = "$archdirname"; then - if test "$KEEP" = n; then - archdirname="makeself-$$-`date +%Y%m%d%H%M%S`" - fi -fi - -test -d "$archdir" || { echo "Error: $archdir does not exist."; rm -f "$tmpfile"; exit 1; } -if test "$QUIET" = "n"; then - echo "About to compress $USIZE KB of data..." - echo "Adding files to archive named \"$archname\"..." -fi - -# See if we have GNU tar -TAR=`exec <&- 2>&-; which gtar || command -v gtar || type gtar` -test -x "$TAR" || TAR=tar - -tmparch="${TMPDIR:-/tmp}/mkself$$.tar" -( - if test "$APPEND" = "y"; then - tail -n "+$OLDSKIP" "$archname" | eval "$GUNZIP_CMD" > "$tmparch" - fi - cd "$archdir" - # "Determining if a directory is empty" - # https://www.etalabs.net/sh_tricks.html - find . \ - \( \ - ! -type d \ - -o \ - \( -links 2 -exec sh -c ' - is_empty () ( - cd "$1" - set -- .[!.]* ; test -f "$1" && return 1 - set -- ..?* ; test -f "$1" && return 1 - set -- * ; test -f "$1" && return 1 - return 0 - ) - is_empty "$0"' {} \; \ - \) \ - \) -print \ - | LC_ALL=C sort \ - | sed 's/./\\&/g' \ - | xargs $TAR $TAR_EXTRA --format $TAR_FORMAT -$TAR_ARGS "$tmparch" -) || { - echo "ERROR: failed to create temporary archive: $tmparch" - rm -f "$tmparch" "$tmpfile" - exit 1 -} - -USIZE=`du $DU_ARGS "$tmparch" | awk '{print $1}'` - -eval "$GZIP_CMD" <"$tmparch" >"$tmpfile" || { - echo "ERROR: failed to create temporary file: $tmpfile" - rm -f "$tmparch" "$tmpfile" - exit 1 -} -rm -f "$tmparch" - -if test x"$ENCRYPT" = x"openssl"; then - echo "About to encrypt archive \"$archname\"..." - { eval "$ENCRYPT_CMD -in $tmpfile -out ${tmpfile}.enc" && mv -f ${tmpfile}.enc $tmpfile; } || \ - { echo Aborting: could not encrypt temporary file: "$tmpfile".; rm -f "$tmpfile"; exit 1; } -fi - -fsize=`cat "$tmpfile" | wc -c | tr -d " "` - -# Compute the checksums - -shasum=0000000000000000000000000000000000000000000000000000000000000000 -md5sum=00000000000000000000000000000000 -crcsum=0000000000 - -if test "$NOCRC" = y; then - if test "$QUIET" = "n"; then - echo "skipping crc at user request" - fi -else - crcsum=`CMD_ENV=xpg4 cksum < "$tmpfile" | sed -e 's/ /Z/' -e 's/ /Z/' | cut -dZ -f1` - if test "$QUIET" = "n"; then - echo "CRC: $crcsum" - fi -fi - -if test "$SHA256" = y; then - SHA_PATH=`exec <&- 2>&-; which shasum || command -v shasum || type shasum` - if test -x "$SHA_PATH"; then - shasum=`eval "$SHA_PATH -a 256" < "$tmpfile" | cut -b-64` - else - SHA_PATH=`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum` - shasum=`eval "$SHA_PATH" < "$tmpfile" | cut -b-64` - fi - if test "$QUIET" = "n"; then - if test -x "$SHA_PATH"; then - echo "SHA256: $shasum" - else - echo "SHA256: none, SHA command not found" - fi - fi -fi -if test "$NOMD5" = y; then - if test "$QUIET" = "n"; then - echo "Skipping md5sum at user request" - fi -else - # Try to locate a MD5 binary - OLD_PATH=$PATH - PATH=${GUESS_MD5_PATH:-"$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} - MD5_ARG="" - MD5_PATH=`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum` - test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which md5 || command -v md5 || type md5` - test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which digest || command -v digest || type digest` - PATH=$OLD_PATH - if test -x "$MD5_PATH"; then - if test `basename ${MD5_PATH}`x = digestx; then - MD5_ARG="-a md5" - fi - md5sum=`eval "$MD5_PATH $MD5_ARG" < "$tmpfile" | cut -b-32` - if test "$QUIET" = "n"; then - echo "MD5: $md5sum" - fi - else - if test "$QUIET" = "n"; then - echo "MD5: none, MD5 command not found" - fi - fi -fi -if test "$SIGN" = y; then - GPG_PATH=`exec <&- 2>&-; which gpg || command -v gpg || type gpg` - if test -x "$GPG_PATH"; then - SIGNATURE=`$GPG_PATH --pinentry-mode=loopback --batch --yes --passphrase "$GPG_PASSPHRASE" --output - --detach-sig $tmpfile | base64 | tr -d \\\\n` - if test "$QUIET" = "n"; then - echo "Signature: $SIGNATURE" - fi - else - echo "Missing gpg command" >&2 - fi -fi - -totalsize=0 -for size in $fsize; -do - totalsize=`expr $totalsize + $size` -done - -if test "$APPEND" = y; then - mv "$archname" "$archname".bak || exit - - # Prepare entry for new archive - filesizes="$fsize" - CRCsum="$crcsum" - MD5sum="$md5sum" - SHAsum="$shasum" - Signature="$SIGNATURE" - # Generate the header - . "$HEADER" - # Append the new data - cat "$tmpfile" >> "$archname" - - chmod +x "$archname" - rm -f "$archname".bak - if test "$QUIET" = "n"; then - echo "Self-extractable archive \"$archname\" successfully updated." - fi -else - filesizes="$fsize" - CRCsum="$crcsum" - MD5sum="$md5sum" - SHAsum="$shasum" - Signature="$SIGNATURE" - - # Generate the header - . "$HEADER" - - # Append the compressed tar data after the stub - if test "$QUIET" = "n"; then - echo - fi - cat "$tmpfile" >> "$archname" - chmod +x "$archname" - if test "$QUIET" = "n"; then - echo Self-extractable archive \"$archname\" successfully created. - fi -fi -rm -f "$tmpfile" diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh deleted file mode 100644 index 31ee16511..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh +++ /dev/null @@ -1,8 +0,0 @@ -#!/bin/sh -# Run every available test - Bash needed -cd test -for test in *test; -do - echo "Running test $test ..." - bash $test || { echo "*** ERROR: Test '$test' failed!"; exit 1; } -done diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh deleted file mode 100755 index a977bd51d..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh +++ /dev/null @@ -1,31 +0,0 @@ -#!/bin/bash - -project_path=$1 -build_path=$2 -vendor_name=customize -echo $@ -if [[ ! -d "$project_path" ]]; then - echo "[ERROR] No projcet path is provided" - exit 1 -fi - -if [[ ! -d "$build_path" ]]; then - echo "[ERROR] No build path is provided" - exit 1 -fi - -if [[ ! -d "$ASCEND_OPP_PATH" ]]; then - echo "[ERROR] No opp install path is provided" - exit 1 -fi -custom_exist_info_json=$ASCEND_OPP_PATH/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json -custom_new_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json -temp_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/temp_cust_aicpu_kernel.json - -if [[ -f "$custom_exist_info_json" ]] && [[ -f "$custom_new_info_json" ]]; then - cp -f $custom_exist_info_json $temp_info_json - chmod +w $temp_info_json - python3 ${project_path}/cmake/util/insert_op_info.py ${custom_new_info_json} ${temp_info_json} - cp -f $temp_info_json $custom_new_info_json - rm -f $temp_info_json -fi diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py deleted file mode 100755 index c38e79edc..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py +++ /dev/null @@ -1,260 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os - - -OP_ALL = '__ALLOP__' -SOC_ALL = '__ALLSOC__' -SOC_TO_SHORT_SOC_MAP = { - "ascend910a": "ascend910", - "ascend910proa": "ascend910", - "ascend910b": "ascend910", - "ascend910prob": "ascend910", - "ascend910premiuma": "ascend910", - "ascend910b1": "ascend910b", - "ascend910b2": "ascend910b", - "ascend910b3": "ascend910b", - "ascend910b4": "ascend910b", - "ascend310p1": "ascend310p", - "ascend310p3": "ascend310p", - "ascend310p3vir01": "ascend310p", - "ascend310p3vir02": "ascend310p", - "ascend310p3vir04": "ascend310p", - "ascend310p3vir08": "ascend310p", - "ascend310b1": "ascend310b", - "bs9sx1aa": "bs9sx1a" -} - - -class OpDesc: - def __init__(self: any, op_type: str): - self.op_type = op_type - self.attr_list = [] - self.attr_val = {} - self.input_name = [] - self.input_type = [] - self.input_dtype = [] - self.input_fmt = [] - self.output_name = [] - self.output_type = [] - self.output_dtype = [] - self.output_fmt = [] - self.op_fmt_sel = False - self.op_chk_support = False - self.op_intf = '' - self.kern_name = '' - self.op_file = '' - self.op_replay_flag = False - self.op_replay_batch = False - self.input_idx = -1 - self.output_idx = -1 - self.max_block_dim = 32 - self.max_shape_size = 268435456 - self.dynamic_shape = False - self.op_range_limit = '' - self.custom_compile_options = {} - self.custom_all_compile_options = {} - - @staticmethod - def _parse_digit(conf: str) -> int: - return int(conf.split('=')[1]) - - @staticmethod - def _parse_flag(conf: str) -> bool: - if 'true' == conf.split('=')[1]: - return True - return False - - @staticmethod - def _parse_str(conf: str) -> str: - return conf.split('=')[1] - - @staticmethod - def _parse_list(conf: str) -> list: - return conf.split('=')[1].split(',') - - def parse_input(self: any, conf: str): - if conf.startswith('input{}.name'.format(int(self.input_idx) + 1)): - self.input_idx += 1 - self.input_name.append(self._parse_str(conf)) - elif conf.startswith('input{}.paramType'.format(int(self.input_idx))): - self.input_type.append(self._parse_str(conf)) - elif conf.startswith('input{}.dtype'.format(int(self.input_idx))): - self.input_dtype.append(self._parse_str(conf)) - elif conf.startswith('input{}.format'.format(int(self.input_idx))): - self.input_fmt.append(self._parse_str(conf)) - else: - return - - def parse_output(self: any, conf: str): - if conf.startswith('output{}.name'.format(int(self.output_idx) + 1)): - self.output_idx += 1 - self.output_name.append(self._parse_str(conf)) - elif conf.startswith('output{}.paramType'.format(int(self.output_idx))): - self.output_type.append(self._parse_str(conf)) - elif conf.startswith('output{}.dtype'.format(int(self.output_idx))): - self.output_dtype.append(self._parse_str(conf)) - elif conf.startswith('output{}.format'.format(int(self.output_idx))): - self.output_fmt.append(self._parse_str(conf)) - else: - return - - def parse_op_format(self: any, conf: str): - self.op_fmt_sel = self._parse_flag(conf) - - def parse_check_support(self: any, conf: str): - self.op_chk_support = self._parse_flag(conf) - - def parse_range_limit(self: any, conf: str): - self.op_range_limit = self._parse_str(conf) - - def parse_kern_name(self: any, conf: str): - self.kern_name = self._parse_str(conf) - - def parse_op_intf(self: any, conf: str): - self.op_intf = self._parse_str(conf) - - def parse_op_file(self: any, conf: str): - self.op_file = self._parse_str(conf) - - def parse_dynamic_shape(self: any, conf: str): - self.dynamic_shape = self._parse_flag(conf) - - def parse_attr_list(self: any, conf: str): - self.attr_list = self._parse_list(conf) - - def parse_attr_val(self: any, conf: str): - for attr in self.attr_list: - if self.attr_val.get(attr) is None: - self.attr_val[attr] = {} - if conf.startswith('attr_{}.type'.format(attr)): - self.attr_val.get(attr)['type'] = self._parse_str(conf) - elif conf.startswith('attr_{}.paramType'.format(attr)): - self.attr_val.get(attr)['paramType'] = self._parse_str(conf) - elif conf.startswith('attr_{}.defaultValue'.format(attr)): - self.attr_val.get(attr)['defaultValue'] = self._parse_str(conf) - - def parse_replay_val(self: any, batch_list: list, iterator_list: list): - if self.op_type in batch_list: - self.op_replay_flag = True - self.op_replay_batch = True - elif self.op_type in iterator_list: - self.op_replay_flag = True - self.op_replay_batch = False - - -def _is_op_type_in_opdesc(op_descs: list, op_type: str): - for op in op_descs: - if op_type == op.op_type: - return True - return False - - -def _set_all_options_to_opdescs(op_descs, soc_ver_compile_options): - for op in op_descs: - op.custom_all_compile_options = soc_ver_compile_options - - -def _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options): - for op in op_descs: - if op.op_type != op_type: - continue - op.custom_compile_options = soc_ver_compile_options - - -def _trans_soc_ver_to_short(soc_ver: str): - low_soc_ver = soc_ver.lower() - if low_soc_ver not in SOC_TO_SHORT_SOC_MAP: - print(f'WARNING: caution: {soc_ver} will trans into ascend910, if not your intention,' - f'use ascend910b1~4 instead') - return SOC_TO_SHORT_SOC_MAP[low_soc_ver] - - -def _get_op_custom_options(op_descs: list, auto_gen_dir: str): - if auto_gen_dir is None: - return {} - file = os.path.join(auto_gen_dir, "custom_compile_options.ini") - if not os.path.exists(file): - print(f'WARNING: cannot find {auto_gen_dir}/custom_compile_options.ini') - return {} - with open (file, 'r') as fd: - lines = fd.readlines() - for line in lines: - param_list = str.split(line.rstrip('\n'), ',') - if len(param_list) != 3: - raise Exception(f'ERROR: custom compile option {param_list} len is not 3') - op_type = param_list[0] - if op_type.upper() == 'ALL': - op_type = OP_ALL - if op_type != OP_ALL and _is_op_type_in_opdesc(op_descs, op_type) == False: - print(f'WARNING: op: {op_type} are not exists in this project') - continue - soc_ver_compile_options = {} - soc_ver = param_list[1] - options_str = param_list[2] - options = str.split(options_str, ';') - if soc_ver == '': - soc_ver_compile_options[SOC_ALL] = options - else: - soc_ver_list = str.split(soc_ver, ';') - for ver in soc_ver_list: - short_ver = _trans_soc_ver_to_short(ver) - soc_ver_compile_options[short_ver] = options - if op_type == OP_ALL: - _set_all_options_to_opdescs(op_descs, soc_ver_compile_options) - else: - _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options) - - -def get_op_desc(file: str, batch_list: list, iterator_list: list, builder: any, - op_type: list, auto_gen_dir: str = None) -> list: - op_descs = [] - op_match = False - with open (file, 'r') as fd: - lines = fd.readlines() - for line in lines: - line = line.strip() - if line.startswith('['): - name = line[1:-1] - if op_type is None or name in op_type: - op_match = True - op_desc = builder(name) - op_desc.parse_replay_val(batch_list, iterator_list) - op_descs.append(op_desc) - else: - op_match = False - if op_type is not None and len(op_descs) == len(op_type): - return op_descs - continue - if not op_match: - continue - if line.startswith('input'): - op_desc.parse_input(line) - elif line.startswith('output'): - op_desc.parse_output(line) - elif line.startswith('dynamicFormat.flag'): - op_desc.parse_op_format(line) - elif line.startswith('needCheckSupport.flag'): - op_desc.parse_check_support(line) - elif line.startswith('rangeLimit.value'): - op_desc.parse_range_limit(line) - elif line.startswith('opInterface.value'): - op_desc.parse_op_intf(line) - elif line.startswith('kernel.name'): - op_desc.parse_kern_name(line) - elif line.startswith('opFile.value'): - op_desc.parse_op_file(line) - elif line.startswith('dynamicShapeSupport.flag'): - op_desc.parse_dynamic_shape(line) - elif line.startswith('attr.list'): - op_desc.parse_attr_list(line) - elif line.startswith('attr_'): - op_desc.parse_attr_val(line) - _get_op_custom_options(op_descs, auto_gen_dir) - return op_descs diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py deleted file mode 100755 index 721465fee..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py +++ /dev/null @@ -1,338 +0,0 @@ -# Copyright 2020-2021 Huawei Technologies Co., Ltd -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -""" -parser ini to json -""" - -import json -import os -import stat -import sys - - -ATTR_TYPE_LIST = ["int", "float", "bool", "str", "listInt", "listFloat", "listBool", "listStr", "listListInt", - "type", "listType", "tensor", "listTensor"] -ATTR_PARAMTYPE_LIST = ["optional", "required"] -BOOL_FLAG_KEY = ["dynamicFormat", "dynamicShapeSupport", "dynamicRankSupport", "precision_reduce", "heavyOp", - "needCheckSupport"] -BOOL_LIST = ["true", "false"] -DTYPE_LIST = ["float16", "float", "float32", "int8", "int16", "int32", "uint8", "uint16", "uint32", "bool", - "int64", "uint64", "qint8", "qint16", "qint32", "quint8", "quint16", "double", "complex64", - "complex128", "string", "resource", "dual", "dual_sub_int8", "dual_sub_uint8", "string_ref", - "int4", "bfloat16", "uint1"] -FORMAT_LIST = ["NCHW", "NHWC", "ND", "NC1HWC0", "FRACTAL_Z", "NC1C0HWPAD", "NHWC1C0", "FSR_NCHW", "FRACTAL_DECONV", - "C1HWNC0", "FRACTAL_DECONV_TRANSPOSE", "FRACTAL_DECONV_SP_STRIDE_TRANS", "NC1HWC0_C04", - "FRACTAL_Z_C04", "CHWN", "FRACTAL_DECONV_SP_STRIDE8_TRANS", "HWCN", "NC1KHKWHWC0", "BN_WEIGHT", - "FILTER_HWCK", "HASHTABLE_LOOKUP_LOOKUPS", "HASHTABLE_LOOKUP_KEYS", "HASHTABLE_LOOKUP_VALUE", - "HASHTABLE_LOOKUP_OUTPUT", "HASHTABLE_LOOKUP_HITS", "C1HWNCoC0", "MD", "NDHWC", "FRACTAL_ZZ", - "FRACTAL_NZ", "NCDHW", "DHWCN", "NDC1HWC0", "FRACTAL_Z_3D", "CN", "NC", "DHWNC", - "FRACTAL_Z_3D_TRANSPOSE", "FRACTAL_ZN_LSTM", "FRACTAL_ZN_RNN", "FRACTAL_Z_G", "NULL"] - - -def parse_ini_files(ini_files): - """ - parse ini files to json - Parameters: - ---------------- - ini_files:input file list - return:ops_info - ---------------- - """ - tbe_ops_info = {} - for ini_file in ini_files: - check_file_size(ini_file) - parse_ini_to_obj(ini_file, tbe_ops_info) - return tbe_ops_info - - -def check_file_size(input_file): - try: - file_size = os.path.getsize(input_file) - except OSError as os_error: - print('[ERROR] Failed to open "%s". %s' % (input_file, str(os_error))) - raise OSError from os_error - if file_size > 10*1024*1024: - print('[WARN] The size of %s exceeds 10MB, it may take more time to run, please wait.' % input_file) - - -def parse_ini_to_obj(ini_file, tbe_ops_info): - """ - parse ini file to json obj - Parameters: - ---------------- - ini_file:ini file path - tbe_ops_info:ops_info - ---------------- - """ - with open(ini_file) as ini_file: - lines = ini_file.readlines() - op_dict = {} - op_name = "" - find_op_type = False - for line in lines: - line = line.rstrip() - if line == "": - continue - if line.startswith("["): - if line.endswith("]"): - op_name = line[1:-1] - op_dict = {} - tbe_ops_info[op_name] = op_dict - find_op_type = True - elif "=" in line: - key1 = line[:line.index("=")] - key2 = line[line.index("=")+1:] - key1_0, key1_1 = key1.split(".") - if key1_0 not in op_dict: - op_dict[key1_0] = {} - if key1_1 in op_dict.get(key1_0): - raise RuntimeError("Op:" + op_name + " " + key1_0 + " " + - key1_1 + " is repeated!") - dic_key = op_dict.get(key1_0) - dic_key[key1_1] = key2 - else: - continue - if not find_op_type: - raise RuntimeError("Not find OpType in .ini file.") - - -def check_output_exist(op_dict, is_valid): - """ - Function Description: - Check output is exist - Parameter: op_dict - Parameter: is_valid - """ - if "output0" in op_dict: - output0_dict = op_dict.get("output0") - if output0_dict.get("name", None) is None: - is_valid = False - print("output0.name is required in .ini file!") - else: - is_valid = False - print("output0 is required in .ini file!") - return is_valid - - -def check_attr_dict(attr_dict, is_valid, attr): - """ - Function Description: - Check attr_dict - Parameter: attr_dict - Parameter: is_valid - Parameter: attr - """ - attr_type = attr_dict.get("type") - value = attr_dict.get("value") - param_type = attr_dict.get("paramType") - if attr_type is None or value is None: - is_valid = False - print("If attr.list is exist, {0}.type and {0}.value is required".format(attr)) - if param_type and param_type not in ATTR_PARAMTYPE_LIST: - is_valid = False - print("{0}.paramType only support {1}.".format(attr, ATTR_PARAMTYPE_LIST)) - if attr_type and attr_type not in ATTR_TYPE_LIST: - is_valid = False - print("{0}.type only support {1}.".format(attr, ATTR_TYPE_LIST)) - return is_valid - - -def check_attr(op_dict, is_valid): - """ - Function Description: - Check attr - Parameter: op_dict - Parameter: is_valid - """ - if "attr" in op_dict: - attr_dict = op_dict.get("attr") - attr_list_str = attr_dict.get("list", None) - if attr_list_str is None: - is_valid = False - print("attr.list is required in .ini file!") - else: - attr_list = attr_list_str.split(",") - for attr_name in attr_list: - attr = "attr_" + attr_name.strip() - attr_dict = op_dict.get(attr) - if attr_dict: - is_valid = check_attr_dict(attr_dict, is_valid, attr) - else: - is_valid = False - print("%s is required in .ini file, when attr.list is %s!" % (attr, attr_list_str)) - return is_valid - - -def check_bool_flag(op_dict, is_valid): - """ - Function Description: - check_bool_flag - Parameter: op_dict - Parameter: is_valid - """ - for key in BOOL_FLAG_KEY: - if key in op_dict: - op_bool_key = op_dict.get(key) - if op_bool_key.get("flag").strip() not in BOOL_LIST: - is_valid = False - print("{0}.flag only support {1}.".format(key, BOOL_LIST)) - return is_valid - - -def check_type_format(op_info, is_valid, op_info_key): - """ - Function Description: - Check type and format - Parameter: op_info - Parameter: is_valid - Parameter: op_info_key - """ - op_info_dtype_str = op_info.get("dtype") - op_info_dtype_num = 0 - op_info_format_num = 0 - if op_info_dtype_str: - op_info_dtype = op_info_dtype_str.split(",") - op_info_dtype_num = len(op_info_dtype) - for dtype in op_info_dtype: - if dtype.strip() not in DTYPE_LIST: - is_valid = False - print("{0}.dtype not support {1}.".format(op_info_key, dtype)) - op_info_format_str = op_info.get("format") - if op_info_format_str: - op_info_format = op_info_format_str.split(",") - op_info_format_num = len(op_info_format) - for op_format in op_info_format: - if op_format.strip() not in FORMAT_LIST: - is_valid = False - print("{0}.format not support {1}.".format(op_info_key, op_format)) - if op_info_dtype_num > 0 and op_info_format_num > 0: - if op_info_dtype_num != op_info_format_num: - is_valid = False - print("The number of {0}.dtype not match the number of {0}.format.".format(op_info_key)) - return is_valid - - -def check_op_info(tbe_ops): - """ - Function Description: - Check info. - Parameter: tbe_ops - Return Value: is_valid - """ - print("\n\n==============check valid for ops info start==============") - required_op_input_info_keys = ["paramType", "name"] - required_op_output_info_keys = ["paramType", "name"] - param_type_valid_value = ["dynamic", "optional", "required"] - is_valid = True - for op_key in tbe_ops: - op_dict = tbe_ops[op_key] - for op_info_key in op_dict: - if op_info_key.startswith("input"): - op_input_info = op_dict[op_info_key] - missing_keys = [] - for required_op_input_info_key in required_op_input_info_keys: - if required_op_input_info_key not in op_input_info: - missing_keys.append(required_op_input_info_key) - if len(missing_keys) > 0: - print("op: " + op_key + " " + op_info_key + " missing: " + - ",".join(missing_keys)) - is_valid = False - else: - if not op_input_info["paramType"] in param_type_valid_value: - print("op: " + op_key + " " + op_info_key + \ - " paramType not valid, valid key:[dynamic, " - "optional, required]") - is_valid = False - is_valid = check_type_format(op_input_info, is_valid, op_info_key) - if op_info_key.startswith("output"): - op_input_info = op_dict[op_info_key] - missing_keys = [] - for required_op_input_info_key in required_op_output_info_keys: - if required_op_input_info_key not in op_input_info: - missing_keys.append(required_op_input_info_key) - if len(missing_keys) > 0: - print("op: " + op_key + " " + op_info_key + " missing: " + - ",".join(missing_keys)) - is_valid = False - else: - if not op_input_info["paramType"] in param_type_valid_value: - print("op: " + op_key + " " + op_info_key + - " paramType not valid, valid key:[dynamic, " - "optional, required]") - is_valid = False - is_valid = check_type_format(op_input_info, is_valid, op_info_key) - is_valid = check_attr(op_dict, is_valid) - is_valid = check_bool_flag(op_dict, is_valid) - print("==============check valid for ops info end================\n\n") - return is_valid - - -def write_json_file(tbe_ops_info, json_file_path): - """ - Save info to json file - Parameters: - ---------------- - tbe_ops_info: ops_info - json_file_path: json file path - ---------------- - """ - json_file_real_path = os.path.realpath(json_file_path) - wr_flag = os.O_WRONLY | os.O_CREAT - wr_mode = stat.S_IWUSR | stat.S_IRUSR - with os.fdopen(os.open(json_file_real_path, wr_flag, wr_mode), 'w') as file_path: - # Only the owner and group have rights - os.chmod(json_file_real_path, stat.S_IWGRP + stat.S_IWUSR + stat.S_IRGRP - + stat.S_IRUSR) - json.dump(tbe_ops_info, file_path, sort_keys=True, indent=4, - separators=(',', ':')) - print("Compile op info cfg successfully.") - - -def parse_ini_to_json(ini_file_paths, outfile_path): - """ - parse ini files to json file - Parameters: - ---------------- - ini_file_paths: list of ini file path - outfile_path: output file path - ---------------- - """ - tbe_ops_info = parse_ini_files(ini_file_paths) - if not check_op_info(tbe_ops_info): - print("Compile op info cfg failed.") - return False - write_json_file(tbe_ops_info, outfile_path) - return True - - -if __name__ == '__main__': - args = sys.argv - - OUTPUT_FILE_PATH = "tbe_ops_info.json" - ini_file_path_list = [] - - for arg in args: - if arg.endswith("ini"): - ini_file_path_list.append(arg) - OUTPUT_FILE_PATH = arg.replace(".ini", ".json") - if arg.endswith("json"): - OUTPUT_FILE_PATH = arg - - if len(ini_file_path_list) == 0: - ini_file_path_list.append("tbe_ops_info.ini") - - if not parse_ini_to_json(ini_file_path_list, OUTPUT_FILE_PATH): - sys.exit(1) - sys.exit(0) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/preset_parse.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/preset_parse.py deleted file mode 100755 index 8f1124b1d..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/preset_parse.py +++ /dev/null @@ -1,23 +0,0 @@ -import json -import sys -import os - - -def get_config_opts(file): - src_dir = os.path.abspath(os.path.dirname(file)) - opts = '' - with open(file, 'r') as fd: - config = json.load(fd) - for conf in config: - if conf == 'configurePresets': - for node in config[conf]: - macros = node.get('cacheVariables') - if macros is not None: - for key in macros: - opts += '-D{}={} '.format(key, macros[key]['value']) - opts = opts.replace('${sourceDir}', src_dir) - print(opts) - - -if __name__ == "__main__": - get_config_opts(sys.argv[1]) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py deleted file mode 100755 index 1baa364ef..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py +++ /dev/null @@ -1,105 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import os -import stat -import collections -import kernel_entry as keb -from tiling_data_def_build import gen_tiling -import code_channel_infer -import const_var - -PYF_PATH = os.path.dirname(__file__) - -ReplayCodeGenParams = collections.namedtuple('ReplayCodeGenParams',\ -['op_type', 'impl', 'tiling_file', 'kernel', 'entry', 'argn', 'op_replay_batch', 'max_block_dim', 'max_shape_size']) - - -class ReplayCodeGen: - def __init__(self, replayCodeGenParams): - self.op_type = replayCodeGenParams.op_type - self.impl = replayCodeGenParams.impl - self.tiling_file = replayCodeGenParams.tiling_file - self.tiling_data_file = '' - self.kernel = replayCodeGenParams.kernel - self.entry = replayCodeGenParams.entry - self.argn = replayCodeGenParams.argn - self.batch = False - self.outdir = '' - self.data_type = 'uint8_t' - self.blknum = 32 - self.op_replay_batch = replayCodeGenParams.op_replay_batch - self.max_block_dim = replayCodeGenParams.max_block_dim - self.max_shape_size = replayCodeGenParams.max_shape_size - - def set_batch(self, is_batch): - self.batch = is_batch - - def set_outdir(self, outdir): - self.outdir = outdir - - def gen_replay(self, ops_product: str): - kerentry = os.path.join(self.outdir, self.kernel + '_entry.cce') - kerimpl = os.path.join(self.outdir, self.kernel + '_impl.cpp') - replayimpl = os.path.join(self.outdir, self.kernel + '_replay.cpp') - if self.batch: - reptmp = os.path.join(PYF_PATH, 'batch_replay_impl.temp') - else: - reptmp = os.path.join(PYF_PATH, 'replay_impl.temp') - kertmp = os.path.join(PYF_PATH, 'kernel_impl.temp') - self._gen_kentry(kerentry) - self._gen_kimpl_code(kerimpl, kertmp) - self._gen_tiling_data_header() - self._gen_replay_code(replayimpl, reptmp, ops_product) - - def _gen_tiling_data_header(self): - self.tiling_data_file = os.path.join(self.outdir, self.kernel + '_tiling_data.h') - gen_tiling(self.tiling_file, self.tiling_data_file) - - def _gen_kimpl_code(self, src, tmpfile): - with open(tmpfile, 'r') as fd: - temp = fd.read() - temp = temp.replace('__CCE_FILE__', self.impl) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(temp) - - def _gen_replay_code(self, src, tmpfile, ops_product: str): - with open(tmpfile, 'r') as fd: - temp = fd.read() - temp = temp.replace('__ARG_NUM__', str(self.argn)) - argdef = [] - kargs = [] - for i in range(0, self.argn): - argdef.append('{} *'.format(self.data_type)) - kargs.append('({} *)GetArg({})'.format(self.data_type, i)) - temp = temp.replace('__ARGS_DEF__', ', '.join(argdef)) - temp = temp.replace('__KERNEL_ARGS__', ', '.join(kargs)) - temp = temp.replace('__KERNEL_FUN__', self.entry) - core_type_infer = 'core_type' - code_channel = code_channel_infer.infer_code_channel(code_channel_infer.InfoCodeChanelParams(self.impl,\ - self.tiling_data_file, self.kernel, self.outdir, ops_product, None)) - if code_channel == code_channel_infer.CODE_VEC: - core_type_infer = '0' - elif code_channel == code_channel_infer.CODE_CUBE: - core_type_infer = '1' - temp = temp.replace('__CORE_TYPE__', core_type_infer) - # regist function - temp = temp.replace('__OPS_PRODUCT__', ops_product) - temp = temp.replace('__OPTYPE__', self.op_type) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(temp) - - def _gen_kentry(self, src): - kf = '' - pre_alloc_str = 'A' * 256 - if self.batch: - kf += keb.batch_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str), self.argn, self.data_type) - else: - kf += keb.mc_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str),\ - self.argn, self.data_type, self.blknum) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(kf) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp deleted file mode 100755 index 1d30dd865..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp +++ /dev/null @@ -1,120 +0,0 @@ -#include -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#include "register/op_check.h" -#define __ASCENDC_REPLAY_CODE__ -using namespace std; -using namespace optiling; -using namespace AscendCReplay; - -extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); -extern "C" int elf_append(char *elf, uint32_t elfSize, char *jit, int kernum, int blknum[], char *atext[], - int alen[], int atlen, const char* kernelname[]); - -#define KERNEL_N 1 -#define ARG_N (__ARG_NUM__) -#define MAX_L (1024 * 1024 * 100) -#define MAX_E (1024 * 1024) - -int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) -{ - // gen type 1 : direct call codes 0: load .o file - if (param.gentype < 0 || param.gentype > 1) { - printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); - return 0; - } else if (param.gentype == 1 && param.objptr == nullptr) { - printf("Error: call replay with direct call mode, but code obj addr is null\n"); - return 0; - } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { - printf("Error: call replay with object file mode, but object file path is null\n"); - return 0; - } - // core_type 0:MIX 1:CUBE 2:VEC - if (core_type < 0 || core_type > 2) { - printf("Error: call replay core type is %d !\n", core_type); - return 0; - } - g_coreType = __CORE_TYPE__; - g_taskRation = param.task_ration; - g_tilingKey = param.tiling_key; - - unsigned char *buf, *jit; - char *kernel[KERNEL_N * 32]; - int len[KERNEL_N * 32]; - int blknum[KERNEL_N]; - int max; - block_num = param.block_dim; - g_ubBase = block_num; - uint8_t *code = (uint8_t *)malloc(MAX_L); - uint8_t *pos = code; - struct timespec tp1, tp2; - - clock_gettime(CLOCK_MONOTONIC, &tp1); - if (block_num > 32) { - printf("Error: block_num > 32\n"); - return 0; - } - //__OP_FOPEN__ - for (int i = 0; i < KERNEL_N; i++) { - for (int j = 0; j < ARG_N; j++) - AddArg(j, ARG_STEP * (j + 1)); - for (block_idx = 0; block_idx < block_num; block_idx++) { - //__OP_SET_KERNEL__ - int code_idx = i * block_num + block_idx; -#ifdef FP_CEILING - SetCtrlFloatEnable(); -#else - SetCtrlFloatDisable(); -#endif - CodeInit(pos, false); - __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); - CodeEnd(); - kernel[code_idx] = (char *)pos; - len[code_idx] = CodeLen(); - pos += len[code_idx]; - printf("kernel %d core %ld code generated len %d\n", i, block_idx, len[code_idx]); - } - blknum[i] = block_num; - } - //__OP_FCLOSE__ - clock_gettime(CLOCK_MONOTONIC, &tp2); - buf = (unsigned char *)malloc(MAX_E); - int fd = open(param.entry_file, O_RDONLY); - if (fd < 0) { - printf("[error]: cannot find entry.o : %s\n", param.entry_file); - return 0; - } - uint32_t bufSize = read(fd, buf, MAX_E); - if (bufSize <= 0) { - printf("[error]: entry.o : %s is too small ! \n", param.entry_file); - } - close(fd); - jit = (unsigned char *)malloc(MAX_L); - printf("total code generated %ld\n", pos - code); - int sz = elf_append((char *)buf, bufSize, (char *)jit, KERNEL_N, blknum, kernel, len, pos - code, ¶m.kernel_name); - if (tp1.tv_sec != tp2.tv_sec) { - printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); - } else { - printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); - } - printf("new elf size %d\n", sz); - if (param.gentype == 0) { - fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); - (void)write(fd, jit, sz); - close(fd); - free(jit); - } else if (param.gentype == 1) { - *param.objptr = (char*)jit; - } - free(buf); - free(code); - return sz; -} - -REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py deleted file mode 100755 index a96304261..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py +++ /dev/null @@ -1,84 +0,0 @@ -#!/usr/bin/env python -# coding=utf-8 -""" -Function: -The replay funtion entry -Copyright Information: -Huawei Technologies Co., Ltd. All Rights Reserved © 2020 -""" - -import sys -import os -import stat -import re -import const_var - - -def gen_tiling(tiling_header_file: str, tiling_file_out: str): - if not os.path.exists(tiling_header_file): - print("warning: no userdef tiling header file: ", tiling_header_file) - return - print("generate tiling def header file: ", tiling_file_out) - tmp_name = os.path.splitext(os.path.basename(tiling_header_file))[0].upper() - tiling_source = '#ifndef __{}_H__\n'.format(tmp_name) - tiling_source += '#define __{}_H__\n\n'.format(tmp_name) - tiling_source += '#include \n' - tiling_source += '#include \n\n' - tiling_source += '#include "kernel_tiling/kernel_tiling.h"\n\n' - end_source = "" - pattern = re.compile(r'[(](.*)[)]', re.S) - with open(tiling_header_file, 'r') as fd: - lines = fd.readlines() - for line in lines: - line = line.strip() - if (line.startswith('BEGIN_TILING_DATA_DEF')): - tiling_source += '#pragma pack(1)\n' - tiling_source += 'struct ' - struct_def = re.findall(pattern, line)[0] - tiling_source += struct_def + ' {\n' - elif (line.startswith('TILING_DATA_FIELD_DEF_ARR')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {}[{}] = {{}};\n'.format(fds[0].strip(), fds[2].strip(), fds[1].strip()) - elif (line.startswith('TILING_DATA_FIELD_DEF_STRUCT')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {};\n'.format(fds[0].strip(), fds[1].strip()) - elif (line.startswith('TILING_DATA_FIELD_DEF')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {} = 0;\n'.format(fds[0].strip(), fds[1].strip()) - elif (line.startswith('END_TILING_DATA_DEF')): - tiling_source += '};\n' - tiling_source += '#pragma pack()\n\n' - tiling_source += '#ifdef __NPU_TILING__\n' - tiling_source += \ - 'inline [aicore] void Init{stru}(const __gm__ uint8_t* tiling, {stru}* const_data)\n'\ - .format(stru=struct_def) - tiling_source += '{\n' - tiling_source += ' const __gm__ uint32_t *src = (const __gm__ uint32_t *)tiling;\n' - tiling_source += ' uint32_t *dst = (uint32_t *)const_data;\n' - tiling_source += ' for (auto i = 0; i < sizeof({}) / 4; i++) *(dst + i) = *(src + i);\n'\ - .format(struct_def) - tiling_source += '}\n' - tiling_source += '#else\n' - tiling_source += 'inline void Init{stru}(uint8_t* tiling, {stru}* const_data)\n'.format(stru=struct_def) - tiling_source += '{\n' - tiling_source += ' memcpy(const_data, tiling, sizeof({}));\n'.format(struct_def) - tiling_source += '}\n' - tiling_source += '#endif\n\n' - end_source = ''' -#define GET_TILING_DATA(tiling_data, tiling_arg) \\ -{stru} tiling_data; \\ -Init{stru}(tiling_arg, &tiling_data)\n -'''.format(stru=struct_def) - tiling_source += end_source - tiling_source += '#endif' - with os.fdopen(os.open(tiling_file_out, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(tiling_source) - - -if __name__ == '__main__': - if len(sys.argv) <= 2: - raise RuntimeError('arguments must greater than 2') - gen_tiling(sys.argv[1], sys.argv[2]) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/CMakeLists.txt deleted file mode 100644 index b6be9b492..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/CMakeLists.txt +++ /dev/null @@ -1,11 +0,0 @@ -if(NOT EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/mindspore") - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/caffe_plugin") - add_subdirectory(caffe_plugin) - endif() - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/tf_plugin") - add_subdirectory(tf_plugin) - endif() - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/onnx_plugin") - add_subdirectory(onnx_plugin) - endif() -endif() diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt deleted file mode 100644 index a6aba5c20..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt +++ /dev/null @@ -1,14 +0,0 @@ - -aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} plugin_srcs) -add_library(cust_tf_parsers SHARED ${plugin_srcs}) -target_compile_definitions(cust_tf_parsers PRIVATE google=ascend_private) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_tf_parsers PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_tf_parsers PRIVATE intf_pub graph) -install(TARGETS cust_tf_parsers - LIBRARY DESTINATION packages/vendors/${vendor_name}/framework/tensorflow -) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc deleted file mode 100644 index 2cd837ce5..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc +++ /dev/null @@ -1,23 +0,0 @@ -/* Copyright (C) 2020-2021. Huawei Technologies Co., Ltd. All -rights reserved. - * - * This program is free software; you can redistribute it and/or modify - * it under the terms of the Apache License Version 2.0. - * You may not use this file except in compliance with the License. - * - * This program is distributed in the hope that it will be useful, - * but WITHOUT ANY WARRANTY; without even the implied warranty of - * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the - * Apache License for more details at - * http://www.apache.org/licenses/LICENSE-2.0 - */ - -#include "register/register.h" - -namespace domi { -// register op info to GE -REGISTER_CUSTOM_OP("AddCustom") - .FrameworkType(TENSORFLOW) // type: CAFFE, TENSORFLOW - .OriginOpType("Add") // name in tf module - .ParseParamsByOperatorFn(AutoMappingByOpFn); -} // namespace domi diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/CMakeLists.txt deleted file mode 100644 index 40dd51cfa..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/CMakeLists.txt +++ /dev/null @@ -1,82 +0,0 @@ - -aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} ops_srcs) - -opbuild(OPS_SRC ${ops_srcs} - OUT_DIR ${ASCEND_AUTOGEN_PATH} -) - -add_library(cust_op_proto SHARED ${ops_srcs} ${ASCEND_AUTOGEN_PATH}/op_proto.cc) -target_compile_definitions(cust_op_proto PRIVATE OP_PROTO_LIB) -target_compile_options(cust_op_proto PRIVATE - -fvisibility=hidden -) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_op_proto PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_op_proto PRIVATE - intf_pub - exe_graph - register - tiling_api - -Wl,--whole-archive - rt2_registry - -Wl,--no-whole-archive -) -set_target_properties(cust_op_proto PROPERTIES OUTPUT_NAME - cust_opsproto_rt2.0 -) -add_library(cust_optiling SHARED ${ops_srcs}) -target_compile_definitions(cust_optiling PRIVATE OP_TILING_LIB) -target_compile_options(cust_optiling PRIVATE - -fvisibility=hidden -) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_optiling PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_optiling PRIVATE - intf_pub - exe_graph - register - tiling_api - -Wl,--whole-archive - rt2_registry - -Wl,--no-whole-archive -) -set_target_properties(cust_optiling PROPERTIES OUTPUT_NAME - cust_opmaster_rt2.0 -) - -file(GLOB aclnn_src ${ASCEND_AUTOGEN_PATH}/aclnn_*.cpp) -file(GLOB aclnn_inc ${ASCEND_AUTOGEN_PATH}/aclnn_*.h) -add_library(cust_opapi SHARED ${aclnn_src}) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_opapi PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_opapi PRIVATE intf_pub ascendcl nnopbase) - -add_custom_target(optiling_compat ALL - COMMAND ln -sf lib/linux/${CMAKE_SYSTEM_PROCESSOR}/$ - ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so -) - -install(TARGETS cust_op_proto - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_proto/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) -install(FILES ${ASCEND_AUTOGEN_PATH}/op_proto.h - DESTINATION packages/vendors/${vendor_name}/op_proto/inc) -install(TARGETS cust_optiling - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) -install(FILES ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so - DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling) -install(TARGETS cust_opapi - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_api/lib) -install(FILES ${aclnn_inc} - DESTINATION packages/vendors/${vendor_name}/op_api/include) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom.cpp b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom.cpp deleted file mode 100644 index b61a67f11..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom.cpp +++ /dev/null @@ -1,117 +0,0 @@ -#include "add_custom_tiling.h" -#include "register/op_def_registry.h" -#include "graph/utils/type_utils.h" -#include "tiling/platform/platform_ascendc.h" - -namespace optiling { -const uint32_t BLOCK_SIZE = 32; -const uint32_t BUFFER_NUM = 2; -static ge::graphStatus TilingFunc(gert::TilingContext* context) -{ - TilingData tiling; - uint64_t ubSize; - auto ascendcPlatform = platform_ascendc::PlatformAscendC(context->GetPlatformInfo()); - ascendcPlatform.GetCoreMemSize(platform_ascendc::CoreMemType::UB, ubSize); - auto coreNum = ascendcPlatform.GetCoreNum(); - - // Based on the input length and the number of inputs, the number of bytes of the input data type is obtained - uint32_t inputNum = context->GetInputShape(0)->GetStorageShape().GetShapeSize(); - uint32_t typeLength = 0; - ge::TypeUtils::GetDataTypeLength(context->GetInputDesc(0)->GetDataType(), typeLength); - uint32_t inputLength = inputNum * typeLength; - uint32_t inputBytes = inputLength / inputNum; - - // There are a total of 3 shared UB spaces in the input and output. If it's int8, there are 2 more TBUFs - uint32_t ubDataNumber = (inputBytes == 1) ? 5 : 3; - // The number of 32B data blocks that can be used for each data. DOUBLE BUFFER is already counted here - uint32_t tileBlockNum = (ubSize / BLOCK_SIZE / BUFFER_NUM) / ubDataNumber; - uint32_t tileDataNum = (tileBlockNum * BLOCK_SIZE) / inputBytes; - - // Input data for 32B alignment - uint32_t inputLengthAlgin32 = (((inputLength + BLOCK_SIZE - 1) / BLOCK_SIZE) * BLOCK_SIZE); - // There is at least 32B of data on each core, satisfying several settings for several cores. The maximum number of audits is the actual number of audits - coreNum = (coreNum < inputLengthAlgin32 / BLOCK_SIZE) ? coreNum : inputLengthAlgin32 / BLOCK_SIZE; - coreNum = (coreNum >= 1) ? coreNum : 1; - uint32_t everyCoreInputBlockNum = inputLengthAlgin32 / BLOCK_SIZE / coreNum; - uint32_t tailBlockNum = (inputLengthAlgin32 / BLOCK_SIZE) % coreNum; - - // Small chunks are calculated and sliced several times using the number of data on each core - uint32_t smallCoreDataNum = everyCoreInputBlockNum * BLOCK_SIZE / inputBytes; - uint32_t smallTileNum = everyCoreInputBlockNum / tileBlockNum; - uint32_t finalSmallTileNum = (everyCoreInputBlockNum % tileBlockNum) == 0 ? smallTileNum : smallTileNum + 1; - // Tail block calculation for small chunks of data - uint32_t smallTailDataNum = smallCoreDataNum - (tileDataNum * smallTileNum); - smallTailDataNum = smallTailDataNum == 0 ? tileDataNum : smallTailDataNum; - - // The total length of a large block of data is 32B larger than that of a small block of data - everyCoreInputBlockNum += 1; - uint32_t bigCoreDataNum = everyCoreInputBlockNum * BLOCK_SIZE / inputBytes; - uint32_t bigTileNum = everyCoreInputBlockNum / tileBlockNum; - uint32_t finalBigTileNum = (everyCoreInputBlockNum % tileBlockNum) == 0 ? bigTileNum : bigTileNum + 1; - uint32_t bigTailDataNum = bigCoreDataNum - tileDataNum * bigTileNum; - bigTailDataNum = bigTailDataNum == 0 ? tileDataNum : bigTailDataNum; - - tiling.set_smallCoreDataNum(smallCoreDataNum); - tiling.set_bigCoreDataNum(bigCoreDataNum); - tiling.set_tileDataNum(tileDataNum); - tiling.set_smallTailDataNum(smallTailDataNum); - tiling.set_bigTailDataNum(bigTailDataNum); - tiling.set_finalSmallTileNum(finalSmallTileNum); - tiling.set_finalBigTileNum(finalBigTileNum); - tiling.set_tailBlockNum(tailBlockNum); - - context->SetBlockDim(coreNum); - tiling.SaveToBuffer(context->GetRawTilingData()->GetData(), context->GetRawTilingData()->GetCapacity()); - context->GetRawTilingData()->SetDataSize(tiling.GetDataSize()); - size_t *currentWorkspace = context->GetWorkspaceSizes(1); - currentWorkspace[0] = 0; - return ge::GRAPH_SUCCESS; -} -} - -namespace ge { -static ge::graphStatus InferShape(gert::InferShapeContext* context) -{ - const gert::Shape* x1_shape = context->GetInputShape(0); - gert::Shape* y_shape = context->GetOutputShape(0); - *y_shape = *x1_shape; - return GRAPH_SUCCESS; -} -static graphStatus InferDataType(gert::InferDataTypeContext* context) -{ - const auto inputDataType = context->GetInputDataType(0); - context->SetOutputDataType(0, inputDataType); - return ge::GRAPH_SUCCESS; -} -} - -namespace ops { -class AddCustom : public OpDef { -public: - explicit AddCustom(const char* name) : OpDef(name) - { - this->Input("x") - .ParamType(REQUIRED) - .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) - .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) - .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); - this->Input("y") - .ParamType(REQUIRED) - .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) - .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) - .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); - this->Output("z") - .ParamType(REQUIRED) - .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) - .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) - .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); - - this->SetInferShape(ge::InferShape).SetInferDataType(ge::InferDataType); - this->AICore() - .SetTiling(optiling::TilingFunc) - .AddConfig("ascend310b") - .AddConfig("ascend910b"); - } -}; -OP_ADD(AddCustom); -} diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h deleted file mode 100644 index 28fd00a37..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h +++ /dev/null @@ -1,22 +0,0 @@ -/* - * Copyright (c) Huawei Technologies Co., Ltd. 2022-2023. All rights reserved. - */ -#ifndef ADD_CUSTOM_TILING_H -#define ADD_CUSTOM_TILING_H -#include "register/tilingdata_base.h" - -namespace optiling { -BEGIN_TILING_DATA_DEF(TilingData) - TILING_DATA_FIELD_DEF(uint32_t, smallCoreDataNum); - TILING_DATA_FIELD_DEF(uint32_t, bigCoreDataNum); - TILING_DATA_FIELD_DEF(uint32_t, finalBigTileNum); - TILING_DATA_FIELD_DEF(uint32_t, finalSmallTileNum); - TILING_DATA_FIELD_DEF(uint32_t, tileDataNum); - TILING_DATA_FIELD_DEF(uint32_t, smallTailDataNum); - TILING_DATA_FIELD_DEF(uint32_t, bigTailDataNum); - TILING_DATA_FIELD_DEF(uint32_t, tailBlockNum); -END_TILING_DATA_DEF; - -REGISTER_TILING_DATA_CLASS(AddCustom, TilingData) -} -#endif // ADD_CUSTOM_TILING_H \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt deleted file mode 100644 index 0d31a444c..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt +++ /dev/null @@ -1,61 +0,0 @@ -# set custom compile options -if ("${CMAKE_BUILD_TYPE}x" STREQUAL "Debugx") - add_ops_compile_options(ALL OPTIONS -g -O0) -endif() - -foreach(compute_unit ${ASCEND_COMPUTE_UNIT}) - - # generate aic-${compute_unit}-ops-info.json - add_ops_info_target(TARGET ops_info_gen_${compute_unit} - OUTPUT ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core/${compute_unit}/aic-${compute_unit}-ops-info.json - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/config/${compute_unit} - ) - - # generate ascendc impl py once - if (NOT TARGET ascendc_impl_gen) - add_ops_impl_target(TARGET ascendc_impl_gen - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl - ) - endif() - - # dynamic shape binary compile - if (${ENABLE_BINARY_PACKAGE}) - add_bin_compile_target(TARGET ascendc_bin_${compute_unit} - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} - ADP_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/dynamic - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/binary/${compute_unit} - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/kernel - COMPUTE_UNIT ${compute_unit} - ) - add_dependencies(ascendc_bin_${compute_unit} ascendc_impl_gen) - endif() - -endforeach() - -# generate npu_supported_ops.json -add_npu_support_target(TARGET npu_supported_ops - OPS_INFO_DIR ${ASCEND_AUTOGEN_PATH} - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core - INSTALL_DIR packages/vendors/${vendor_name}/framework/${ASCEND_FRAMEWORK_TYPE} -) - -if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) - add_subdirectory(testcases) -endif() - -# install kernel file -if (${ENABLE_SOURCE_PACKAGE}) - file(GLOB KERNEL_FILES - ${CMAKE_CURRENT_SOURCE_DIR}/*.cpp - ${CMAKE_CURRENT_SOURCE_DIR}/*.h - ${CMAKE_CURRENT_SOURCE_DIR}/*.py - ) - install(FILES ${KERNEL_FILES} - DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl/dynamic - ) -endif() diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp deleted file mode 100644 index 37c80a825..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp +++ /dev/null @@ -1,134 +0,0 @@ -/* - * Copyright (c) Huawei Technologies Co., Ltd. 2022-2023. All rights reserved. - * - * Function : z = x + y - * This sample is a very basic sample that implements vector add on Ascend plaform. - */ -#include "kernel_operator.h" -// tensor num for each queue -constexpr int32_t BUFFER_NUM = 2; - -template class KernelAdd { - using T = TYPE_X; -public: - __aicore__ inline KernelAdd() {} - __aicore__ inline void Init(GM_ADDR x, GM_ADDR y, GM_ADDR z, uint32_t smallCoreDataNum, - uint32_t bigCoreDataNum, uint32_t finalBigTileNum, - uint32_t finalSmallTileNum, uint32_t tileDataNum, - uint32_t smallTailDataNum, uint32_t bigTailDataNum, - uint32_t tailBlockNum) - { - ASSERT(AscendC::GetBlockNum() != 0 && "block dim can not be zero!"); - uint32_t coreNum = AscendC::GetBlockIdx(); - uint32_t globalBufferIndex = bigCoreDataNum * AscendC::GetBlockIdx(); - this->tileDataNum = tileDataNum; - if (coreNum < tailBlockNum) { - this->coreDataNum = bigCoreDataNum; - this->tileNum = finalBigTileNum; - this->tailDataNum = bigTailDataNum; - } - else { - this->coreDataNum = smallCoreDataNum; - this->tileNum = finalSmallTileNum; - this->tailDataNum = smallTailDataNum; - globalBufferIndex -= (bigCoreDataNum - smallCoreDataNum) * (AscendC::GetBlockIdx() - tailBlockNum); - } - xGm.SetGlobalBuffer((__gm__ TYPE_X*)x + globalBufferIndex, this->coreDataNum); - yGm.SetGlobalBuffer((__gm__ TYPE_Y*)y + globalBufferIndex, this->coreDataNum); - zGm.SetGlobalBuffer((__gm__ TYPE_Z*)z + globalBufferIndex, this->coreDataNum); - pipe.InitBuffer(inQueueX, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_X)); - pipe.InitBuffer(inQueueY, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_Y)); - pipe.InitBuffer(outQueueZ, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_Z)); - pipe.InitBuffer(tmp1, this->tileDataNum * sizeof(half)); - pipe.InitBuffer(tmp2, this->tileDataNum * sizeof(half)); - } - __aicore__ inline void Process() - { - int32_t loopCount = this->tileNum; - this->processDataNum = this->tileDataNum; - for (int32_t i = 0; i < loopCount; i++) { - if (i == this->tileNum - 1) { - this->processDataNum = this->tailDataNum; - } - CopyIn(i); - Compute(i); - CopyOut(i); - } - } - -private: - __aicore__ inline void CopyIn(int32_t progress) - { - AscendC::LocalTensor xLocal = inQueueX.AllocTensor(); - AscendC::LocalTensor yLocal = inQueueY.AllocTensor(); - AscendC::DataCopy(xLocal, xGm[progress * this->tileDataNum], this->processDataNum); - AscendC::DataCopy(yLocal, yGm[progress * this->tileDataNum], this->processDataNum); - inQueueX.EnQue(xLocal); - inQueueY.EnQue(yLocal); - } - __aicore__ inline void Compute(int32_t progress) - { - AscendC::LocalTensor xLocal = inQueueX.DeQue(); - AscendC::LocalTensor yLocal = inQueueY.DeQue(); - AscendC::LocalTensor zLocal = outQueueZ.AllocTensor(); - if constexpr (std::is_same_v) { - auto p1 = tmp1.Get(); - auto p2 = tmp2.Get(); - AscendC::Cast(p1, xLocal, AscendC::RoundMode::CAST_NONE, this->processDataNum); - AscendC::Cast(p2, yLocal, AscendC::RoundMode::CAST_NONE, this->processDataNum); - AscendC::Add(p2, p1, p2, this->processDataNum); - AscendC::Cast(p1.ReinterpretCast(), p2, AscendC::RoundMode::CAST_RINT, this->processDataNum); - AscendC::ShiftLeft(p1.ReinterpretCast(), p1.ReinterpretCast(), int16_t(8), this->processDataNum); - AscendC::ShiftRight(p1.ReinterpretCast(), p1.ReinterpretCast(), int16_t(8), this->processDataNum); - AscendC::Cast(p2, p1.ReinterpretCast(), AscendC::RoundMode::CAST_NONE, this->processDataNum); - AscendC::Cast(zLocal, p2, AscendC::RoundMode::CAST_NONE, this->processDataNum); - } - else { - AscendC::Add(zLocal, xLocal, yLocal, this->processDataNum); - } - outQueueZ.EnQue(zLocal); - inQueueX.FreeTensor(xLocal); - inQueueY.FreeTensor(yLocal); - } - __aicore__ inline void CopyOut(int32_t progress) - { - AscendC::LocalTensor zLocal = outQueueZ.DeQue(); - AscendC::DataCopy(zGm[progress * this->tileDataNum], zLocal, this->processDataNum); - outQueueZ.FreeTensor(zLocal); - } - -private: - AscendC::TPipe pipe; - AscendC::TQue inQueueX, inQueueY; - AscendC::TQue outQueueZ; - AscendC::TBuf tmp1, tmp2; - AscendC::GlobalTensor xGm; - AscendC::GlobalTensor yGm; - AscendC::GlobalTensor zGm; - uint32_t coreDataNum; - uint32_t tileNum; - uint32_t tileDataNum; - uint32_t tailDataNum; - uint32_t processDataNum; -}; - -extern "C" __global__ __aicore__ void add_custom(GM_ADDR x, GM_ADDR y, GM_ADDR z, GM_ADDR workspace, GM_ADDR tiling) -{ - GET_TILING_DATA(tiling_data, tiling); - KernelAdd op; - op.Init(x, y, z, tiling_data.smallCoreDataNum, - tiling_data.bigCoreDataNum, tiling_data.finalBigTileNum, - tiling_data.finalSmallTileNum, tiling_data.tileDataNum, - tiling_data.smallTailDataNum, tiling_data.bigTailDataNum, - tiling_data.tailBlockNum); - op.Process(); -} - -#ifndef ASCENDC_CPU_DEBUG -// call of kernel function -void add_custom_do(uint32_t blockDim, void* l2ctrl, void* stream, uint8_t* x, uint8_t* y, uint8_t* z, - uint8_t* workspace, uint8_t* tiling) -{ - add_custom<<>>(x, y, z, workspace, tiling); -} -#endif diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/install.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/install.sh deleted file mode 100644 index 9d45a8052..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/install.sh +++ /dev/null @@ -1,318 +0,0 @@ -#!/bin/bash -vendor_name=customize -targetdir=/usr/local/Ascend/opp -target_custom=0 - -sourcedir=$PWD/packages -vendordir=vendors/$vendor_name - -QUIET="y" - -while true -do - case $1 in - --quiet) - QUIET="y" - shift - ;; - --install-path=*) - INSTALL_PATH=$(echo $1 | cut -d"=" -f2-) - INSTALL_PATH=${INSTALL_PATH%*/} - shift - ;; - --*) - shift - ;; - *) - break - ;; - esac -done - -log() { - cur_date=`date +"%Y-%m-%d %H:%M:%S"` - echo "[runtime] [$cur_date] "$1 -} - -if [ -n "${INSTALL_PATH}" ]; then - if [[ ! "${INSTALL_PATH}" = /* ]]; then - log "[ERROR] use absolute path for --install-path argument" - exit 1 - fi - if [ ! -d ${INSTALL_PATH} ]; then - mkdir ${INSTALL_PATH} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${INSTALL_PATH} failed" - exit 1 - fi - fi - targetdir=${INSTALL_PATH} -elif [ -n "${ASCEND_CUSTOM_OPP_PATH}" ]; then - if [ ! -d ${ASCEND_CUSTOM_OPP_PATH} ]; then - mkdir -p ${ASCEND_CUSTOM_OPP_PATH} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${ASCEND_CUSTOM_OPP_PATH} failed" - fi - fi - targetdir=${ASCEND_CUSTOM_OPP_PATH} -else - if [ "x${ASCEND_OPP_PATH}" == "x" ]; then - log "[ERROR] env ASCEND_OPP_PATH no exist" - exit 1 - fi - targetdir="${ASCEND_OPP_PATH}" -fi - -if [ ! -d $targetdir ];then - log "[ERROR] $targetdir no exist" - exit 1 -fi - -upgrade() -{ - if [ ! -d ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 files" - return 0 - fi - - if [ ! -d ${targetdir}/$vendordir/$1 ];then - log "[INFO] create ${targetdir}/$vendordir/$1." - mkdir -p ${targetdir}/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/$1 failed" - return 1 - fi - else - has_same_file=-1 - for file_a in ${sourcedir}/$vendordir/$1/*; do - file_b=${file_a##*/}; - if [ "ls ${targetdir}/$vendordir/$1" = "" ]; then - log "[INFO] ${targetdir}/$vendordir/$1 is empty !!" - return 1 - fi - grep -q $file_b <<<`ls ${targetdir}/$vendordir/$1`; - if [[ $? -eq 0 ]]; then - echo -n "${file_b} " - has_same_file=0 - fi - done - if [ 0 -eq $has_same_file ]; then - if test $QUIET = "n"; then - echo "[INFO]: has old version in ${targetdir}/$vendordir/$1, \ - you want to Overlay Installation , please enter:[o]; \ - or replace directory installation , please enter: [r]; \ - or not install , please enter:[n]." - - while true - do - read orn - if [ "$orn" = n ]; then - return 0 - elif [ "$orn" = m ]; then - break; - elif [ "$0rn" = r ]; then - [ -n "${targetdir}/$vendordir/$1/" ] && rm -rf "${targetdir}/$vendordir/$1"/* - break; - else - echo "[ERROR] input error, please input again!" - fi - done - fi - fi - log "[INFO] replace or merge old ops $1 files .g....." - fi - - log "copy new ops $1 files ......" - if [ -d ${targetdir}/$vendordir/$1/ ]; then - chmod -R +w "$targetdir/$vendordir/$1/" >/dev/null 2>&1 - fi - cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 files failed" - return 1 - fi - - return 0 -} -upgrade_proto() -{ - if [ ! -f ${sourcedir}/$vendordir/custom.proto ]; then - log "[INFO] no need to upgrade custom.proto files" - return 0 - fi - if [ ! -d ${targetdir}/$vendordir/framework/caffe ];then - log "[INFO] create ${targetdir}/$vendordir/framework/caffe." - mkdir -p ${targetdir}/$vendordir/framework/caffe - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/framework/caffe failed" - return 1 - fi - else - if [ -f ${targetdir}/$vendordir/framework/caffe/custom.proto ]; then - # 有老版本,判断是否要覆盖式安装 - if test $QUIET = "n"; then - echo "[INFO] ${targetdir}/$vendordir/framework/caffe has old version"\ - "custom.proto file. Do you want to replace? [y/n] " - - while true - do - read yn - if [ "$yn" = n ]; then - return 0 - elif [ "$yn" = y ]; then - break; - else - echo "[ERROR] input error, please input again!" - fi - done - fi - fi - log "[INFO] replace old caffe.proto files ......" - fi - chmod -R +w "$targetdir/$vendordir/framework/caffe/" >/dev/null 2>&1 - cp -rf ${sourcedir}/$vendordir/custom.proto ${targetdir}/$vendordir/framework/caffe/ - if [ $? -ne 0 ];then - log "[ERROR] copy new custom.proto failed" - return 1 - fi - log "[INFO] copy custom.proto success" - - return 0 -} - -upgrade_file() -{ - if [ ! -e ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 file" - return 0 - fi - - log "copy new $1 files ......" - cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 file failed" - return 1 - fi - - return 0 -} - -delete_optiling_file() -{ - if [ ! -d ${targetdir}/vendors ];then - log "[INFO] $1 not exist, no need to uninstall" - return 0 - fi - sys_info=$(uname -m) - if [ ! -d ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/lib/linux/${sys_info} ];then - rm -rf ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/liboptiling.so - fi - return 0 -} - -log "[INFO] copy uninstall sh success" - -if [ ! -d ${targetdir}/vendors ];then - log "[INFO] create ${targetdir}/vendors." - mkdir -p ${targetdir}/vendors - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/vendors failed" - return 1 - fi -fi -chmod u+w ${targetdir}/vendors - -echo "[ops_custom]upgrade framework" -upgrade framework -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op proto" -upgrade op_proto -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade version.info" -upgrade_file version.info -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op impl" -delete_optiling_file op_impl -upgrade op_impl -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op api" -upgrade op_api -if [ $? -ne 0 ];then - exit 1 -fi - -upgrade_proto -if [ $? -ne 0 ];then - exit 1 -fi - -# set the set_env.bash -if [ -n "${INSTALL_PATH}" ] && [ -d ${INSTALL_PATH} ]; then - _ASCEND_CUSTOM_OPP_PATH=${targetdir}/${vendordir} - bin_path="${_ASCEND_CUSTOM_OPP_PATH}/bin" - set_env_variable="#!/bin/bash\nexport ASCEND_CUSTOM_OPP_PATH=${_ASCEND_CUSTOM_OPP_PATH}:\${ASCEND_CUSTOM_OPP_PATH}" - if [ ! -d ${bin_path} ]; then - mkdir -p ${bin_path} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${bin_path} failed" - exit 1 - fi - fi - echo -e ${set_env_variable} > ${bin_path}/set_env.bash - if [ $? -ne 0 ]; then - log "[ERROR] write ASCEND_CUSTOM_OPP_PATH to set_env.bash failed" - exit 1 - else - log "[INFO] using requirements: when custom module install finished or before you run the custom module, \ - execute the command [ source ${bin_path}/set_env.bash ] to set the environment path" - fi -else - config_file=${targetdir}/vendors/config.ini - if [ ! -f ${config_file} ]; then - touch ${config_file} - chmod 640 ${config_file} - echo "load_priority=$vendor_name" > ${config_file} - if [ $? -ne 0 ];then - echo "echo load_priority failed" - exit 1 - fi - else - found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" - found_vendor=$(echo $found_vendors | sed "s/\<$vendor_name\>//g" | tr ',' ' ') - vendor=$(echo $found_vendor | tr -s ' ' ',') - if [ "$vendor" != "" ]; then - sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" - fi - fi -fi - -chmod u-w ${targetdir}/vendors - -if [ -d ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/ ]; then - chmod -R 440 ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/* >/dev/null 2>&1 -fi -if [ -f ${targetdir}/ascend_install.info ]; then - chmod -R 440 ${targetdir}/ascend_install.info -fi -if [ -f ${targetdir}/scene.info ]; then - chmod -R 440 ${targetdir}/scene.info -fi -if [ -f ${targetdir}/version.info ]; then - chmod -R 440 ${targetdir}/version.info -fi - -echo "SUCCESS" -exit 0 - diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/upgrade.sh b/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/upgrade.sh deleted file mode 100644 index 34801efc4..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddAscendC/scripts/upgrade.sh +++ /dev/null @@ -1,151 +0,0 @@ -#!/bin/bash -vendor_name=customize -targetdir=/usr/local/Ascend/opp -target_custom=0 - -sourcedir=$PWD/packages -vendordir=vendors/$vendor_name - -log() { - cur_date=`date +"%Y-%m-%d %H:%M:%S"` - echo "[runtime] [$cur_date] "$1 -} - -if [[ "x${ASCEND_OPP_PATH}" == "x" ]];then - log "[ERROR] env ASCEND_OPP_PATH no exist" - exit 1 -fi - -targetdir=${ASCEND_OPP_PATH} - -if [ ! -d $targetdir ];then - log "[ERROR] $targetdir no exist" - exit 1 -fi - -upgrade() -{ - if [ ! -d ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 files" - return 0 - fi - - if [ ! -d ${targetdir}/$vendordir/$1 ];then - log "[INFO] create ${targetdir}/$vendordir/$1." - mkdir -p ${targetdir}/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/$1 failed" - return 1 - fi - else - vendor_installed_dir=$(ls "$targetdir/vendors" 2> /dev/null) - for i in $vendor_installed_dir;do - vendor_installed_file=$(ls "$vendor_installed_dir/$vendor_name/$i" 2> /dev/null) - if [ "$i" = "$vendor_name" ] && [ "$vendor_installed_file" != "" ]; then - echo "[INFO]: $vendor_name custom opp package has been installed on the path $vendor_installed_dir, \ - you want to Overlay Installation , please enter:[o]; \ - or replace directory installation , please enter: [r]; \ - or not install , please enter:[n]." - fi - while true - do - read mrn - if [ "$mrn" = m ]; then - break - elif [ "$mrn" = r ]; then - [ -n "$vendor_installed_file"] && rm -rf "$vendor_installed_file" - break - elif [ "$mrn" = n ]; then - return 0 - else - echo "[WARNING]: Input error, please input m or r or n to choose!" - fi - done - done - log "[INFO] replace old ops $1 files ......" - fi - - log "copy new ops $1 files ......" - cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 files failed" - return 1 - fi - - return 0 -} - -upgrade_file() -{ - if [ ! -e ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 file" - return 0 - fi - - log "copy new $1 files ......" - cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 file failed" - return 1 - fi - - return 0 -} - -log "[INFO] copy uninstall sh success" - -echo "[ops_custom]upgrade framework" -upgrade framework -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op proto" -upgrade op_proto -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op impl" -upgrade op_impl -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op api" -upgrade op_api -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade version.info" -upgrade_file version.info -if [ $? -ne 0 ];then - exit 1 -fi - -config_file=${targetdir}/vendors/config.ini -found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" -found_vendor=$(echo $found_vendors | sed "s/\<$vendor_name\>//g" | tr ',' ' ') -vendor=$(echo $found_vendor | tr -s ' ' ',') -if [ "$vendor" != "" ]; then - sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" -fi - -changemode() -{ - if [ -d ${targetdir} ];then - chmod -R 550 ${targetdir}>/dev/null 2>&1 - fi - - return 0 -} -echo "[ops_custom]changemode..." -#changemode -if [ $? -ne 0 ];then - exit 1 -fi - -echo "SUCCESS" -exit 0 - diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt deleted file mode 100644 index c89578fcd..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright 2024 Tencent Inc. All rights reserved. -# -# ============================================================================== -cmake_minimum_required(VERSION 3.14) -set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") -file(GLOB_RECURSE atb_plugin_operations_SRCS "./*.cpp") -set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") -message(STATUS "atb_plugin_operations_SRCS: ${atb_plugin_operations_SRCS}") -include_directories( - $ENV{ATB_HOME_PATH}/include - $ENV{ASCEND_HOME_PATH}/include - ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 - ${CMAKE_CURRENT_SOURCE_DIR} - ) -link_directories( - $ENV{ATB_HOME_PATH}/lib - $ENV{ASCEND_HOME_PATH}/lib64 - ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 - ) -add_library(atb_plugin_operations STATIC ${atb_plugin_operations_SRCS}) -target_link_libraries(atb_plugin_operations PRIVATE ${ACL_SHARED_LIBS} ${ACL_SHARED_LIBS} ascendcl) \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp b/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp deleted file mode 100644 index 468b84aff..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp +++ /dev/null @@ -1,142 +0,0 @@ -#include "aclnn_add_operation.h" -#include "aclnn_add_custom.h" - -AddOperation::AddOperation(const std::string &name, AddAttrParam param){ - attrParam = param; - opName_ = name; -} - -atb::SVector GetCopyTensorStride(atb::Dims &tensorDims) -{ - atb::SVector tmpStrides(tensorDims.dimNum, 1); - if (tensorDims.dimNum > 8) { // 8: tensor最大维度数量 - printf("tensor's dimNum is larger than 8, GetCopyTensorStride failed."); - return tmpStrides; - } - for (int64_t i = static_cast(tensorDims.dimNum) - 2; i >= 0; i--) { - tmpStrides[i] = (tensorDims.dims[i + 1] * tmpStrides[i + 1]); - } - return tmpStrides; -} - -std::shared_ptr AddOperation::CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx) -{ - auto aclnnTensor = std::make_shared(); - aclnnTensor->tensorIdx = static_cast(tensorIdx); - aclnnTensor->needUpdateTensorDataPtr = true; - aclnnTensor->atbTensor = atbTensor; - aclnnTensor->strides = GetCopyTensorStride(atbTensor.desc.shape); - - // 创建Aclnn tensor - aclnnTensor->tensor = aclCreateTensor(atbTensor.desc.shape.dims, - atbTensor.desc.shape.dimNum, - atbTensor.desc.dtype, - aclnnTensor->strides.data(), - 0, - atbTensor.desc.format, - atbTensor.desc.shape.dims, - atbTensor.desc.shape.dimNum, - atbTensor.deviceData); - return aclnnTensor; -} - -atb::Status AddOperation::UpdateAclnnVariantPack(const atb::VariantPack &variantPack) -{ - // 更新inTensor的device地址 - for (size_t i = 0; i < aclInTensors_.size(); ++i) { - int ret = -1; - if (!aclInTensors_[i]->needUpdateTensorDataPtr) { - continue; - } - aclInTensors_[i]->atbTensor = variantPack.inTensors.at(i); - ret = aclSetInputTensorAddr(aclExecutor_, - aclInTensors_[i]->tensorIdx, - aclInTensors_[i]->tensor, - aclInTensors_[i]->atbTensor.deviceData); - - if (ret != 0) { - printf("set input fail"); - return atb::ERROR_CANN_ERROR; - } - } - - // 更新outTensor的device地址 - for (size_t i = 0; i < aclOutTensors_.size(); ++i) { - int ret = -1; - if (!aclOutTensors_[i]->needUpdateTensorDataPtr) { - continue; - } - aclOutTensors_[i]->atbTensor = variantPack.outTensors.at(i); - ret = aclSetOutputTensorAddr(aclExecutor_, - aclOutTensors_[i]->tensorIdx, - aclOutTensors_[i]->tensor, - aclOutTensors_[i]->atbTensor.deviceData); - - if (ret != 0) { - printf("set output fail"); - return atb::ERROR_CANN_ERROR; - } - } - return atb::NO_ERROR; -} - -atb::Status AddOperation::Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) { - - aclInTensors_.resize(GetInputNum()); - for (size_t i = 0; i < aclInTensors_.size(); ++i) { - auto aclnnTensor = CreateAclnnTensor(variantPack.inTensors.at(i), i); - if (aclnnTensor->tensor == nullptr) { - printf("creat input tensor %ld fail", i); - return atb::ERROR_INTERNAL_ERROR; - } - aclInTensors_[i] = aclnnTensor; - } - - aclOutTensors_.resize(GetOutputNum()); - for (size_t i = 0; i < aclOutTensors_.size(); ++i) { - auto aclnnTensor = CreateAclnnTensor(variantPack.outTensors.at(i), i); - if (aclnnTensor->tensor == nullptr) { - printf("creat output tensor %ld fail", i); - return atb::ERROR_INTERNAL_ERROR; - } - aclOutTensors_[i] = aclnnTensor; - } - - - auto ret = aclnnAddCustomGetWorkspaceSize(aclInTensors_.at(0)->tensor, - aclInTensors_.at(1)->tensor, - aclOutTensors_.at(0)->tensor, - &workspaceSize_, - &aclExecutor_); - - workspaceSize = workspaceSize_; - return ret; - -} - -atb::Status AddOperation::Execute(const atb::VariantPack &variantPack, uint8_t *workspace, uint64_t workspaceSize, atb::Context *context) { - - - - aclrtStream stream = context->GetExecuteStream(); - if (!stream) { - printf("get stream fail"); - return atb::ERROR_INVALID_PARAM; - } - // 更新数据传入的地址 - int ret = UpdateAclnnVariantPack(variantPack); - if (ret != 0) { - printf("UpdateAclnnVariantPack fail"); - return atb::ERROR_CANN_ERROR; - } - ret = aclnnAddCustom(workspace, workspaceSize_, aclExecutor_, stream); - - return ret; -} - -atb::Status AddOperation::InferShape( - const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const -{ - outTensorDesc.at(0) = inTensorDesc.at(0); - return atb::NO_ERROR; -} diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h b/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h deleted file mode 100644 index 8670088d3..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h +++ /dev/null @@ -1,56 +0,0 @@ -#include -#include -#include -#include -#include -#include "atb/infer_op_params.h" - - -struct AddAttrParam -{ - // add没属性,此处空 -}; - -struct AclnnTensor -{ -public: - atb::Tensor atbTensor; // - aclTensor *tensor = nullptr; - int tensorIdx = -1; // aclTensor在aclExecutor中的index - bool needUpdateTensorDataPtr = false; - atb::SVector strides = {}; -}; - -class AddOperation: public atb::Operation{ -public: - AddOperation(const std::string &name, AddAttrParam param); - atb::Status Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) override; - atb::Status Execute(const atb::VariantPack &variantPack, uint8_t *workspace, - uint64_t workspaceSize, atb::Context *context) override; - atb::Status InferShape( - const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const; - std::shared_ptr CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx); - atb::Status UpdateAclnnVariantPack(const atb::VariantPack &variantPack); - - uint32_t GetInputNum() const - { - return 2; // 算子入参个数 - } - - uint32_t GetOutputNum() const - { - return 1; // 算子出参个数 - } - std::string GetName() const - { - return opName_; - } - - aclOpExecutor *aclExecutor_ = nullptr; - AddAttrParam attrParam; - std::string opName_; - uint64_t workspaceSize_; - - atb::SVector> aclInTensors_; - atb::SVector> aclOutTensors_; -}; \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/build.sh b/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/build.sh deleted file mode 100755 index ba28cc143..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddOperationATBPlugin/build.sh +++ /dev/null @@ -1,33 +0,0 @@ -#!/bin/bash - -# 定义构建目录 -BUILD_DIR="build" - -# 创建构建目录 -mkdir -p "$BUILD_DIR" -cd "$BUILD_DIR" - -# 运行 CMake 配置和编译 -cmake .. -make - -# 查找生成的 .a 文件 -A_FILE=$(find . -name "*.a" -type f) - -# 检查是否找到了 .a 文件 -if [ -z "$A_FILE" ]; then - echo "未找到 .a 文件,编译可能失败。" - exit 1 -fi - -# 复制头文件到 /usr/include -HEADER_FILES=$(find .. -name "*.h" -type f) -for header in $HEADER_FILES; do - cp "$header" /usr/include/ -done - -# 复制 .a 文件到 /usr/local/lib -cp "$A_FILE" /usr/local/lib/ - -echo "构建完成,头文件和 .a 文件已复制到目标目录。" - diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/CMakeLists.txt b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/CMakeLists.txt deleted file mode 100644 index b9c2cec46..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/CMakeLists.txt +++ /dev/null @@ -1,40 +0,0 @@ -cmake_minimum_required(VERSION 3.16) -project("test_model") -option(USE_CXX11_ABI "USE_CXX11_ABI" OFF) -set(CMAKE_BUILD_TYPE Debug) -if(USE_CXX11_ABI) - set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=1") -else() - set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") -endif() - -set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -std=c++17") - -set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") - -include_directories( - $ENV{ATB_HOME_PATH}/include - $ENV{ASCEND_HOME_PATH}/include - ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 - ${CMAKE_CURRENT_SOURCE_DIR} - ./src/ - ) - -message(".h path = ${CUST_PKG_PATH}/include") -link_directories( - $ENV{ATB_HOME_PATH}/lib - $ENV{ASCEND_HOME_PATH}/lib64 - ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 - ) - - -link_libraries(atb_plugin_operations) -add_executable(test_model ./src/main.cpp) - -# 列出所有的头文件目录 -target_link_libraries(test_model PRIVATE atb ascendcl cust_opapi opapi nnopbase pthread) # 添加自定义算子so及适配so - - - - - diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/gen_data.py b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/gen_data.py deleted file mode 100644 index 83ecafbce..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/gen_data.py +++ /dev/null @@ -1,20 +0,0 @@ -import torch -import os -def gen_golden_data_simple(): - dtype = "float" - input1 = torch.randn(133, 4095, dtype=torch.float) - input2 = torch.randn(133, 4095, dtype=torch.float) - - - golden = input1 + input2 - os.system("mkdir -p input") - os.system("mkdir -p output") - input1.numpy().tofile('./script/input/input0.bin') - input2.numpy().tofile('./script/input/input1.bin') - golden.numpy().tofile("./script/output/golden0.bin") - - with open("./output/meta", "w") as fp: - print(dtype, file=fp) - -if __name__ == "__main__": - gen_golden_data_simple() diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/run.sh b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/run.sh deleted file mode 100644 index 450b717f2..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/run.sh +++ /dev/null @@ -1,52 +0,0 @@ -#!/bin/bash - - -# 步骤1: 运行gen_data.py生成输入bin文件和golden标杆输出数据 -echo "正在生成输入数据和golden标杆数据..." -mkdir -p script/input -mkdir -p script/output -python3 script/gen_data.py -if [ $? -ne 0 ]; then - echo "生成数据失败,脚本终止。" - exit 1 -fi - -# 步骤2: 创建构建目录并进入 -mkdir -p build -cd build -if [ $? -ne 0 ]; then - echo "无法进入构建目录,脚本终止。" - exit 1 -fi - -# 步骤3: 使用CMake配置项目 -echo "正在配置CMake项目..." -cmake .. -if [ $? -ne 0 ]; then - echo "CMake配置失败,脚本终止。" - exit 1 -fi - -# 步骤4: 编译代码 -echo "正在编译代码..." -make -if [ $? -ne 0 ]; then - echo "编译失败,脚本终止。" - exit 1 -fi - -mv test_model ../ -cd .. - -# 步骤5: 运行可执行文件生成实际输出文件 -echo "正在运行可执行文件生成实际输出..." -./test_model -if [ $? -ne 0 ]; then - echo "运行可执行文件失败,脚本终止。" - exit 1 -fi - -# 步骤6: 调用verify_result.py进行golden标杆数据和实际输出数据的比对 -echo "正在验证结果..." -python3 script/verify_result.py script/output/output_0.bin script/output/golden0.bin - diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/verify_result.py b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/verify_result.py deleted file mode 100644 index 629845e13..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/script/verify_result.py +++ /dev/null @@ -1,28 +0,0 @@ -import os -import sys -import numpy - -loss = 1e-3 # 容忍偏差,一般fp16要求绝对误差和相对误差均不超过千分之一 -minimum = 10e-10 - -def verify_result(real_result, golden): - with open("output/meta", "r") as fp: - dtype_str = fp.readline().strip() - dtype = eval(dtype_str) - real_result = numpy.fromfile(real_result, dtype=dtype) # 从bin文件读取实际运算结果 - golden = numpy.fromfile(golden, dtype=dtype) # 从bin文件读取预期运算结果 - print("=" * 50, real_result[:5], golden[:5], "=" * 50, sep='\n', end='\n', file=sys.stderr) - result = numpy.abs(real_result - golden) # 计算运算结果和预期结果偏差 - deno = numpy.maximum(numpy.abs(real_result), numpy.abs(golden)) # 获取最大值并组成新数组 - result_atol = numpy.less_equal(result, loss) # 计算绝对误差 - result_rtol = numpy.less_equal(result / numpy.add(deno, minimum), loss) # 计算相对误差 - if not result_rtol.all() and not result_atol.all(): - if numpy.sum(result_rtol == False) > real_result.size * loss and numpy.sum(result_atol == False) > real_result.size * loss: # 误差超出预期时返回打印错误,返回对比失败 - print("[ERROR] result error") - return False - print("test Operation success!") - return True - - -if __name__ == '__main__': - verify_result(sys.argv[1],sys.argv[2]) diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.cpp b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.cpp deleted file mode 100644 index 258c0cb3c..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.cpp +++ /dev/null @@ -1,217 +0,0 @@ -/* - * Copyright (c) Huawei Technologies Co., Ltd. 2023. All rights reserved. - * - * Licensed under the Apache License, Version 2.0 (the "License"); - * you may not use this file except in compliance with the License. - * You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ -#include "main.h" - -bool SetInputData(std::vector &inputData){ - char *xPath = "./script/input/input0.bin"; - char *yPath = "./script/input/input1.bin"; - InputData inputX; - InputData inputY; - inputX.data = ReadBinFile(xPath,inputX.size); - inputY.data = ReadBinFile(yPath,inputY.size); - inputData.push_back(inputX); - inputData.push_back(inputY); - return true; -} - -bool SetOperationInputDesc(atb::SVector &intensorDescs){ - atb::TensorDesc xDesc; - xDesc.dtype = ACL_FLOAT; - xDesc.format = ACL_FORMAT_ND; - xDesc.shape.dimNum = 2; - xDesc.shape.dims[0] = 133; - xDesc.shape.dims[1] = 4095; - - atb::TensorDesc yDesc; - yDesc.dtype = ACL_FLOAT; - yDesc.format = ACL_FORMAT_ND; - yDesc.shape.dimNum = 2; - yDesc.shape.dims[0] = 133; - yDesc.shape.dims[1] = 4095; - - intensorDescs.at(0) = xDesc; - intensorDescs.at(1) = yDesc; -} - - - -static void SetCurrentDevice() -{ - const int deviceId = 0; - std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; - int ret = aclrtSetDevice(deviceId); - if (ret != 0) { - std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; - return; - } - std::cout << "[INFO]: aclrtSetDevice success" << std::endl; -} - - -static void FreeTensor(atb::Tensor &tensor) -{ - if (tensor.deviceData) { - int ret = aclrtFree(tensor.deviceData); - if (ret != 0) { - std::cout << "[ERROR]: aclrtFree fail" << std::endl; - } - tensor.deviceData = nullptr; - tensor.dataSize = 0; - } - if (tensor.hostData) { - int ret = aclrtFreeHost(tensor.hostData); - if (ret != 0) { - std::cout << "[ERROR]: aclrtFreeHost fail, ret = " << ret << std::endl; - } - tensor.hostData = nullptr; - tensor.dataSize = 0; - } -} - -static void FreeTensors(atb::SVector &inTensors, atb::SVector &outTensors) -{ - for (size_t i = 0; i < inTensors.size(); ++i) { - FreeTensor(inTensors.at(i)); - } - for (size_t i = 0; i < outTensors.size(); ++i) { - FreeTensor(outTensors.at(i)); - } -} -bool SaveMemoryToBinFile(void* memoryAddress, size_t memorySize, size_t i) { - // 创建 output 目录(如果不存在) - std::filesystem::create_directories("output"); - - // 生成文件名 - std::string filename = "script/output/output_" + std::to_string(i) + ".bin"; - - // 打开文件以二进制写入模式 - std::ofstream file(filename, std::ios::binary); - if (!file) { - std::cerr << "无法打开文件: " << filename << std::endl; - return false; - } - - // 写入数据 - file.write(static_cast(memoryAddress), memorySize); - if (!file) { - std::cerr << "写入文件时出错: " << filename << std::endl; - file.close(); - return false; - } - - // 关闭文件 - file.close(); - std::cout << "数据已成功保存到: " << filename << std::endl; - return true; -} - -int main(int argc, const char *argv[]) -{ - const int deviceId = 0; - std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; - int ret = aclrtSetDevice(deviceId); - if (ret != 0) { - std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; - return 1; - } - std::cout << "[INFO]: aclrtSetDevice success" << std::endl; - atb::Context *context = nullptr; - ret = atb::CreateContext(&context); - void *stream = nullptr; - ret = aclrtCreateStream(&stream); - if (ret != 0) { - std::cout << "[ERROR]: AsdRtStreamCreate fail, ret:" << ret << std::endl; - return 1; - } - context->SetExecuteStream(stream); - - std::vector input; - SetInputData(input); - - AddAttrParam addAttrParam; - AddOperation *op = new AddOperation("Add",addAttrParam); - std::cout << "[INFO]: complete CreateOp!" << std::endl; - - if(input.size() != op->GetInputNum()) std::cout << "[ERROR]: Operation actual input num is not equal to GetInputNum()"; - - atb::SVector intensorDescs; - atb::SVector outtensorDescs; - intensorDescs.resize(op->GetInputNum()); - outtensorDescs.resize(op->GetOutputNum()); - SetOperationInputDesc(intensorDescs); - atb::Status st = op->InferShape(intensorDescs,outtensorDescs); - if (st != 0) { - std::cout << "[ERROR]: Operation InferShape fail" << std::endl; - return -1; - } - std::cout << "[INFO]: Operation InferShape success" << std::endl; - atb::VariantPack variantPack; - variantPack.inTensors.resize(op->GetInputNum()); - variantPack.outTensors.resize(op->GetOutputNum()); - for(size_t i=0;iGetInputNum();i++){ - variantPack.inTensors.at(i).desc = intensorDescs.at(i); - variantPack.inTensors.at(i).hostData = input[i].data; - variantPack.inTensors.at(i).dataSize = input[i].size; - CheckAcl(aclrtMalloc(&variantPack.inTensors.at(i).deviceData, input[i].size, ACL_MEM_MALLOC_HUGE_FIRST)); - CheckAcl(aclrtMemcpy(variantPack.inTensors.at(i).deviceData, input[i].size, input[i].data, input[i].size, ACL_MEMCPY_HOST_TO_DEVICE)); - } - std::cout << "[INFO]: Operation Input prepare sucess" << std::endl; - for(size_t i=0;iGetOutputNum();i++){ - int64_t *dims = new int64_t[outtensorDescs.at(i).shape.dimNum]; - for(size_t j=0;jSetup(variantPack, workspaceSize, context); - if (st != 0) { - std::cout << "[ERROR]: Operation setup fail" << std::endl; - return -1; - } - std::cout << "[INFO]: Operation setup success" << std::endl; - void *workspace = nullptr; - if (workspaceSize > 0) { - aclrtMalloc(&workspace, workspaceSize, ACL_MEM_MALLOC_HUGE_FIRST); - } - - std::cout << "[INFO]: Operation execute start" << std::endl; - st = op->Execute(variantPack, (uint8_t*)workspace, workspaceSize, context); - if (st != 0) { - std::cout << "[ERROR]: Operation execute fail" << std::endl; - return -1; - } - std::cout << "[INFO]: Operation execute success" << std::endl; - for(size_t i = 0; i < op->GetOutputNum(); i++){ - CheckAcl(aclrtMemcpy(variantPack.outTensors.at(i).hostData, variantPack.outTensors.at(i).dataSize, variantPack.outTensors.at(0).deviceData, - variantPack.outTensors.at(i).dataSize, ACL_MEMCPY_DEVICE_TO_HOST)); - SaveMemoryToBinFile(variantPack.outTensors.at(i).hostData,variantPack.outTensors.at(i).dataSize,i); - } - - FreeTensors(variantPack.inTensors, variantPack.outTensors); - st = atb::DestroyContext(context); - CheckAcl(aclrtDestroyStream(stream)); - CheckAcl(aclrtResetDevice(0)); - CheckAcl(aclFinalize()); - return atb::ErrorType::NO_ERROR; -} diff --git a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.h b/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.h deleted file mode 100644 index 44a94e9ec..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/AddOperationTest/src/main.h +++ /dev/null @@ -1,55 +0,0 @@ -#include -#include -#include -#include -#include -#include -#include -#include "securec.h" -#include "atb/atb_infer.h" -#include "aclnn_add_operation.h" -#include - -struct InputData{ - void* data; - uint64_t size; -}; -aclError CheckAcl(aclError ret) -{ - if (ret != ACL_ERROR_NONE) { - std::cerr << __FILE__ << ":" << __LINE__ << " aclError:" << ret << std::endl; - } - return ret; -} -void* ReadBinFile(const char* filename, size_t& size) { - std::ifstream file(filename, std::ios::binary | std::ios::ate); - if (!file) { - std::cerr << "无法打开文件: " << filename << std::endl; - return nullptr; - } - - // 获取文件大小 - size = file.tellg(); - file.seekg(0, std::ios::beg); - - // 分配内存 - void* buffer; - int ret = aclrtMallocHost(&buffer,size); - if (!buffer) { - std::cerr << "内存分配失败" << std::endl; - file.close(); - return nullptr; - } - - // 读取文件内容到内存 - file.read(static_cast(buffer), size); - if (!file) { - std::cerr << "读取文件失败" << std::endl; - delete[] static_cast(buffer); - file.close(); - return nullptr; - } - - file.close(); - return buffer; -} \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AtbAddSample/readme.md b/atb_operator/AtbAddSample/AtbAddSample/readme.md deleted file mode 100644 index d205e5429..000000000 --- a/atb_operator/AtbAddSample/AtbAddSample/readme.md +++ /dev/null @@ -1,81 +0,0 @@ -### 概述 - -本样例基于AscendC自定义Add算子,开发了ATB插件并进行了插件调用测试. - -### 项目结构介绍 - -├── AddAscendC //Add AscendC自定义算子工程 - -├── AddOperationATBPlugin //AddOperation ATB插件代码 - -├── AddOperationTest //AddOperation 测试代码 - -### 样例运行 - -#### Add AscendC自定义算子部署 - -- 运行生成算子run包 - - ```shell - cd AddAscendC - bash build.sh - ``` - -- 算子run包部署 - - ```shell - cd AddAscendC/build_out - ./custom_opp_hce_aarch64.run - ``` - -#### AddOperation ATB插件部署 - -- 运行编译脚本完成部署(脚本会生成静态库.a文件,同时将头文件拷贝到/usr/include,.a文件拷贝到/usr/local/lib下) - - ``` - cd AddOperationATBPlugin - bash build.sh - ``` - -#### AddOperation测试 - -- 运行脚本完成算子测试 - - ```shell - cd AddOperationTest - bash run.sh - ``` - -### AddOperation算子介绍 - -#### 功能 - -实现两个输入张量相加 - -#### 定义 - -``` -struct AddParam { -}; -``` - -#### 参数列表 - -该算子参数为空 - -#### 输入 - -| **参数** | **维度** | **数据类型** | **格式** | 描述 | -| -------- | -------------------------- | --------------------- | -------- | ---------- | -| x | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | -| y | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | - -#### 输出 - -| **参数** | **维度** | **数据类型** | **格式** | 描述 | -| -------- | -------------------------- | --------------------- | -------- | ---------------------------------------- | -| output | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输出tensor。数据类型和shape与x保持一致。 | - -#### 规格约束 - -暂无 \ No newline at end of file -- Gitee From 74f9e21243f2c329caffd3ebdf0ff3b42fe20948 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Mon, 31 Mar 2025 06:28:41 +0000 Subject: [PATCH 06/26] update atb_operator/AtbAddSample/AddAscendC/README.md. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: 万祖涛 --- .../AtbAddSample/AddAscendC/README.md | 45 +++---------------- 1 file changed, 6 insertions(+), 39 deletions(-) diff --git a/atb_operator/AtbAddSample/AddAscendC/README.md b/atb_operator/AtbAddSample/AddAscendC/README.md index da59a83e7..ea2794454 100644 --- a/atb_operator/AtbAddSample/AddAscendC/README.md +++ b/atb_operator/AtbAddSample/AddAscendC/README.md @@ -1,25 +1,16 @@ -<<<<<<< HEAD -## 概述 -本样例基于AddCustom算子工程,介绍了单算子工程及单算子调用。 -## 算子描述 -======= # 概述 本样例基于AddCustom算子工程,介绍了单算子工程及aclnn接口说明。 ## 算子工程介绍 ### 算子描述 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd Add算子实现了两个数据相加,返回相加结果的功能,该算子实现了非32B对齐shape下的算子功能,完成了算子的泛化实现。 对应的数学表达式为: ``` z = x + y ``` -<<<<<<< HEAD -## 算子规格描述 -======= + ### 算子规格描述 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd @@ -34,11 +25,8 @@ z = x + y
算子类型(OpType)Add
-<<<<<<< HEAD -## 算子工程介绍 -======= + ### 算子工程文件介绍 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd 其中,算子工程目录AddCustom包含算子实现的模板文件、编译脚本等,如下所示: ``` ├── AddCustom //Add自定义算子工程 @@ -51,28 +39,19 @@ z = x + y │ ├── CMakeLists.txt // 算子工程的CMakeLists.txt │ └── CMakePresets.json // 编译配置项 ``` -<<<<<<< HEAD CANN软件包中提供了工程创建工具msopgen,AddCustom算子工程可通过AddCustom.json自动创建,具体请参考[Ascend C算子开发](https://hiascend.com/document/redirect/CannCommunityOpdevAscendC)>算子开发>算子开发工程>基于自定义算子工程的算子开发>创建算子工程 章节。 -## 编译运行样例算子 -======= ### 编译运行样例算子 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd 针对自定义算子工程,编译运行包含如下步骤: - 编译自定义算子工程生成算子安装包; - 安装自定义算子到算子库中; - 调用执行自定义算子; 详细操作如下所示。 -<<<<<<< HEAD -### 1. 获取源码包 -编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 -### 2. 编译算子工程 -======= + #### 1. 获取源码包 编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 #### 2. 编译算子工程 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd 编译自定义算子工程,构建生成自定义算子包。 - 执行如下命令,切换到算子工程AddCustom目录。 @@ -108,11 +87,8 @@ CANN软件包中提供了工程创建工具msopgen,AddCustom算子工程可通 备注:如果要使用dump调试功能,需要移除op_host内和CMakeLists.txt内的Atlas 200/500 A2 推理产品的配置。 -<<<<<<< HEAD -### 3. 部署算子包 -======= + #### 3. 部署算子包 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd 执行如下命令,在自定义算子安装包所在路径下,安装自定义算子包。 ```bash @@ -121,11 +97,7 @@ CANN软件包中提供了工程创建工具msopgen,AddCustom算子工程可通 ``` 命令执行成功后,自定义算子包中的相关文件将部署至当前环境的OPP算子库的vendors/customize目录中。 -<<<<<<< HEAD -### 4. 配置环境变量 -======= #### 4. 配置环境变量 ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd 请根据当前环境上CANN开发套件包的[安装方式](https://hiascend.com/document/redirect/CannCommunityInstSoftware),选择对应配置环境变量的命令。 - 默认路径,root用户安装CANN软件包 @@ -149,8 +121,6 @@ CANN软件包中提供了工程创建工具msopgen,AddCustom算子工程可通 aclnnStatus aclnnAddCustom(void *workspace, int64_t workspaceSize, aclOpExecutor **executor, aclrtStream stream); ``` 其中aclnnAddCustomGetWorkspaceSize为第一段接口,主要用于计算本次API调用计算过程中需要多少的workspace内存。获取到本次API计算需要的workspace大小之后,按照workspaceSize大小申请Device侧内存,然后调用第二段接口aclnnAddCustom执行计算。具体参考[AscendCL单算子调用](https://hiascend.com/document/redirect/CannCommunityAscendCInVorkSingleOp)>单算子API执行 章节。 - -<<<<<<< HEAD ======= ### 功能描述 * 算子功能:完成加法计算。 @@ -228,13 +198,10 @@ z = x + y | ACL_SUCCESS | 成功 | | ACL_ERROR | 失败 | ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd ## 更新说明 | 时间 | 更新事项 | |----|------| -<<<<<<< HEAD | 2025/03/27 | 修改不同用户环境配置 | -======= -| 2025/03/27 | 新增README | ->>>>>>> 6ca29cfc70d50859b7f7fa898b258ad71733f5bd + + -- Gitee From ff3d48ca5c1335020b928b04b4023ac59f16ed3a Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Mon, 31 Mar 2025 06:29:43 +0000 Subject: [PATCH 07/26] update atb_operator/AtbAddSample/AddAscendC/README.md. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: 万祖涛 --- atb_operator/AtbAddSample/AddAscendC/README.md | 1 - 1 file changed, 1 deletion(-) diff --git a/atb_operator/AtbAddSample/AddAscendC/README.md b/atb_operator/AtbAddSample/AddAscendC/README.md index ea2794454..425aaf5f8 100644 --- a/atb_operator/AtbAddSample/AddAscendC/README.md +++ b/atb_operator/AtbAddSample/AddAscendC/README.md @@ -121,7 +121,6 @@ CANN软件包中提供了工程创建工具msopgen,AddCustom算子工程可通 aclnnStatus aclnnAddCustom(void *workspace, int64_t workspaceSize, aclOpExecutor **executor, aclrtStream stream); ``` 其中aclnnAddCustomGetWorkspaceSize为第一段接口,主要用于计算本次API调用计算过程中需要多少的workspace内存。获取到本次API计算需要的workspace大小之后,按照workspaceSize大小申请Device侧内存,然后调用第二段接口aclnnAddCustom执行计算。具体参考[AscendCL单算子调用](https://hiascend.com/document/redirect/CannCommunityAscendCInVorkSingleOp)>单算子API执行 章节。 -======= ### 功能描述 * 算子功能:完成加法计算。 * 计算公式: -- Gitee From 5c4a8a8dc7366f508f8b2f0492aff536130be3dc Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Mon, 31 Mar 2025 06:32:24 +0000 Subject: [PATCH 08/26] update atb_operator/AtbAddSample/readme.md. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: 万祖涛 --- atb_operator/AtbAddSample/readme.md | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/atb_operator/AtbAddSample/readme.md b/atb_operator/AtbAddSample/readme.md index d205e5429..dbab64271 100644 --- a/atb_operator/AtbAddSample/readme.md +++ b/atb_operator/AtbAddSample/readme.md @@ -3,13 +3,14 @@ 本样例基于AscendC自定义Add算子,开发了ATB插件并进行了插件调用测试. ### 项目结构介绍 +``` -├── AddAscendC //Add AscendC自定义算子工程 +├── AddAscendC //Add AscendC自定义算子工程 ├── AddOperationATBPlugin //AddOperation ATB插件代码 -├── AddOperationTest //AddOperation 测试代码 - +├── AddOperationTest //AddOperation 测试代码 +``` ### 样例运行 #### Add AscendC自定义算子部署 -- Gitee From 5165179bcedf8b303bba7a84a7487a445c02dc7b Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Mon, 31 Mar 2025 06:33:10 +0000 Subject: [PATCH 09/26] update atb_operator/AtbAddSample/readme.md. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: 万祖涛 --- .../AtbAddSample/AddAscendC/CMakeLists.txt | 69 -- .../AtbAddSample/AddAscendC/CMakePresets.json | 63 -- .../AtbAddSample/AddAscendC/README.md | 206 ----- atb_operator/AtbAddSample/AddAscendC/build.sh | 37 - .../AddAscendC/cmake/config.cmake | 25 - .../AtbAddSample/AddAscendC/cmake/func.cmake | 228 ----- .../AtbAddSample/AddAscendC/cmake/intf.cmake | 26 - .../AddAscendC/cmake/makeself.cmake | 17 - .../AddAscendC/cmake/util/__init__.py | 8 - .../cmake/util/ascendc_bin_param_build.py | 129 --- .../cmake/util/ascendc_impl_build.py | 446 ---------- .../cmake/util/ascendc_ops_config.py | 114 --- .../cmake/util/ascendc_replay_build.py | 65 -- .../cmake/util/batch_replay_impl.temp | 117 --- .../cmake/util/code_channel_infer.py | 115 --- .../AddAscendC/cmake/util/const_var.py | 33 - .../cmake/util/gen_impl_and_mrege_json.sh | 57 -- .../AddAscendC/cmake/util/gen_ops_filter.sh | 62 -- .../AddAscendC/cmake/util/gen_version_info.sh | 6 - .../AddAscendC/cmake/util/insert_op_info.py | 36 - .../cmake/util/insert_simplified_keys.py | 248 ------ .../AddAscendC/cmake/util/kernel_entry.py | 115 --- .../AddAscendC/cmake/util/kernel_impl.temp | 10 - .../AddAscendC/cmake/util/makeself/COPYING | 339 -------- .../AddAscendC/cmake/util/makeself/README.md | 246 ------ .../AddAscendC/cmake/util/makeself/VERSION | 1 - .../cmake/util/makeself/make-release.sh | 9 - .../cmake/util/makeself/makeself-header.sh | 660 -------------- .../AddAscendC/cmake/util/makeself/makeself.1 | 110 --- .../cmake/util/makeself/makeself.lsm | 16 - .../cmake/util/makeself/makeself.sh | 822 ------------------ .../cmake/util/makeself/run-tests.sh | 8 - .../cmake/util/merge_aicpu_info_json.sh | 31 - .../AddAscendC/cmake/util/opdesc_parser.py | 260 ------ .../cmake/util/parse_ini_to_json.py | 338 ------- .../AddAscendC/cmake/util/preset_parse.py | 23 - .../AddAscendC/cmake/util/replay_codegen.py | 105 --- .../AddAscendC/cmake/util/replay_impl.temp | 120 --- .../cmake/util/tiling_data_def_build.py | 84 -- .../AddAscendC/framework/CMakeLists.txt | 11 - .../framework/tf_plugin/CMakeLists.txt | 14 - .../tf_plugin/tensorflow_add_custom_plugin.cc | 23 - .../AddAscendC/op_host/CMakeLists.txt | 82 -- .../AddAscendC/op_host/add_custom.cpp | 117 --- .../AddAscendC/op_host/add_custom_tiling.h | 22 - .../AddAscendC/op_kernel/CMakeLists.txt | 61 -- .../AddAscendC/op_kernel/add_custom.cpp | 134 --- .../AddAscendC/scripts/install.sh | 318 ------- .../AddAscendC/scripts/upgrade.sh | 151 ---- atb_operator/AtbAddSample/readme.md | 42 +- .../AtbEyeSample/EyeOp/CMakeLists.txt | 69 -- .../AtbEyeSample/EyeOp/CMakePresets.json | 63 -- atb_operator/AtbEyeSample/EyeOp/README.md | 216 ----- atb_operator/AtbEyeSample/EyeOp/build.sh | 80 -- .../AtbEyeSample/EyeOp/cmake/config.cmake | 25 - .../AtbEyeSample/EyeOp/cmake/func.cmake | 192 ---- .../AtbEyeSample/EyeOp/cmake/intf.cmake | 26 - .../AtbEyeSample/EyeOp/cmake/makeself.cmake | 17 - .../AtbEyeSample/EyeOp/cmake/util/__init__.py | 8 - .../cmake/util/ascendc_bin_param_build.py | 127 --- .../EyeOp/cmake/util/ascendc_impl_build.py | 501 ----------- .../EyeOp/cmake/util/ascendc_ops_config.py | 120 --- .../EyeOp/cmake/util/ascendc_replay_build.py | 65 -- .../EyeOp/cmake/util/batch_replay_impl.temp | 117 --- .../EyeOp/cmake/util/code_channel_infer.py | 58 -- .../EyeOp/cmake/util/const_var.py | 33 - .../cmake/util/gen_impl_and_mrege_json.sh | 20 - .../EyeOp/cmake/util/gen_ops_filter.sh | 62 -- .../EyeOp/cmake/util/gen_version_info.sh | 6 - .../EyeOp/cmake/util/insert_op_info.py | 36 - .../cmake/util/insert_simplified_keys.py | 242 ------ .../EyeOp/cmake/util/kernel_entry.py | 115 --- .../EyeOp/cmake/util/kernel_impl.temp | 10 - .../EyeOp/cmake/util/makeself/COPYING | 339 -------- .../EyeOp/cmake/util/makeself/README.md | 246 ------ .../EyeOp/cmake/util/makeself/VERSION | 1 - .../EyeOp/cmake/util/makeself/make-release.sh | 9 - .../cmake/util/makeself/makeself-header.sh | 660 -------------- .../EyeOp/cmake/util/makeself/makeself.1 | 110 --- .../EyeOp/cmake/util/makeself/makeself.lsm | 16 - .../EyeOp/cmake/util/makeself/makeself.sh | 822 ------------------ .../EyeOp/cmake/util/makeself/run-tests.sh | 8 - .../EyeOp/cmake/util/merge_aicpu_info_json.sh | 31 - .../EyeOp/cmake/util/opdesc_parser.py | 298 ------- .../EyeOp/cmake/util/parse_ini_to_json.py | 338 ------- .../EyeOp/cmake/util/preset_parse.py | 23 - .../EyeOp/cmake/util/replay_codegen.py | 105 --- .../EyeOp/cmake/util/replay_impl.temp | 120 --- .../EyeOp/cmake/util/tiling_data_def_build.py | 87 -- .../EyeOp/framework/CMakeLists.txt | 11 - .../EyeOp/framework/tf_plugin/CMakeLists.txt | 14 - .../tf_plugin/tensorflow_eye_plugin.cc | 12 - .../AtbEyeSample/EyeOp/op_host/CMakeLists.txt | 82 -- .../AtbEyeSample/EyeOp/op_host/eye.cpp | 173 ---- .../AtbEyeSample/EyeOp/op_host/eye_tiling.h | 31 - .../EyeOp/op_kernel/CMakeLists.txt | 68 -- .../AtbEyeSample/EyeOp/op_kernel/eye.cpp | 91 -- .../AtbEyeSample/EyeOp/scripts/install.sh | 318 ------- .../AtbEyeSample/EyeOp/scripts/upgrade.sh | 151 ---- .../AtbTrilSample/TrilOp/CMakeLists.txt | 69 -- .../AtbTrilSample/TrilOp/CMakePresets.json | 63 -- atb_operator/AtbTrilSample/TrilOp/README.md | 193 ---- atb_operator/AtbTrilSample/TrilOp/build.sh | 74 -- .../AtbTrilSample/TrilOp/cmake/config.cmake | 25 - .../AtbTrilSample/TrilOp/cmake/func.cmake | 192 ---- .../AtbTrilSample/TrilOp/cmake/intf.cmake | 26 - .../AtbTrilSample/TrilOp/cmake/makeself.cmake | 17 - .../TrilOp/cmake/util/__init__.py | 8 - .../cmake/util/ascendc_bin_param_build.py | 127 --- .../TrilOp/cmake/util/ascendc_impl_build.py | 496 ----------- .../TrilOp/cmake/util/ascendc_ops_config.py | 114 --- .../TrilOp/cmake/util/ascendc_replay_build.py | 65 -- .../TrilOp/cmake/util/batch_replay_impl.temp | 117 --- .../TrilOp/cmake/util/code_channel_infer.py | 58 -- .../TrilOp/cmake/util/const_var.py | 33 - .../cmake/util/gen_impl_and_mrege_json.sh | 20 - .../TrilOp/cmake/util/gen_ops_filter.sh | 62 -- .../TrilOp/cmake/util/gen_version_info.sh | 6 - .../TrilOp/cmake/util/insert_op_info.py | 36 - .../cmake/util/insert_simplified_keys.py | 242 ------ .../TrilOp/cmake/util/kernel_entry.py | 115 --- .../TrilOp/cmake/util/kernel_impl.temp | 10 - .../TrilOp/cmake/util/makeself/COPYING | 339 -------- .../TrilOp/cmake/util/makeself/README.md | 246 ------ .../TrilOp/cmake/util/makeself/VERSION | 1 - .../cmake/util/makeself/make-release.sh | 9 - .../cmake/util/makeself/makeself-header.sh | 660 -------------- .../TrilOp/cmake/util/makeself/makeself.1 | 110 --- .../TrilOp/cmake/util/makeself/makeself.lsm | 16 - .../TrilOp/cmake/util/makeself/makeself.sh | 822 ------------------ .../TrilOp/cmake/util/makeself/run-tests.sh | 8 - .../cmake/util/merge_aicpu_info_json.sh | 31 - .../TrilOp/cmake/util/opdesc_parser.py | 277 ------ .../TrilOp/cmake/util/parse_ini_to_json.py | 338 ------- .../TrilOp/cmake/util/preset_parse.py | 23 - .../TrilOp/cmake/util/replay_codegen.py | 105 --- .../TrilOp/cmake/util/replay_impl.temp | 120 --- .../cmake/util/tiling_data_def_build.py | 87 -- .../TrilOp/framework/CMakeLists.txt | 11 - .../TrilOp/framework/tf_plugin/CMakeLists.txt | 14 - .../tf_plugin/tensorflow_tril_plugin.cc | 20 - .../TrilOp/op_host/CMakeLists.txt | 82 -- .../AtbTrilSample/TrilOp/op_host/tril.cpp | 215 ----- .../TrilOp/op_host/tril_tiling.h | 34 - .../TrilOp/op_kernel/CMakeLists.txt | 68 -- .../AtbTrilSample/TrilOp/op_kernel/tril.cpp | 278 ------ .../AtbTrilSample/TrilOp/scripts/install.sh | 318 ------- .../AtbTrilSample/TrilOp/scripts/upgrade.sh | 151 ---- 148 files changed, 15 insertions(+), 19287 deletions(-) delete mode 100644 atb_operator/AtbAddSample/AddAscendC/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AddAscendC/CMakePresets.json delete mode 100644 atb_operator/AtbAddSample/AddAscendC/README.md delete mode 100755 atb_operator/AtbAddSample/AddAscendC/build.sh delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/config.cmake delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/func.cmake delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/intf.cmake delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/makeself.cmake delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/__init__.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/const_var.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp delete mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING delete mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/README.md delete mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION delete mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh delete mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh delete mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 delete mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh delete mode 100644 atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/preset_parse.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp delete mode 100755 atb_operator/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py delete mode 100644 atb_operator/AtbAddSample/AddAscendC/framework/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc delete mode 100644 atb_operator/AtbAddSample/AddAscendC/op_host/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AddAscendC/op_host/add_custom.cpp delete mode 100644 atb_operator/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h delete mode 100644 atb_operator/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt delete mode 100644 atb_operator/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp delete mode 100644 atb_operator/AtbAddSample/AddAscendC/scripts/install.sh delete mode 100644 atb_operator/AtbAddSample/AddAscendC/scripts/upgrade.sh delete mode 100755 atb_operator/AtbEyeSample/EyeOp/CMakeLists.txt delete mode 100755 atb_operator/AtbEyeSample/EyeOp/CMakePresets.json delete mode 100644 atb_operator/AtbEyeSample/EyeOp/README.md delete mode 100755 atb_operator/AtbEyeSample/EyeOp/build.sh delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/config.cmake delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/func.cmake delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/intf.cmake delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/makeself.cmake delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/__init__.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_bin_param_build.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_impl_build.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_ops_config.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_replay_build.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/batch_replay_impl.temp delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/code_channel_infer.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/const_var.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_impl_and_mrege_json.sh delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_ops_filter.sh delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_version_info.sh delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_op_info.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_simplified_keys.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_entry.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_impl.temp delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/COPYING delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/README.md delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/VERSION delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/make-release.sh delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself-header.sh delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.1 delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.lsm delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.sh delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/run-tests.sh delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/merge_aicpu_info_json.sh delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/opdesc_parser.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/parse_ini_to_json.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/preset_parse.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_codegen.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_impl.temp delete mode 100755 atb_operator/AtbEyeSample/EyeOp/cmake/util/tiling_data_def_build.py delete mode 100755 atb_operator/AtbEyeSample/EyeOp/framework/CMakeLists.txt delete mode 100755 atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/CMakeLists.txt delete mode 100755 atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/tensorflow_eye_plugin.cc delete mode 100755 atb_operator/AtbEyeSample/EyeOp/op_host/CMakeLists.txt delete mode 100755 atb_operator/AtbEyeSample/EyeOp/op_host/eye.cpp delete mode 100755 atb_operator/AtbEyeSample/EyeOp/op_host/eye_tiling.h delete mode 100755 atb_operator/AtbEyeSample/EyeOp/op_kernel/CMakeLists.txt delete mode 100755 atb_operator/AtbEyeSample/EyeOp/op_kernel/eye.cpp delete mode 100755 atb_operator/AtbEyeSample/EyeOp/scripts/install.sh delete mode 100755 atb_operator/AtbEyeSample/EyeOp/scripts/upgrade.sh delete mode 100755 atb_operator/AtbTrilSample/TrilOp/CMakeLists.txt delete mode 100755 atb_operator/AtbTrilSample/TrilOp/CMakePresets.json delete mode 100644 atb_operator/AtbTrilSample/TrilOp/README.md delete mode 100755 atb_operator/AtbTrilSample/TrilOp/build.sh delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/config.cmake delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/func.cmake delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/intf.cmake delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/makeself.cmake delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/__init__.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_bin_param_build.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_impl_build.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_ops_config.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_replay_build.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/batch_replay_impl.temp delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/code_channel_infer.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/const_var.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_impl_and_mrege_json.sh delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_ops_filter.sh delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_version_info.sh delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_op_info.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_simplified_keys.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_entry.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_impl.temp delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/COPYING delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/README.md delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/VERSION delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/make-release.sh delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself-header.sh delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.1 delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.lsm delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.sh delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/run-tests.sh delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/merge_aicpu_info_json.sh delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/opdesc_parser.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/parse_ini_to_json.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/preset_parse.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_codegen.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_impl.temp delete mode 100755 atb_operator/AtbTrilSample/TrilOp/cmake/util/tiling_data_def_build.py delete mode 100755 atb_operator/AtbTrilSample/TrilOp/framework/CMakeLists.txt delete mode 100755 atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/CMakeLists.txt delete mode 100755 atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/tensorflow_tril_plugin.cc delete mode 100755 atb_operator/AtbTrilSample/TrilOp/op_host/CMakeLists.txt delete mode 100755 atb_operator/AtbTrilSample/TrilOp/op_host/tril.cpp delete mode 100755 atb_operator/AtbTrilSample/TrilOp/op_host/tril_tiling.h delete mode 100755 atb_operator/AtbTrilSample/TrilOp/op_kernel/CMakeLists.txt delete mode 100755 atb_operator/AtbTrilSample/TrilOp/op_kernel/tril.cpp delete mode 100755 atb_operator/AtbTrilSample/TrilOp/scripts/install.sh delete mode 100755 atb_operator/AtbTrilSample/TrilOp/scripts/upgrade.sh diff --git a/atb_operator/AtbAddSample/AddAscendC/CMakeLists.txt b/atb_operator/AtbAddSample/AddAscendC/CMakeLists.txt deleted file mode 100644 index 584132d80..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/CMakeLists.txt +++ /dev/null @@ -1,69 +0,0 @@ -cmake_minimum_required(VERSION 3.16.0) -project(opp) -if(ENABLE_CROSS_COMPILE) - if(${CMAKE_SYSTEM_PROCESSOR} STREQUAL x86_64) - set(CROSS_COMPILE_PLATFORM aarch64) - else() - set(CROSS_COMPILE_PLATFORM x86_64) - endif() - set(PLATFORM ${CMAKE_SYSTEM_PROCESSOR}) - set(CMAKE_COMPILE_COMPILER_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/linux/${CROSS_COMPILE_PLATFORM}/) - set(CMAKE_COMPILE_RUNTIME_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/${CROSS_COMPILE_PLATFORM}/) - set(CMAKE_SYSTEM_PROCESSOR ${CROSS_COMPILE_PLATFORM}) - set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) - set(CMAKE_CXX_COMPILER ${CMAKE_CROSS_PLATFORM_COMPILER}) -else() - set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) -endif() - -include(cmake/config.cmake) -include(cmake/func.cmake) -include(cmake/intf.cmake) - -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/framework) - add_subdirectory(framework) -endif() -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_host) - add_subdirectory(op_host) -endif() -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_kernel) - add_subdirectory(op_kernel) -endif() -if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) - add_subdirectory(testcases) -endif() - -# modify vendor_name in install.sh and upgrade.sh -add_custom_command(OUTPUT ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh - COMMAND mkdir -p ${CMAKE_BINARY_DIR}/scripts - COMMAND cp -r ${CMAKE_SOURCE_DIR}/scripts/* ${CMAKE_BINARY_DIR}/scripts/ - COMMAND sed -i "s/vendor_name=customize/vendor_name=${vendor_name}/g" ${CMAKE_BINARY_DIR}/scripts/* -) -add_custom_target(modify_vendor ALL DEPENDS ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh) -install(DIRECTORY ${CMAKE_BINARY_DIR}/scripts/ DESTINATION . FILE_PERMISSIONS OWNER_EXECUTE OWNER_READ GROUP_READ) - -install(FILES ${CMAKE_SOURCE_DIR}/custom.proto DESTINATION packages OPTIONAL) - -get_system_info(SYSTEM_INFO) - -# gen version.info -add_custom_target(gen_version_info ALL - COMMAND bash ${CMAKE_CURRENT_SOURCE_DIR}/cmake/util/gen_version_info.sh ${ASCEND_CANN_PACKAGE_PATH} ${CMAKE_CURRENT_BINARY_DIR} -) - -install(FILES ${CMAKE_CURRENT_BINARY_DIR}/version.info - DESTINATION packages/vendors/${vendor_name}/) - -# CPack config -set(CPACK_PACKAGE_NAME ${CMAKE_PROJECT_NAME}) -set(CPACK_PACKAGE_VERSION ${CMAKE_PROJECT_VERSION}) -set(CPACK_PACKAGE_DESCRIPTION "CPack opp project") -set(CPACK_PACKAGE_DESCRIPTION_SUMMARY "CPack opp project") -set(CPACK_PACKAGE_DIRECTORY ${CMAKE_INSTALL_PREFIX}) -set(CPACK_PACKAGE_FILE_NAME "custom_opp_${SYSTEM_INFO}.run") -set(CPACK_GENERATOR External) -set(CPACK_CMAKE_GENERATOR "Unix Makefiles") -set(CPACK_EXTERNAL_ENABLE_STAGING TRUE) -set(CPACK_EXTERNAL_PACKAGE_SCRIPT ${CMAKE_SOURCE_DIR}/cmake/makeself.cmake) -set(CPACK_EXTERNAL_BUILT_PACKAGES ${CPACK_PACKAGE_DIRECTORY}/_CPack_Packages/Linux/External/${CPACK_PACKAGE_FILE_NAME}/${CPACK_PACKAGE_FILE_NAME}) -include(CPack) diff --git a/atb_operator/AtbAddSample/AddAscendC/CMakePresets.json b/atb_operator/AtbAddSample/AddAscendC/CMakePresets.json deleted file mode 100644 index ddec9431b..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/CMakePresets.json +++ /dev/null @@ -1,63 +0,0 @@ -{ - "version": 1, - "cmakeMinimumRequired": { - "major": 3, - "minor": 19, - "patch": 0 - }, - "configurePresets": [ - { - "name": "default", - "displayName": "Default Config", - "description": "Default build using Unix Makefiles generator", - "generator": "Unix Makefiles", - "binaryDir": "${sourceDir}/build_out", - "cacheVariables": { - "CMAKE_BUILD_TYPE": { - "type": "STRING", - "value": "Release" - }, - "ENABLE_SOURCE_PACKAGE": { - "type": "BOOL", - "value": "True" - }, - "ENABLE_BINARY_PACKAGE": { - "type": "BOOL", - "value": "True" - }, - "ASCEND_COMPUTE_UNIT": { - "type": "STRING", - "value": "ascend310b;ascend910b" - }, - "ENABLE_TEST": { - "type": "BOOL", - "value": "True" - }, - "vendor_name": { - "type": "STRING", - "value": "customize" - }, - "ASCEND_CANN_PACKAGE_PATH": { - "type": "PATH", - "value": "/usr/local/Ascend/ascend-toolkit/latest" - }, - "ASCEND_PYTHON_EXECUTABLE": { - "type": "STRING", - "value": "python3" - }, - "CMAKE_INSTALL_PREFIX": { - "type": "PATH", - "value": "${sourceDir}/build_out" - }, - "ENABLE_CROSS_COMPILE": { - "type": "BOOL", - "value": "False" - }, - "CMAKE_CROSS_PLATFORM_COMPILER": { - "type": "PATH", - "value": "/usr/bin/aarch64-linux-gnu-g++" - } - } - } - ] -} diff --git a/atb_operator/AtbAddSample/AddAscendC/README.md b/atb_operator/AtbAddSample/AddAscendC/README.md deleted file mode 100644 index 425aaf5f8..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/README.md +++ /dev/null @@ -1,206 +0,0 @@ - -# 概述 -本样例基于AddCustom算子工程,介绍了单算子工程及aclnn接口说明。 - -## 算子工程介绍 -### 算子描述 -Add算子实现了两个数据相加,返回相加结果的功能,该算子实现了非32B对齐shape下的算子功能,完成了算子的泛化实现。 -对应的数学表达式为: -``` -z = x + y -``` - -### 算子规格描述 - - - - - - - - - - - -
算子类型(OpType)Add
算子输入nameshapedata typeformat
x-float32,float16,int32,int8ND
y-float32,float16,int32,int8ND
算子输出z-float32,float16,int32,int8ND
核函数名add_custom
- - - -### 算子工程文件介绍 -其中,算子工程目录AddCustom包含算子实现的模板文件、编译脚本等,如下所示: -``` -├── AddCustom //Add自定义算子工程 -│ ├── cmake -│ ├── framework // 算子插件实现文件目录,单算子模型文件的生成不依赖算子适配插件,无需关注 -│ ├── op_host // host侧实现文件 -│ ├── op_kernel // kernel侧实现文件 -│ ├── scripts // 自定义算子工程打包相关脚本所在目录 -│ ├── build.sh // 编译入口脚本 -│ ├── CMakeLists.txt // 算子工程的CMakeLists.txt -│ └── CMakePresets.json // 编译配置项 -``` -CANN软件包中提供了工程创建工具msopgen,AddCustom算子工程可通过AddCustom.json自动创建,具体请参考[Ascend C算子开发](https://hiascend.com/document/redirect/CannCommunityOpdevAscendC)>算子开发>算子开发工程>基于自定义算子工程的算子开发>创建算子工程 章节。 - -### 编译运行样例算子 -针对自定义算子工程,编译运行包含如下步骤: -- 编译自定义算子工程生成算子安装包; -- 安装自定义算子到算子库中; -- 调用执行自定义算子; - -详细操作如下所示。 - -#### 1. 获取源码包 -编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 -#### 2. 编译算子工程 - 编译自定义算子工程,构建生成自定义算子包。 - - - 执行如下命令,切换到算子工程AddCustom目录。 - - ```bash - cd ${git_clone_path}/samples/atb_operator/AtbAddSample/AddOp - ``` - - - 修改CMakePresets.json中ASCEND_CANN_PACKAGE_PATH为CANN软件包安装后的实际路径。 - - - ```json - { - …… - "configurePresets": [ - { - …… - "ASCEND_CANN_PACKAGE_PATH": { - "type": "PATH", - "value": "/usr/local/Ascend/ascend-toolkit/latest" //请替换为CANN软件包安装后的实际路径。eg:/home/HwHiAiUser/Ascend/ascend-toolkit/latest - }, - …… - } - ] - } - ``` - - 在算子工程AddOp目录下执行如下命令,进行算子工程编译。 - - ```bash - ./build.sh - ``` -编译成功后,会在当前目录下创建build_out目录,并在build_out目录下生成自定义算子安装包custom_opp_\_\.run,例如“custom_opp_ubuntu_x86_64.run”。 - -备注:如果要使用dump调试功能,需要移除op_host内和CMakeLists.txt内的Atlas 200/500 A2 推理产品的配置。 - - -#### 3. 部署算子包 - -执行如下命令,在自定义算子安装包所在路径下,安装自定义算子包。 - ```bash - cd build_out - ./custom_opp__.run - ``` -命令执行成功后,自定义算子包中的相关文件将部署至当前环境的OPP算子库的vendors/customize目录中。 - -#### 4. 配置环境变量 - - 请根据当前环境上CANN开发套件包的[安装方式](https://hiascend.com/document/redirect/CannCommunityInstSoftware),选择对应配置环境变量的命令。 - - 默认路径,root用户安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=/usr/local/Ascend/ascend-toolkit/latest - ``` - - 默认路径,非root用户安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=$HOME/Ascend/ascend-toolkit/latest - ``` - - 指定路径install_path,安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=${install_path}/ascend-toolkit/latest - ``` - -## aclnn接口说明 - -自定义算子编译部署后,会自动生成单算子API,可以直接在应用程序中调用。算子API的形式一般定义为“两段式接口”,形如: - ```cpp - aclnnStatus aclnnAddCustomGetWorkspaceSize(const aclTensor *x, const aclTensor *y, const alcTensor *out, uint64_t workspaceSize, aclOpExecutor **executor); - aclnnStatus aclnnAddCustom(void *workspace, int64_t workspaceSize, aclOpExecutor **executor, aclrtStream stream); - ``` -其中aclnnAddCustomGetWorkspaceSize为第一段接口,主要用于计算本次API调用计算过程中需要多少的workspace内存。获取到本次API计算需要的workspace大小之后,按照workspaceSize大小申请Device侧内存,然后调用第二段接口aclnnAddCustom执行计算。具体参考[AscendCL单算子调用](https://hiascend.com/document/redirect/CannCommunityAscendCInVorkSingleOp)>单算子API执行 章节。 -### 功能描述 -* 算子功能:完成加法计算。 -* 计算公式: -``` -z = x + y -``` - -#### 参数说明 -##### aclnnAddCustomGetWorkspaceSize: - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
接口参数描述
aclnnAddCustomGetWorkspaceSizex输入x的Tensor,支持flaot/half/int8/int32类型,ND排布格式
y输入y的Tensor,支持flaot/half/int8/int32类型,ND排布格式
out输出z的Tensor,支持flaot/half/int8/int32类型,ND排布格式
workspaceSize返回需要在Device侧申请的workspace大小
executor返回op执行器,包含了算子计算流程
- - -| 返回值 | 描述 | -| ---- | ---- | -| ACL_SUCCESS | 成功 | -| ACL_ERROR | 失败 | - -##### aclnnAddCustom - - - - - - - - - - - - - - - - - - - - - - - -
接口参数描述
aclnnAddCustomworkspace输入Device侧申请的workspace内存
workspaceSizeworkspace大小
executor输入op执行器,包含了算子计算流程
stream输入执行流
- -| 返回值 | 描述 | -| ---- | ---- | -| ACL_SUCCESS | 成功 | -| ACL_ERROR | 失败 | - - -## 更新说明 -| 时间 | 更新事项 | -|----|------| -| 2025/03/27 | 修改不同用户环境配置 | - - diff --git a/atb_operator/AtbAddSample/AddAscendC/build.sh b/atb_operator/AtbAddSample/AddAscendC/build.sh deleted file mode 100755 index 4be96d7d8..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/build.sh +++ /dev/null @@ -1,37 +0,0 @@ -#!/bin/bash -script_path=$(realpath $(dirname $0)) - - -mkdir -p build_out -rm -rf build_out/* -cd build_out - -cmake_version=$(cmake --version | grep "cmake version" | awk '{print $3}') -if [ "$cmake_version" \< "3.19.0" ] ; then - opts=$(python3 $script_path/cmake/util/preset_parse.py $script_path/CMakePresets.json) - echo $opts - cmake .. $opts -else - cmake .. --preset=default -fi -target=package -if [ "$1"x != ""x ]; then target=$1; fi - -cmake --build . --target $target -j16 -if [ $? -ne 0 ]; then exit 1; fi - -if [ $target = "package" ]; then - if test -d ./op_kernel/binary ; then - ./cust*.run - if [ $? -ne 0 ]; then exit 1; fi - cmake --build . --target binary -j16 - if [ $? -ne 0 ]; then exit 1; fi - cmake --build . --target $target -j16 - fi -fi - -# for debug -# cd build_out -# make -# cpack -# verbose append -v diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/config.cmake b/atb_operator/AtbAddSample/AddAscendC/cmake/config.cmake deleted file mode 100755 index 886119daa..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/config.cmake +++ /dev/null @@ -1,25 +0,0 @@ - -set(CMAKE_CXX_FLAGS_DEBUG "") -set(CMAKE_CXX_FLAGS_RELEASE "") - -if (NOT DEFINED vendor_name) - set(vendor_name customize CACHE STRING "") -endif() -if (NOT DEFINED ASCEND_CANN_PACKAGE_PATH) - set(ASCEND_CANN_PACKAGE_PATH /usr/local/Ascend/latest CACHE PATH "") -endif() -if (NOT DEFINED ASCEND_PYTHON_EXECUTABLE) - set(ASCEND_PYTHON_EXECUTABLE python3 CACHE STRING "") -endif() -if (NOT DEFINED ASCEND_COMPUTE_UNIT) - message(FATAL_ERROR "ASCEND_COMPUTE_UNIT not set in CMakePreset.json ! -") -endif() -set(ASCEND_TENSOR_COMPILER_PATH ${ASCEND_CANN_PACKAGE_PATH}/compiler) -set(ASCEND_CCEC_COMPILER_PATH ${ASCEND_TENSOR_COMPILER_PATH}/ccec_compiler/bin) -set(ASCEND_AUTOGEN_PATH ${CMAKE_BINARY_DIR}/autogen) -set(ASCEND_FRAMEWORK_TYPE tensorflow) -file(MAKE_DIRECTORY ${ASCEND_AUTOGEN_PATH}) -set(CUSTOM_COMPILE_OPTIONS "custom_compile_options.ini") -execute_process(COMMAND rm -rf ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} - COMMAND touch ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS}) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/func.cmake b/atb_operator/AtbAddSample/AddAscendC/cmake/func.cmake deleted file mode 100755 index ad187e7d6..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/func.cmake +++ /dev/null @@ -1,228 +0,0 @@ - -function(get_system_info SYSTEM_INFO) - if (UNIX) - execute_process(COMMAND grep -i ^id= /etc/os-release OUTPUT_VARIABLE TEMP) - string(REGEX REPLACE "\n|id=|ID=|\"" "" SYSTEM_NAME ${TEMP}) - set(${SYSTEM_INFO} ${SYSTEM_NAME}_${CMAKE_SYSTEM_PROCESSOR} PARENT_SCOPE) - elseif (WIN32) - message(STATUS "System is Windows. Only for pre-build.") - else () - message(FATAL_ERROR "${CMAKE_SYSTEM_NAME} not support.") - endif () -endfunction() - -function(opbuild) - message(STATUS "Opbuild generating sources") - cmake_parse_arguments(OPBUILD "" "OUT_DIR;PROJECT_NAME;ACCESS_PREFIX" "OPS_SRC" ${ARGN}) - execute_process(COMMAND ${CMAKE_COMPILE} -g -fPIC -shared -std=c++11 ${OPBUILD_OPS_SRC} -D_GLIBCXX_USE_CXX11_ABI=0 - -I ${ASCEND_CANN_PACKAGE_PATH}/include -L ${ASCEND_CANN_PACKAGE_PATH}/lib64 -lexe_graph -lregister -ltiling_api - -o ${OPBUILD_OUT_DIR}/libascend_all_ops.so - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("build ops lib info: ${EXEC_INFO}") - message("build ops lib error: ${EXEC_ERROR}") - message(FATAL_ERROR "opbuild run failed!") - endif() - set(proj_env "") - set(prefix_env "") - if (NOT "${OPBUILD_PROJECT_NAME}x" STREQUAL "x") - set(proj_env "OPS_PROJECT_NAME=${OPBUILD_PROJECT_NAME}") - endif() - if (NOT "${OPBUILD_ACCESS_PREFIX}x" STREQUAL "x") - set(prefix_env "OPS_DIRECT_ACCESS_PREFIX=${OPBUILD_ACCESS_PREFIX}") - endif() - execute_process(COMMAND ${proj_env} ${prefix_env} ${ASCEND_CANN_PACKAGE_PATH}/toolkit/tools/opbuild/op_build - ${OPBUILD_OUT_DIR}/libascend_all_ops.so ${OPBUILD_OUT_DIR} - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("opbuild ops info: ${EXEC_INFO}") - message("opbuild ops error: ${EXEC_ERROR}") - endif() - message(STATUS "Opbuild generating sources - done") -endfunction() - -function(add_ops_info_target) - cmake_parse_arguments(OPINFO "" "TARGET;OPS_INFO;OUTPUT;INSTALL_DIR" "" ${ARGN}) - get_filename_component(opinfo_file_path "${OPINFO_OUTPUT}" DIRECTORY) - add_custom_command(OUTPUT ${OPINFO_OUTPUT} - COMMAND mkdir -p ${opinfo_file_path} - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/parse_ini_to_json.py - ${OPINFO_OPS_INFO} ${OPINFO_OUTPUT} - ) - add_custom_target(${OPINFO_TARGET} ALL - DEPENDS ${OPINFO_OUTPUT} - ) - install(FILES ${OPINFO_OUTPUT} - DESTINATION ${OPINFO_INSTALL_DIR} - ) -endfunction() - -function(add_ops_compile_options OP_TYPE) - cmake_parse_arguments(OP_COMPILE "" "OP_TYPE" "COMPUTE_UNIT;OPTIONS" ${ARGN}) - file(APPEND ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} - "${OP_TYPE},${OP_COMPILE_COMPUTE_UNIT},${OP_COMPILE_OPTIONS}\n") -endfunction() - -function(add_ops_impl_target) - cmake_parse_arguments(OPIMPL "" "TARGET;OPS_INFO;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) - add_custom_command(OUTPUT ${OPIMPL_OUT_DIR}/.impl_timestamp - COMMAND mkdir -m 700 -p ${OPIMPL_OUT_DIR}/dynamic - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py - ${OPIMPL_OPS_INFO} - \"${OPIMPL_OPS_BATCH}\" \"${OPIMPL_OPS_ITERATE}\" - ${OPIMPL_IMPL_DIR} - ${OPIMPL_OUT_DIR}/dynamic - ${ASCEND_AUTOGEN_PATH} - - COMMAND rm -rf ${OPIMPL_OUT_DIR}/.impl_timestamp - COMMAND touch ${OPIMPL_OUT_DIR}/.impl_timestamp - DEPENDS ${OPIMPL_OPS_INFO} - ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py - ) - add_custom_target(${OPIMPL_TARGET} ALL - DEPENDS ${OPIMPL_OUT_DIR}/.impl_timestamp) - if (${ENABLE_SOURCE_PACKAGE}) - install(DIRECTORY ${OPIMPL_OUT_DIR}/dynamic - DESTINATION ${OPIMPL_INSTALL_DIR} - ) - endif() -endfunction() - -function(add_ops_replay_targets) - cmake_parse_arguments(OPREPLAY "" "OPS_INFO;COMPUTE_UNIT;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) - # ccec compile options - set(ccec_base_opts -c -O2 --cce-aicore-only -mllvm -cce-aicore-function-stack-size=16000 - -mllvm -cce-aicore-record-overflow=false -std=c++17) - set(ccec_extopts_ascend310p --cce-aicore-arch=dav-m200 -mllvm -cce-aicore-fp-ceiling=2) - set(ccec_extopts_ascend910 --cce-aicore-arch=dav-c100) - set(ccec_extopts_ascend910b --cce-aicore-arch=dav-c220-cube) - file(MAKE_DIRECTORY ${OPREPLAY_OUT_DIR}) - execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_replay_build.py - ${OPREPLAY_OPS_INFO} - "${OPREPLAY_OPS_BATCH}" "${OPREPLAY_OPS_ITERATE}" - ${OPREPLAY_IMPL_DIR} - ${OPREPLAY_OUT_DIR} - ${OPREPLAY_COMPUTE_UNIT} - ) - file(GLOB replay_kernel_entries ${OPREPLAY_OUT_DIR}/*.cce) - if (NOT "${replay_kernel_entries}x" STREQUAL "x") - foreach(replay_kernel_file ${replay_kernel_entries}) - get_filename_component(replay_kernel_file_name "${replay_kernel_file}" NAME) - string(REPLACE "_entry.cce" "" op_kerne_name ${replay_kernel_file_name}) - file(GLOB replay_lib_src ${OPREPLAY_OUT_DIR}/${op_kerne_name}*.cpp) - set(OP_TILING_DATA_H_PATH ${OPREPLAY_OUT_DIR}/${op_kerne_name}_tiling_data.h) - add_library(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} SHARED ${replay_lib_src}) - if(EXISTS ${OP_TILING_DATA_H_PATH}) - target_compile_options(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE - -include ${OP_TILING_DATA_H_PATH} - ) - endif() - target_compile_definitions(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE - ${op_kerne_name}=${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} - ) - target_compile_options(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE - -D__ASCENDC_REPLAY__ - ) - target_link_libraries(replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} PRIVATE intf_pub - tikreplaylib::${OPREPLAY_COMPUTE_UNIT} - register - ) - add_custom_command(OUTPUT ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o - COMMAND ccec ${ccec_base_opts} ${ccec_extopts_${OPREPLAY_COMPUTE_UNIT}} ${replay_kernel_file} - -o ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o - DEPENDS ${replay_kernel_file} - ) - add_custom_target(replay_kernel_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} ALL - DEPENDS ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o - ) - install(TARGETS replay_${op_kerne_name}_${OPREPLAY_COMPUTE_UNIT} - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_replay - ) - install(FILES ${OPREPLAY_OUT_DIR}/${op_kerne_name}_entry_${OPREPLAY_COMPUTE_UNIT}.o - DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_replay - ) - endforeach() - endif() -endfunction() - -function(add_npu_support_target) - cmake_parse_arguments(NPUSUP "" "TARGET;OPS_INFO_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) - get_filename_component(npu_sup_file_path "${NPUSUP_OUT_DIR}" DIRECTORY) - add_custom_command(OUTPUT ${NPUSUP_OUT_DIR}/npu_supported_ops.json - COMMAND mkdir -p ${NPUSUP_OUT_DIR} - COMMAND ${CMAKE_SOURCE_DIR}/cmake/util/gen_ops_filter.sh - ${NPUSUP_OPS_INFO_DIR} - ${NPUSUP_OUT_DIR} - ) - add_custom_target(npu_supported_ops ALL - DEPENDS ${NPUSUP_OUT_DIR}/npu_supported_ops.json - ) - install(FILES ${NPUSUP_OUT_DIR}/npu_supported_ops.json - DESTINATION ${NPUSUP_INSTALL_DIR} - ) -endfunction() - -function(add_bin_compile_target) - cmake_parse_arguments(BINCMP "" "TARGET;OPS_INFO;COMPUTE_UNIT;IMPL_DIR;ADP_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/src) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/gen) - execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_bin_param_build.py - ${BINCMP_OPS_INFO} ${BINCMP_OUT_DIR}/gen ${BINCMP_COMPUTE_UNIT} - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("ops binary compile scripts gen info: ${EXEC_INFO}") - message("ops binary compile scripts gen error: ${EXEC_ERROR}") - message(FATAL_ERROR "ops binary compile scripts gen failed!") - endif() - if (NOT TARGET binary) - add_custom_target(binary) - endif() - add_custom_target(${BINCMP_TARGET} - COMMAND cp -r ${BINCMP_IMPL_DIR}/*.* ${BINCMP_OUT_DIR}/src - ) - add_custom_target(${BINCMP_TARGET}_gen_ops_config - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/insert_simplified_keys.py -p ${BINCMP_OUT_DIR}/bin - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_ops_config.py -p ${BINCMP_OUT_DIR}/bin - -s ${BINCMP_COMPUTE_UNIT} - ) - add_dependencies(binary ${BINCMP_TARGET}_gen_ops_config) - file(GLOB bin_scripts ${BINCMP_OUT_DIR}/gen/*.sh) - foreach(bin_script ${bin_scripts}) - get_filename_component(bin_file ${bin_script} NAME_WE) - string(REPLACE "-" ";" bin_sep ${bin_file}) - list(GET bin_sep 0 op_type) - list(GET bin_sep 1 op_file) - list(GET bin_sep 2 op_index) - if (NOT TARGET ${BINCMP_TARGET}_${op_file}_copy) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file}) - add_custom_target(${BINCMP_TARGET}_${op_file}_copy - COMMAND cp ${BINCMP_ADP_DIR}/${op_file}.py ${BINCMP_OUT_DIR}/src/${op_type}.py - ) - install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} - DESTINATION ${BINCMP_INSTALL_DIR}/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) - install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json - DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT}/ OPTIONAL - ) - endif() - add_custom_target(${BINCMP_TARGET}_${op_file}_${op_index} - COMMAND export HI_PYTHON=${ASCEND_PYTHON_EXECUTABLE} && bash ${bin_script} ${BINCMP_OUT_DIR}/src/${op_type}.py ${BINCMP_OUT_DIR}/bin/${op_file} - WORKING_DIRECTORY ${BINCMP_OUT_DIR} - ) - add_dependencies(${BINCMP_TARGET}_${op_file}_${op_index} ${BINCMP_TARGET} ${BINCMP_TARGET}_${op_file}_copy) - add_dependencies(${BINCMP_TARGET}_gen_ops_config ${BINCMP_TARGET}_${op_file}_${op_index}) - endforeach() - install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json - DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) -endfunction() diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/intf.cmake b/atb_operator/AtbAddSample/AddAscendC/cmake/intf.cmake deleted file mode 100755 index 2f362c396..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/intf.cmake +++ /dev/null @@ -1,26 +0,0 @@ - -add_library(intf_pub INTERFACE) -target_compile_options(intf_pub INTERFACE - -fPIC - -fvisibility=hidden - -fvisibility-inlines-hidden - $<$:-O2> - $<$:-O0 -g> - $<$:-std=c++11> - $<$,$>:-ftrapv -fstack-check> - $<$:-pthread -Wfloat-equal -Wshadow -Wformat=2 -Wno-deprecated -Wextra> - $,-fstack-protector-strong,-fstack-protector-all> -) -target_compile_definitions(intf_pub INTERFACE - _GLIBCXX_USE_CXX11_ABI=0 - $<$:_FORTIFY_SOURCE=2> -) -target_include_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/include) -target_link_options(intf_pub INTERFACE - $<$,EXECUTABLE>:-pie> - $<$:-s> - -Wl,-z,relro - -Wl,-z,now - -Wl,-z,noexecstack -) -target_link_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/lib64) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/makeself.cmake b/atb_operator/AtbAddSample/AddAscendC/cmake/makeself.cmake deleted file mode 100755 index 48c565bfb..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/makeself.cmake +++ /dev/null @@ -1,17 +0,0 @@ -execute_process(COMMAND chmod +x ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh) -execute_process(COMMAND ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh - --header ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself-header.sh - --help-header ./help.info - --gzip --complevel 4 --nomd5 --sha256 - ./ ${CPACK_PACKAGE_FILE_NAME} "version:1.0" ./install.sh - WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} - RESULT_VARIABLE EXEC_RESULT - ERROR_VARIABLE EXEC_ERROR -) -if (NOT "${EXEC_RESULT}x" STREQUAL "0x") - message(FATAL_ERROR "CPack Command error: ${EXEC_RESULT}\n${EXEC_ERROR}") -endif() -execute_process(COMMAND cp ${CPACK_EXTERNAL_BUILT_PACKAGES} ${CPACK_PACKAGE_DIRECTORY}/ - COMMAND echo "Copy ${CPACK_EXTERNAL_BUILT_PACKAGES} to ${CPACK_PACKAGE_DIRECTORY}/" - WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} -) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/__init__.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/__init__.py deleted file mode 100755 index c4ddc893a..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/__init__.py +++ /dev/null @@ -1,8 +0,0 @@ -#!/usr/bin/env python -# -*- coding: utf-8 -*- - -import sys -import os - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) -sys.path.append(PYF_PATH) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py deleted file mode 100755 index decf34544..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_bin_param_build.py +++ /dev/null @@ -1,129 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import json -import hashlib -import const_var -import opdesc_parser - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - - -class BinParamBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - super().__init__(op_type) - self.soc = '' - self.out_path = '' - - def set_soc_version(self: any, soc: str): - self.soc = soc - - def set_out_path(self: any, out_path: str): - self.out_path = out_path - - def gen_input_json(self: any): - key_map = {} - count = len(self.input_dtype[0].split(',')) - for i in range(0, count): - inputs = [] - outputs = [] - attrs = [] - op_node = {} - for idx in range(0, len(self.input_name)): - idtypes = self.input_dtype[idx].split(',') - ifmts = self.input_fmt[idx].split(',') - itype = self.input_type[idx] - para = {} - para['name'] = self.input_name[idx] - para['index'] = idx - para['dtype'] = idtypes[i] - para['format'] = ifmts[i] - para['paramType'] = itype - para['shape'] = [-2] - if itype == 'dynamic': - inputs.append([para]) - else: - inputs.append(para) - for idx in range(0, len(self.output_name)): - odtypes = self.output_dtype[idx].split(',') - ofmts = self.output_fmt[idx].split(',') - otype = self.output_type[idx] - para = {} - para['name'] = self.output_name[idx] - para['index'] = idx - para['dtype'] = odtypes[i] - para['format'] = ofmts[i] - para['paramType'] = otype - para['shape'] = [-2] - if otype == 'dynamic': - outputs.append([para]) - else: - outputs.append(para) - for attr in self.attr_list: - att = {} - att['name'] = attr - atype = self.attr_val.get(attr).get('type').lower() - atype = atype.replace('list', 'list_') - att['dtype'] = atype - att['value'] = const_var.ATTR_DEF_VAL.get(atype) - attrs.append(att) - op_node['bin_filename'] = '' - op_node['inputs'] = inputs - op_node['outputs'] = outputs - if len(attrs) > 0: - op_node['attrs'] = attrs - param = {} - param['op_type'] = self.op_type - param['op_list'] = [op_node] - objstr = json.dumps(param, indent=' ') - md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() - while key_map.get(md5sum) is not None: - objstr += '1' - md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() - key_map[md5sum] = md5sum - bin_file = self.op_type + '_' + md5sum - op_node['bin_filename'] = bin_file - param_file = os.path.join(self.out_path, bin_file + '_param.json') - param_file = os.path.realpath(param_file) - with os.fdopen(os.open(param_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - json.dump(param, fd, indent=' ') - self._write_buld_cmd(param_file, bin_file, i) - - - def _write_buld_cmd(self: any, param_file: str, bin_file: str, index: int): - hard_soc = const_var.SOC_MAP_EXT.get(self.soc) - if not hard_soc: - hard_soc = soc.capitalize() - name_com = [self.op_type, self.op_file, str(index)] - compile_file = os.path.join(self.out_path, '-'.join(name_com) + '.sh') - compile_file = os.path.realpath(compile_file) - with os.fdopen(os.open(compile_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - fd.write('#!/bin/bash\n') - fd.write('echo "[{}] Generating {} ..."\n'.format(hard_soc, bin_file)) - cmd = const_var.BIN_CMD.format(fun=self.op_intf, soc=hard_soc, param=param_file, impl='""') - fd.write(cmd) - chk = const_var.CHK_CMD.format(res_file=bin_file + '.json') - fd.write(chk) - chk = const_var.CHK_CMD.format(res_file=bin_file + '.o') - fd.write(chk) - fd.write('echo "[{}] Generating {} Done"\n'.format(hard_soc, bin_file)) - - -def gen_bin_param_file(cfgfile: str, out_dir: str, soc: str): - op_descs = opdesc_parser.get_op_desc(cfgfile, [], [], BinParamBuilder, None) - for op_desc in op_descs: - op_desc.set_soc_version(soc) - op_desc.set_out_path(out_dir) - op_desc.gen_input_json() - - -if __name__ == '__main__': - if len(sys.argv) <= 3: - raise RuntimeError('arguments must greater than 3') - gen_bin_param_file(sys.argv[1], sys.argv[2], sys.argv[3]) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py deleted file mode 100755 index 7fe177da1..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_impl_build.py +++ /dev/null @@ -1,446 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import re -import stat -import opdesc_parser -import const_var - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - -IMPL_HEAD = ''' -import os, sys -import ctypes -import json -import shutil -from tbe.common.platform import get_soc_spec -from tbe.common.utils import para_check -from tbe.tikcpp import compile_op, replay_op, check_op_cap, generalize_op_params, get_code_channel, OpInfo -from tbe.common.buildcfg import get_default_build_config -from impl.util.platform_adapter import tbe_register -from tbe.common.buildcfg import get_current_build_config -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - -DTYPE_MAP = {"float32": ["DT_FLOAT", "float"], - "float16": ["DT_FLOAT16", "half"], - "int8": ["DT_INT8", "int8_t"], - "int16": ["DT_INT16", "int16_t"], - "int32": ["DT_INT32", "int32_t"], - "int64": ["DT_INT64", "int64_t"], - "uint1": ["DT_UINT1", "uint8_t"], - "uint8": ["DT_UINT8", "uint8_t"], - "uint16": ["DT_UINT16", "uint16_t"], - "uint32": ["DT_UINT32", "uint32_t"], - "uint64": ["DT_UINT64", "uint64_t"], - "bool": ["DT_BOOL", "bool"], - "double": ["DT_DOUBLE", "double"], - "dual": ["DT_DUAL", "unknown"], - "dual_sub_int8": ["DT_DUAL_SUB_INT8", "unknown"], - "dual_sub_uint8": ["DT_DUAL_SUB_UINT8", "unknown"], - "string": ["DT_STRING", "unknown"], - "complex64": ["DT_COMPLEX64", "unknown"], - "complex128": ["DT_COMPLEX128", "unknown"], - "qint8": ["DT_QINT8", "unknown"], - "qint16": ["DT_QINT16", "unknown"], - "qint32": ["DT_QINT32", "unknown"], - "quint8": ["DT_QUINT8", "unknown"], - "quint16": ["DT_QUINT16", "unknown"], - "resource": ["DT_RESOURCE", "unknown"], - "string_ref": ["DT_STRING_REF", "unknown"], - "int4": ["DT_INT4", "int8_t"], - "bfloat16": ["DT_BF16", "bfloat16_t"]} - -def get_dtype_fmt_options(__inputs__, __outputs__): - options = [] - for x in __inputs__ + __outputs__: - x_n = x.get("param_name").upper() - x_fmt = x.get("format") - x_dtype = x.get("dtype") - options.append("-DDTYPE_{n}={t}".format(n=x_n, t=DTYPE_MAP.get(x_dtype)[1])) - options.append("-DORIG_DTYPE_{n}={ot}".format(n=x_n, ot=DTYPE_MAP.get(x_dtype)[0])) - options.append("-DFORMAT_{n}=FORMAT_{f}".format(n=x_n, f=x_fmt)) - return options - -def load_dso(so_path): - try: - ctypes.CDLL(so_path) - except OSError as error : - print(error) - raise RuntimeError("cannot open %s" %(so_path)) - else: - print("load so succ ", so_path) - -def get_shortsoc_compile_option(compile_option_list: list, shortsoc:str): - compile_options = [] - if shortsoc in compile_option_list: - compile_options = compile_option_list[shortsoc] - elif '__ALLSOC__' in compile_option_list: - compile_options = compile_option_list['__ALLSOC__'] - return compile_options - -''' - -IMPL_API = ''' -@tbe_register.register_operator("{}") -@para_check.check_op_params({}) -def {}({}, kernel_name="{}", impl_mode=""): - if get_current_build_config("enable_op_prebuild"): - return - __inputs__, __outputs__, __attrs__ = _build_args({}) - options = get_dtype_fmt_options(__inputs__, __outputs__) - options += ["-x", "cce"] - ccec = os.environ.get('CCEC_REAL_PATH') - if ccec is None: - ccec = shutil.which("ccec") - if ccec != None: - ccec_path = os.path.dirname(ccec) - tikcpp_path = os.path.realpath(os.path.join(ccec_path, "..", "..", "tikcpp")) - else: - tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") - options.append("-I" + tikcpp_path) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) - options.append("-I" + os.path.join(PYF_PATH, "..", "ascendc", "common")) - if impl_mode == "high_performance": - options.append("-DHIGH_PERFORMANCE=1") - elif impl_mode == "high_precision": - options.append("-DHIGH_PRECISION=1") - if get_default_build_config("enable_deterministic_mode") == 1: - options.append("-DDETEMINISTIC_MODE=1") - - custom_compile_options = {}, - custom_all_compile_options = {}, - soc_version = get_soc_spec("SOC_VERSION") - soc_short = get_soc_spec("SHORT_SOC_VERSION").lower() - custom_compile_options_soc = get_shortsoc_compile_option(custom_compile_options[0], soc_short) - custom_all_compile_options_soc = get_shortsoc_compile_option(custom_all_compile_options[0], soc_short) - options += custom_all_compile_options_soc - options += custom_compile_options_soc - - origin_func_name = "{}" - ascendc_src_dir = "{}" - ascendc_src_file = "{}" - src = os.path.join(PYF_PATH, "..", "ascendc", ascendc_src_dir, ascendc_src_file) - if not os.path.exists(src): - src = os.path.join(PYF_PATH, ascendc_src_file) -''' - -REPLAY_OP_API = ''' - print("start replay Acend C Operator {}, kernel name is {}") - tikreplay_codegen_path = tikcpp_path + "/tikreplaylib/lib" - tikreplay_stub_path = tikcpp_path + "/tikreplaylib/lib/" + soc_version - print("start load libtikreplaylib_codegen.so and libtikreplaylib_stub.so") - codegen_so_path = tikreplay_codegen_path + "/libtikreplaylib_codegen.so" - replaystub_so_path = tikreplay_stub_path + "/libtikreplaylib_stub.so" - if PYF_PATH.endswith("dynamic"): - op_replay_path = os.path.join(PYF_PATH, "..", "..", "op_replay") - else: - op_replay_path = os.path.join(PYF_PATH, "..", "op_replay") - replayapi_so_path = os.path.join(op_replay_path, "libreplay_{}_" + soc_short + ".so") - load_dso(codegen_so_path) - load_dso(replaystub_so_path) - load_dso(replayapi_so_path) - op_type = "{}" - entry_obj = os.path.join(op_replay_path, "{}_entry_" + soc_short + ".o") - code_channel = get_code_channel(src, kernel_name, op_type, options) - op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ - attrs = __attrs__, impl_mode = impl_mode) - res, msg = replay_op(op_info, entry_obj, code_channel, src, options) - if not res: - print("call replay op failed for %s and get into call compile op" %(msg)) - compile_op(src, origin_func_name, op_info, options, code_channel, '{}') -''' - -COMPILE_OP_API = ''' - print("start compile Ascend C operator {}. kernel name is {}") - op_type = "{}" - code_channel = get_code_channel(src, kernel_name, op_type, options) - op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ - attrs = __attrs__, impl_mode = impl_mode, origin_inputs=[{}], origin_outputs = [{}]) - compile_op(src, origin_func_name, op_info, options, code_channel, '{}') -''' - -SUP_API = ''' -def {}({}, impl_mode=""): - __inputs__, __outputs__, __attrs__ = _build_args({}) - ret_str = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) - ret_dict = json.loads(ret_str) - err_code = ret_dict.get("ret_code") - sup = "Unknown" - reason = "Unknown reason" - if err_code is not None: - if err_code is 0: - sup = "True" - reason = "" - elif err_code is 1: - sup = "False" - reason = ret_dict.get("reason") - else: - sup = "Unknown" - reason = ret_dict.get("reason") - return sup, reason -''' -CAP_API = ''' -def {}({}, impl_mode=""): - __inputs__, __outputs__, __attrs__ = _build_args({}) - result = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) - return result.decode("utf-8") -''' -GLZ_API = ''' -@tbe_register.register_param_generalization("{}") -def {}_generalization({}, generalize_config=None): - __inputs__, __outputs__, __attrs__ = _build_args({}) - ret_str = generalize_op_params("{}", __inputs__, __outputs__, __attrs__, generalize_config) - return [json.loads(ret_str)] -''' - -ATTR_DEFAULT = {'bool': 'False', 'int': '0', 'float': '0.0', 'listInt': '[]', - 'listFloat': '[]', 'listBool': '[]', 'listListInt': '[[]]', 'str': ''} - - -def optype_snake(origin_str): - temp_str = origin_str[0].lower() + origin_str[1:] - new_str = re.sub(r'([A-Z])', r'_\1', temp_str).lower() - return new_str - - -class AdpBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - self.argsname = [] - self.argsdefv = [] - self.op_compile_option:str = '{}' - super().__init__(op_type) - - - def write_adapt(self: any, impl_path, path: str, op_compile_option_all:list = None): - self._build_paradefault() - if impl_path != "": - src_file = os.path.join(impl_path, self.op_file + '.cpp') - if not os.path.exists(src_file): - return - out_path = os.path.abspath(path) - if self.dynamic_shape and not out_path.endswith('dynamic'): - out_path = os.path.join(path, 'dynamic') - os.makedirs(out_path, mode=0o700, exist_ok=True) - adpfile = os.path.join(out_path, self.op_file + '.py') - self._gen_op_compile_option(op_compile_option_all) - with os.fdopen(os.open(adpfile, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - self._write_head(fd) - self._write_argparse(fd) - self._write_impl(fd) - if self.op_chk_support: - self._write_cap('check_supported', fd) - self._write_cap('get_op_support_info', fd) - if self.op_fmt_sel: - self._write_cap('op_select_format', fd) - self._write_cap('get_op_specific_info', fd) - if self.op_range_limit == 'limited' or self.op_range_limit == 'dynamic': - self._write_glz(fd) - - - def _gen_op_compile_option(self:any, op_compile_option_all:list =None): - if op_compile_option_all is not None: - if self.op_type in op_compile_option_all: - self.op_compile_option = op_compile_option_all[self.op_type] - elif "__all__" in op_compile_option_all: - self.op_compile_option = op_compile_option_all["__all__"] - - - def _ip_argpack(self: any, default: bool = True) -> list: - args = [] - for i in range(len(self.input_name)): - arg = self.input_name[i] - if default and self.argsdefv[i] is not None: - arg += '=' + self.argsdefv[i] - args.append(arg) - return args - - def _op_argpack(self: any, default: bool = True) -> list: - args = [] - argidx = len(self.input_name) - for i in range(len(self.output_name)): - arg = self.output_name[i] - if default and self.argsdefv[i + argidx] is not None: - arg += '=' + self.argsdefv[i + argidx] - args.append(arg) - return args - - def _attr_argpack(self: any, default: bool = True) -> list: - args = [] - argidx = len(self.input_name) + len(self.output_name) - for i in range(len(self.attr_list)): - att = self.attr_list[i] - arg = att - if default and self.argsdefv[i + argidx] is not None: - if self.attr_val.get(att).get('type') == 'str': - arg += '="' + self.argsdefv[i + argidx] + '"' - elif self.attr_val.get(att).get('type') == 'bool': - arg += '=' + self.argsdefv[i + argidx].capitalize() - else: - arg += '=' + self.argsdefv[i + argidx] - args.append(arg) - return args - - def _build_paralist(self: any, default: bool = True) -> str: - args = [] - args.extend(self._ip_argpack(default)) - args.extend(self._op_argpack(default)) - args.extend(self._attr_argpack(default)) - return ', '.join(args) - - def _io_parachk(self: any, types: list, type_name: str) -> list: - chk = [] - for iot in types: - if iot == 'optional': - ptype = 'OPTION' - else: - ptype = iot.upper() - chk.append('para_check.{}_{}'.format(ptype, type_name)) - return chk - - def _attr_parachk(self: any) -> list: - chk = [] - for att in self.attr_list: - if self.attr_val.get(att).get('paramType') == 'optional': - pt = 'OPTION' - else: - pt = self.attr_val.get(att).get('paramType').upper() - att_type = self.attr_val.get(att).get('type').upper() - att_type = att_type.replace('LIST', 'LIST_') - chk.append('para_check.{}_ATTR_{}'.format(pt, att_type)) - return chk - - def _build_parachk(self: any) -> str: - chk = [] - chk.extend(self._io_parachk(self.input_type, 'INPUT')) - chk.extend(self._io_parachk(self.output_type, 'OUTPUT')) - chk.extend(self._attr_parachk()) - chk.append('para_check.KERNEL_NAME') - return ', '.join(chk) - - def _build_paradefault(self: any): - optional = False - argtypes = [] - argtypes.extend(self.input_type) - argtypes.extend(self.output_type) - for atype in argtypes: - if atype == 'optional': - optional = True - if optional: - self.argsdefv.append('None') - else: - self.argsdefv.append(None) - for attr in self.attr_list: - atype = self.attr_val.get(attr).get('paramType') - if atype == 'optional': - optional = True - attrval = self.attr_val.get(attr).get('defaultValue') - if attrval is not None: - optional = True - if type == "bool": - attrval = attrval.capitalize() - elif type == "str": - attrval = "\"" + attrval + "\"" - self.argsdefv.append(attrval) - continue - if optional: - self.argsdefv.append(ATTR_DEFAULT.get(self.attr_val.get(attr).get('type'))) - else: - self.argsdefv.append(None) - - def _write_head(self: any, fd: object): - fd.write(IMPL_HEAD) - - def _write_argparse(self: any, fd: object): - args = self._build_paralist(False) - fd.write('def _build_args({}):\n'.format(args)) - fd.write(' __inputs__ = []\n') - fd.write(' for arg in [{}]:\n'.format(', '.join(self.input_name))) - fd.write(' if arg != None:\n') - fd.write(' if isinstance(arg, (list, tuple)):\n') - fd.write(' if len(arg) == 0:\n') - fd.write(' continue\n') - fd.write(' __inputs__.append(arg[0])\n') - fd.write(' else:\n') - fd.write(' __inputs__.append(arg)\n') - fd.write(' __outputs__ = []\n') - fd.write(' for arg in [{}]:\n'.format(', '.join(self.output_name))) - fd.write(' if arg != None:\n') - fd.write(' if isinstance(arg, (list, tuple)):\n') - fd.write(' if len(arg) == 0:\n') - fd.write(' continue\n') - fd.write(' __outputs__.append(arg[0])\n') - fd.write(' else:\n') - fd.write(' __outputs__.append(arg)\n') - fd.write(' __attrs__ = []\n') - for attr in self.attr_list: - fd.write(' if {} != None:\n'.format(attr)) - fd.write(' attr = {}\n') - fd.write(' attr["name"] = "{}"\n'.format(attr)) - fd.write(' attr["dtype"] = "{}"\n'.format(self.attr_val.get(attr).get('type'))) - fd.write(' attr["value"] = {}\n'.format(attr)) - fd.write(' __attrs__.append(attr)\n') - fd.write(' return __inputs__, __outputs__, __attrs__\n') - - def _write_impl(self: any, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - pchk = self._build_parachk() - if len(self.kern_name) > 0: - kern_name = self.kern_name - else: - kern_name = self.op_intf - src = self.op_file + '.cpp' - fd.write(IMPL_API.format(self.op_type, pchk, self.op_intf, argsdef, kern_name, argsval,\ - self.custom_compile_options, self.custom_all_compile_options, self.op_intf,\ - optype_snake(self.op_type), src)) - if self.op_replay_flag: - fd.write(REPLAY_OP_API.format(self.op_type, kern_name, self.op_file, self.op_type, self.op_file,\ - self.op_compile_option)) - else: - fd.write(COMPILE_OP_API.format(self.op_type, kern_name, self.op_type, ', '.join(self.input_name),\ - ', '.join(self.output_name), self.op_compile_option)) - - def _write_cap(self: any, cap_name: str, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - if cap_name == 'check_supported': - fd.write(SUP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) - else: - fd.write(CAP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) - - def _write_glz(self: any, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - fd.write(GLZ_API.format(self.op_type, self.op_intf, argsdef, argsval, self.op_type)) - - -def write_scripts(cfgfile: str, cfgs: dict, dirs: dict, ops: list = None, op_compile_option:list = None): - batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') - iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') - file_map = {} - op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, AdpBuilder,\ - ops, dirs.get(const_var.AUTO_GEN_DIR)) - for op_desc in op_descs: - op_desc.write_adapt(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), op_compile_option) - file_map[op_desc.op_type] = op_desc.op_file - return file_map - -if __name__ == '__main__': - if len(sys.argv) <= 6: - raise RuntimeError('arguments must greater equal than 6') - rep_cfg = {} - rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] - rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] - cfg_dir = {} - cfg_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] - cfg_dir[const_var.CFG_OUT_DIR] = sys.argv[5] - cfg_dir[const_var.AUTO_GEN_DIR] = sys.argv[6] - write_scripts(cfgfile=sys.argv[1], cfgs=rep_cfg, dirs=cfg_dir) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py deleted file mode 100755 index 7a97180be..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_ops_config.py +++ /dev/null @@ -1,114 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import glob -import json -import argparse -import const_var - - -def load_json(json_file: str): - with open(json_file, encoding='utf-8') as file: - json_content = json.load(file) - return json_content - - -def get_specified_suffix_file(root_dir, suffix): - specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) - all_suffix_files = glob.glob(specified_suffix, recursive=True) - return all_suffix_files - - -def add_simplified_config(op_type, key, core_type, objfile, config): - simple_cfg = config.get('binary_info_config.json') - op_cfg = simple_cfg.get(op_type) - if not op_cfg: - op_cfg = {} - op_cfg['dynamicRankSupport'] = True - op_cfg['simplifiedKeyMode'] = 0 - op_cfg['binaryList'] = [] - simple_cfg[op_type] = op_cfg - bin_list = op_cfg.get('binaryList') - bin_list.append({'coreType': core_type, 'simplifiedKey': key, 'binPath': objfile}) - - -def add_op_config(op_file, bin_info, config): - op_cfg = config.get(op_file) - if not op_cfg: - op_cfg = {} - op_cfg['binList'] = [] - config[op_file] = op_cfg - op_cfg.get('binList').append(bin_info) - - -def gen_ops_config(json_file, soc, config): - core_type_map = {"MIX": 0, "AiCore": 1, "VectorCore": 2} - contents = load_json(json_file) - if ('binFileName' not in contents) or ('supportInfo' not in contents): - return - json_base_name = os.path.basename(json_file) - op_dir = os.path.basename(os.path.dirname(json_file)) - support_info = contents.get('supportInfo') - bin_name = contents.get('binFileName') - bin_suffix = contents.get('binFileSuffix') - core_type = core_type_map.get(contents.get("coreType")) - bin_file_name = bin_name + bin_suffix - op_type = bin_name.split('_')[0] - op_file = op_dir + '.json' - bin_info = {} - keys = support_info.get('simplifiedKey') - if keys: - bin_info['simplifiedKey'] = keys - for key in keys: - add_simplified_config(op_type, key, core_type, os.path.join(soc, op_dir, bin_file_name), config) - bin_info['staticKey'] = support_info.get('staticKey') - bin_info['int64Mode'] = support_info.get('int64Mode') - bin_info['inputs'] = support_info.get('inputs') - bin_info['outputs'] = support_info.get('outputs') - if support_info.get('attrs'): - bin_info['attrs'] = support_info.get('attrs') - bin_info['binInfo'] = {'jsonFilePath': os.path.join(soc, op_dir, json_base_name)} - add_op_config(op_file, bin_info, config) - - -def gen_all_config(root_dir, soc): - suffix = 'json' - config = {} - config['binary_info_config.json'] = {} - all_json_files = get_specified_suffix_file(root_dir, suffix) - for _json in all_json_files: - gen_ops_config(_json, soc, config) - for cfg_key in config.keys(): - cfg_file = os.path.join(root_dir, cfg_key) - with os.fdopen(os.open(cfg_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - json.dump(config.get(cfg_key), fd, indent=' ') - - -def args_prase(): - parser = argparse.ArgumentParser() - parser.add_argument('-p', - '--path', - nargs='?', - required=True, - help='Parse the path of the json file.') - parser.add_argument('-s', - '--soc', - nargs='?', - required=True, - help='Parse the soc_version of ops.') - return parser.parse_args() - - -def main(): - args = args_prase() - gen_all_config(args.path, args.soc) - - -if __name__ == '__main__': - main() diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py deleted file mode 100755 index 1cac7d911..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/ascendc_replay_build.py +++ /dev/null @@ -1,65 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import opdesc_parser -import replay_codegen -import const_var -from replay_codegen import ReplayCodeGenParams - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - - -class ReplayBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - super().__init__(op_type) - - def gen_replay_source(self: any, impl_path: str, out_path: str, ops_product: str): - if not self.op_replay_flag: - print('{} replay not enabled'.format(self.op_type)) - return - argn = len(self.input_name) + len(self.output_name) + 1 - if self.op_replay_batch: - print('{} replay in batch mode'.format(self.op_type)) - else: - print('{} replay in normal mode'.format(self.op_type)) - if impl_path.endswith('op_kernel'): - implf = os.path.join(impl_path, self.op_file + '.cpp') - tiling_file = os.path.join(impl_path, "../op_host", self.op_file + '_tiling.h') - else: - if self.dynamic_shape: - dyn_path = 'dynamic' - else: - dyn_path = '' - implf = os.path.join(impl_path, dyn_path, self.op_file + '.cpp') - tiling_file = os.path.join(impl_path, "../../op_tiling", self.op_file + '_tiling.h') - rep_conf = replay_codegen.ReplayCodeGen(ReplayCodeGenParams(self.op_type, implf, tiling_file, self.op_file, \ - self.op_intf, argn, self.op_replay_batch, self.max_block_dim, self.max_shape_size)) - rep_conf.set_batch(self.op_replay_batch) - rep_conf.set_outdir(out_path) - rep_conf.gen_replay(ops_product) - - -def gen_replay(cfgfile: str, cfgs: dict, dirs: dict, ops_product: str, ops: list = None): - batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') - iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') - op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, ReplayBuilder, ops) - for op_desc in op_descs: - op_desc.gen_replay_source(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), ops_product) - - -if __name__ == '__main__': - if len(sys.argv) <= 6: - raise RuntimeError('arguments must greater than 6') - rep_cfg = {} - rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] - rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] - rep_dir = {} - rep_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] - rep_dir[const_var.CFG_OUT_DIR] = sys.argv[5] - gen_replay(sys.argv[1], rep_cfg, rep_dir, sys.argv[6]) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp b/atb_operator/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp deleted file mode 100755 index 0e8834664..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/batch_replay_impl.temp +++ /dev/null @@ -1,117 +0,0 @@ -#include -#include -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#include "register/op_check.h" -#define __ASCENDC_REPLAY_CODE__ -#include - -using namespace std; -using namespace optiling; -using namespace AscendCReplay; - -extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); -extern "C" int elf_batch_append(char *elf, uint32_t elfSize, char *jit, int kernum, char *atext[], int alen[], - int atlen, const char* kernelname[]); - -#define KERNEL_N 1 -#define ARG_N (__ARG_NUM__) -#define MAX_L (1024 * 1024 * 100) -#define MAX_E (1024 * 1024) - -int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) -{ - // gen type 1 : direct call codes 0: load .o file - if (param.gentype < 0 || param.gentype > 1) { - printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); - return 0; - } else if (param.gentype == 1 && param.objptr == nullptr) { - printf("Error: call replay with direct call mode, but code obj addr is null\n"); - return 0; - } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { - printf("Error: call replay with object file mode, but object file path is null\n"); - return 0; - } - // core_type 0:MIX 1:CUBE 2:VEC - if (core_type < 0 || core_type > 2) { - printf("Error: call replay core type is %d !\n", core_type); - return 0; - } - g_coreType = __CORE_TYPE__; - g_taskRation = param.task_ration; - g_tilingKey = param.tiling_key; - - unsigned char *buf, *jit; - char *kernel[KERNEL_N]; - int len[KERNEL_N]; - block_idx = 0; - block_num = param.block_dim; - g_ubBase = block_num; - uint8_t *code = (uint8_t *)malloc(MAX_L); - uint8_t *pos = code; - struct timespec tp1, tp2; - - clock_gettime(CLOCK_MONOTONIC, &tp1); - if (block_num > 32) { - printf("Error: block_num > 32\n"); - return 0; - } - //__OP_FOPEN__ - for (int i = 0; i < KERNEL_N; i++) { - //__OP_SET_KERNEL__ - for (int j = 0; j < ARG_N; j++) - AddArg(j, ARG_STEP * (j + 1)); -#ifdef FP_CEILING - SetCtrlFloatEnable(); -#else - SetCtrlFloatDisable(); -#endif - CodeInit(pos, true); - __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); - CodeEnd(); - kernel[i] = (char *)pos; - len[i] = CodeLen(); - pos += len[i]; - } - //__OP_FCLOSE__ - clock_gettime(CLOCK_MONOTONIC, &tp2); - buf = (unsigned char *)malloc(MAX_E); - int fd = open(param.entry_file, O_RDONLY); - if (fd < 0) { - printf("[error]: cannot find entry.o : %s\n", param.entry_file); - return 0; - } - uint32_t bufSize = read(fd, buf, MAX_E); - if (bufSize <= 0) { - printf("[error]: entry.o : %s is too small ! \n", param.entry_file); - } - close(fd); - jit = (unsigned char *)malloc(MAX_L); - printf("total code generated %ld\n", pos - code); - int sz = elf_batch_append((char *)buf, bufSize, (char *)jit, KERNEL_N, kernel, len, pos - code, ¶m.kernel_name); - if (tp1.tv_sec != tp2.tv_sec) { - printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); - } else { - printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); - } - printf("new elf size %d\n", sz); - if (param.gentype == 0) { - fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); - (void)write(fd, jit, sz); - close(fd); - free(jit); - } else if (param.gentype == 1) { - *param.objptr = (char*)jit; - } - free(buf); - free(code); - return sz; -} - -REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py deleted file mode 100755 index 137c9f39f..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/code_channel_infer.py +++ /dev/null @@ -1,115 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" -import os -import stat -import ctypes -import collections -import shutil -import subprocess -import copy - -"""CODE_* is used to cube/vector api is called in operator code -CODE_MIX means both cube and vector api is called -CODE_CUBE means only cube api is called -CODE_VEC means only vector api is called -""" -CODE_MIX = 0 -CODE_CUBE = 1 -CODE_VEC = 2 - - -def _is_v220(op_product: str): - """return if current soc version is V220 - - Returns: - res: True means V220 - """ - if op_product in ["ascend910b"]: - return True - return False - - -InfoCodeChanelParams = collections.namedtuple('InfoCodeChanelParams',\ -['src_file', 'tiling_header', 'kernel_name', 'outdir', 'op_product', 'compile_options']) - - -def infer_code_channel(params: InfoCodeChanelParams): - """get code channel for v220, return CODE_MIX if soc version is not V220 - - Args: - src_file (str): AscendC operator code file - src_file (str): AscendC operator tiling header file - kernel_name (str): kernel function name - optype (str): operator type - compile_options (list): compile options for ccec cmd - - Raises: - Exception: if not exist L1/L0/UB if code, it's not a aicore code - - Returns: - res (int): CODE_MIX/CODE_CUBE/CODE_VEC - """ - if not _is_v220(params.op_product): - return CODE_MIX - return CODE_VEC - if params.compile_options is None: - compile_options = [] - else: - compile_options = params.compile_options - ccec = shutil.which("ccec") - if ccec is not None: - ccec_path = os.path.dirname(ccec) - tikcpp_path = os.path.realpath(os.path.join(ccec_path, "..", "..", "tikcpp")) - else: - tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") - compile_options.append("-I" + tikcpp_path) - compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) - compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) - compile_options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) - compile_options += ["-include", params.tiling_header] - arch = "dav-c220-cube" - sub_core_type = "AIC" - optional_core = "AiCore" - compile_cmd = [shutil.which("ccec"), '-c', '-O3'] - compile_cmd += compile_options - temp_file_name_tag = "_" + str(os.getpid()) + "_temp.o" - dst_file = os.path.join(kernel_meta_dir, kernel_name + temp_file_name_tag) - compile_cmd += [params.src_file, "--cce-aicore-arch={}".format(arch), - "--cce-aicore-only", "-o", dst_file, - "-mllvm", "-cce-aicore-function-stack-size=16000", - "-mllvm", "-cce-aicore-record-overflow=true", - "-mllvm", "-cce-aicore-addr-transform"] - compile_cmd += ["-std=c++17"] - print('get_code_channel: ', ' '.join(compile_cmd)) - proc = subprocess.Popen( - compile_cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) - (out, _) = proc.communicate() - if proc.returncode != 0: - print('get_code_channel coretype compile error: ', out.decode()) - msg = "compile %s error :%s\n" % (params.src_file, out.decode()) - raise Exception(f"get_code_channel coretype error, msg is{msg}") - objdump_cmd = ['objdump', '-s', '-j', '.text', '{}'.format(dst_file)] - - proc = subprocess.Popen( - objdump_cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) - (out, _) = proc.communicate() - if proc.returncode != 0: - print('get_code_channel objdump error: ', out.decode()) - msg = "get_code_channel objdump %s error :%s\n" % (src_file, out.decode()) - raise Exception(f"get_code_channel objdump error, msg is{msg}") - os.remove(dst_file) - lines = out.decode('utf-8').split('\n') - for line in lines: - insts = line.strip().split() - if len(insts) < 5: - continue - for inst in insts[1:]: - if len(inst) != 8: - continue - if inst[6] == 'f': - return CODE_MIX - return CODE_VEC diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/const_var.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/const_var.py deleted file mode 100755 index 8b32c3b91..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/const_var.py +++ /dev/null @@ -1,33 +0,0 @@ - -#!/usr/bin/env python -# coding=utf-8 -""" -Function: -The replay funtion entry -Copyright Information: -Huawei Technologies Co., Ltd. All Rights Reserved © 2020 -""" - -import os -import stat - - -REPLAY_BATCH = 'batch' -REPLAY_ITERATE = 'iterate' -CFG_IMPL_DIR = 'impl_dir' -CFG_OUT_DIR = 'out_dir' -AUTO_GEN_DIR = 'auto_gen_dir' -WFLAGS = os.O_WRONLY | os.O_CREAT | os.O_TRUNC -WMODES = stat.S_IWUSR | stat.S_IRUSR -SOC_MAP_EXT = {'ascend310p': 'Ascend310P3', 'ascend310b': 'Ascend310B1', - 'ascend910': 'Ascend910A', 'ascend910b': 'Ascend910B1'} -BIN_CMD = 'opc $1 --main_func={fun} --input_param={param} --soc_version={soc} \ ---output=$2 --impl_mode={impl} --simplified_key_mode=0 --op_mode=dynamic\n' -CHK_CMD = ''' -if ! test -f $2/{res_file} ; then - echo "$2/{res_file} not generated!" - exit 1 -fi -''' -ATTR_DEF_VAL = {'str' : '', 'int': 0, 'float': 0.0, 'bool': False, 'list_bool': [], - 'list_int': [], 'list_float': [], 'list_list_int': [[]]} diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh deleted file mode 100755 index 55e12e5ed..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_impl_and_mrege_json.sh +++ /dev/null @@ -1,57 +0,0 @@ -#!/usr/bin/bash - -project_path=$1 -build_path=$2 -vendor_name=customize -if [[ ! -d "$project_path" ]]; then - echo "[ERROR] No projcet path is provided" - exit 1 -fi - -if [[ ! -d "$build_path" ]]; then - echo "[ERROR] No build path is provided" - exit 1 -fi - -# copy ai_core operators implements -# tbe_impl_files_num=$(ls $project_path/tbe/impl/* 2> /dev/null | wc -l) -# if [[ "$tbe_impl_files_num" -gt 0 ]];then -# cp -r ${project_path}/tbe/impl/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/ai_core/tbe/customize_impl -# cp -r ${project_path}/tbe/impl/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/vector_core/tbe/customize_impl -# fi - -# copy aicpu kernel so operators -if [[ -d "${project_path}/cpukernel/aicpu_kernel_lib" ]]; then - cp -f ${project_path}/cpukernel/aicpu_kernel_lib/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/cpu/aicpu_kernel/impl - rm -rf ${project_path}/cpukernel/aicpu_kernel_lib -fi - -# merge aicpu.ini and aicore.ini to generate npu_supported_ops.json -# mkdir -p ${build_path}/framework/op_info_cfg -# mkdir -p ${build_path}/framework/op_info_cfg/aicpu_kernel -# mkdir -p ${build_path}/framework/op_info_cfg/ai_core - -# if [[ -d "${project_path}/tbe/op_info_cfg/ai_core" ]]; then -# bash ${project_path}/cmake/util/gen_ops_filter.sh ${project_path}/tbe/op_info_cfg/ai_core ${build_path}/framework/op_info_cfg/ai_core -# fi - -# if [[ -d "${project_path}/cpukernel/op_info_cfg/aicpu_kernel" ]]; then -# bash ${project_path}/cmake/util/gen_ops_filter.sh ${project_path}/cpukernel/op_info_cfg/aicpu_kernel ${build_path}/framework/op_info_cfg/aicpu_kernel -# fi - -# aicpu_filter_file=${build_path}/framework/op_info_cfg/aicpu_kernel/npu_supported_ops.json -# aicore_filter_file=${build_path}/framework/op_info_cfg/ai_core/npu_supported_ops.json -# if [[ -f "${aicpu_filter_file}" ]] && [[ ! -f "${aicore_filter_file}" ]]; then -# cp $aicpu_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow -# fi -# if [[ -f "${aicore_filter_file}" ]] && [[ ! -f "${aicpu_filter_file}" ]]; then -# cp $aicore_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow -# fi - -# if [[ -f "${aicore_filter_file}" ]] && [[ -f "${aicpu_filter_file}" ]]; then -# chmod u+w ${aicpu_filter_file} -# python3 ${project_path}/cmake/util/insert_op_info.py ${aicore_filter_file} ${aicpu_filter_file} -# chmod u-w ${aicpu_filter_file} -# cp $aicpu_filter_file ${build_path}/makepkg/packages/vendors/$vendor_name/framework/tensorflow -# fi - diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh deleted file mode 100755 index d4c27d17f..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_ops_filter.sh +++ /dev/null @@ -1,62 +0,0 @@ -#!/bin/bash -# Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -# Description: Generate npu_supported_ops.json -# ============================================================================== - -if [[ -z "$1" ]]; then - echo -e "[ERROR] No source dir provided" - exit 1 -fi - -if [[ -z "$2" ]]; then - echo -e "[ERROR] No destination dir provided" - exit 1 -fi - -src=$1 -dest_file=$2/npu_supported_ops.json - -if [ -f "$dest_file" ];then - chmod u+w $dest_file -fi - -echo $* - -add_ops() { - name=$1 - isHeavy=$2 - file=$3 - grep -w "\"$name\"" ${file} >/dev/null - if [ $? == 0 ];then - return - fi - echo " \"${name}\": {" >> ${file} - echo " \"isGray\": false," >> ${file} - echo " \"isHeavy\": ${isHeavy}" >> ${file} - echo " }," >> ${file} -} - -echo "{" > ${dest_file} -ini_files=$(find ${src} -name "*.ini") -for file in ${ini_files} ; do - name=$(grep '^\[' ${file} | sed 's/\[//g' | sed 's/]//g' | sed 's/\r//g') - grep 'heavyOp.flag' ${file} >/dev/null - if [ $? == 0 ];then - isHeavy=$(grep 'heavyOp.flag' ${file} | awk -F= '{print $2}') - else - isHeavy="false" - fi - for op in ${name} ; do - add_ops ${op} "false" ${dest_file} - done -done -echo "}" >> ${dest_file} -file_count=$(cat ${dest_file} | wc -l) -line=$(($file_count-1)) -sed -i "${line}{s/,//g}" ${dest_file} - -chmod 640 "${dest_file}" -echo -e "[INFO] Succed generated ${dest_file}" - -exit 0 - diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh deleted file mode 100755 index a06cfc78d..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/gen_version_info.sh +++ /dev/null @@ -1,6 +0,0 @@ -ascend_install_dir=$1 -gen_file_dir=$2 - -# create version.info -compiler_version=$(grep "Version" -w ${ascend_install_dir}/compiler/version.info | awk -F = '{print $2}') -echo "custom_opp_compiler_version=${compiler_version}" > ${gen_file_dir}/version.info \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py deleted file mode 100755 index 28ba08757..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_op_info.py +++ /dev/null @@ -1,36 +0,0 @@ -# -*- coding: utf-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" -import json -import os -import sys -import stat -import const_var - - -if __name__ == '__main__': - if len(sys.argv) != 3: - print(sys.argv) - print('argv error, inert_op_info.py your_op_file lib_op_file') - sys.exit(2) - - with open(sys.argv[1], 'r') as load_f: - insert_operator = json.load(load_f) - - all_operators = {} - if os.path.exists(sys.argv[2]): - if os.path.getsize(sys.argv[2]) != 0: - with open(sys.argv[2], 'r') as load_f: - all_operators = json.load(load_f) - - for k in insert_operator.keys(): - if k in all_operators.keys(): - print('replace op:[', k, '] success') - else: - print('insert op:[', k, '] success') - all_operators[k] = insert_operator[k] - - with os.fdopen(os.open(sys.argv[2], const_var.WFLAGS, const_var.WMODES), 'w') as json_file: - json_file.write(json.dumps(all_operators, indent=4)) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py deleted file mode 100755 index ace727b90..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/insert_simplified_keys.py +++ /dev/null @@ -1,248 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import re -import glob -import json -import argparse -import const_var - - -DATA_TPYE_DICT = { - 'float32': 0, - 'float16': 1, - 'int8': 2, - 'int16': 6, - 'uint16': 7, - 'uint8': 4, - 'int32': 3, - 'int64': 9, - 'uint32': 8, - 'uint64': 10, - 'bool': 12, - 'double': 11, - 'string': 13, - 'dual': 14, - 'dual': 15, - 'complex64': 16, - 'complex128': 17, - 'qint8': 18, - 'qint16': 19, - 'qint32': 20, - 'quint8': 21, - 'quint16': 22, - 'resource': 23, - 'string': 24, - 'dual': 25, - 'variant': 26, - 'bf16': 27, - 'bfloat16': 27, - 'undefined': 28, - 'int4': 29, - 'uint1': 30, - 'int2': 31 -} - -FORMAT_DICT = { - 'NCHW': 0, - 'NHWC': 1, - 'ND': 2, - 'NC1HWC0': 3, - 'FRACTAL_Z': 4, - 'NC1C0HWPAD': 5, - 'NHWC1C0': 6, - 'FSR_NCHW': 7, - 'FRACTAL_DECONV': 8, - 'C1HWNC0': 9, - 'FRACTAL_DECONV_TRANSPOSE': 10, - 'FRACTAL_DECONV_SP_STRIDE_TRANS': 11, - 'NC1HWC0_C04': 12, - 'FRACTAL_Z_C04': 13, - 'CHWN': 14, - 'FRACTAL_DECONV_SP_STRIDE8_TRANS': 15, - 'HWCN': 16, - 'NC1KHKWHWC0': 17, - 'BN_WEIGHT': 18, - 'FILTER_HWCK': 19, - 'HASHTABLE_LOOKUP_LOOKUPS': 20, - 'HASHTABLE_LOOKUP_KEYS': 21, - 'HASHTABLE_LOOKUP_VALUE': 22, - 'HASHTABLE_LOOKUP_OUTPUT': 23, - 'HASHTABLE_LOOKUP_HITS': 24, - 'C1HWNCoC0': 25, - 'MD': 26, - 'NDHWC': 27, - 'FRACTAL_ZZ': 28, - 'FRACTAL_NZ': 29, - 'NCDHW': 30, - 'DHWCN': 31, - 'NDC1HWC0': 32, - 'FRACTAL_Z_3D': 33, - 'CN': 34, - 'NC': 35, - 'DHWNC': 36, - 'FRACTAL_Z_3D_TRANSPOSE': 37, - 'FRACTAL_ZN_LSTM': 38, - 'FRACTAL_Z_G': 39, - 'RESERVED': 40, - 'ALL': 41, - 'NULL': 42, - 'ND_RNN_BIAS': 43, - 'FRACTAL_ZN_RNN': 44, - 'NYUV': 45, - 'NYUV_A': 46 -} - - -def load_json(json_file: str): - with open(json_file, encoding='utf-8') as file: - json_content = json.load(file) - return json_content - - -def get_specified_suffix_file(root_dir, suffix): - specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) - all_suffix_files = glob.glob(specified_suffix, recursive=True) - return all_suffix_files - - -def get_deterministic_value(support_info): - deterministic_key = 'deterministic' - if deterministic_key not in support_info: - return 0 - deterministic_value = support_info.get(deterministic_key) - if deterministic_value == 'true': - return 1 - else: - return 0 - - -def get_precision_value(support_info): - precision_key = 'implMode' - precision_value = support_info.get(precision_key) - if precision_value == 'high_performance': - _value = 1 - elif precision_value == 'high_precision': - _value = 2 - else: - _value = 0 - return _value - - -def get_overflow_value(support_info): - return 0 - - -def get_parameters(info): - if info: - if 'dtype' in info: - data_type = info['dtype'] - data_type_value = DATA_TPYE_DICT.get(data_type) - else: - data_type_value = 0 - if 'format' in info: - _format = info['format'] - _format_value = FORMAT_DICT.get(_format) - else: - _format_value = 0 - else: - data_type_value = 0 - _format_value = 0 - return str(data_type_value), str(_format_value) - - -def get_dynamic_parameters(info): - # 动态输入时只需获取第一个参数 - return get_parameters(info[0]) - - -def get_all_parameters(support_info, _type): - result_list = list() - info_lists = support_info.get(_type) - if info_lists: - for _info in info_lists: - # 输入为列表时是动态输入 - if isinstance(_info, (list, tuple)): - data_type_value, _format_value = get_dynamic_parameters(_info) - else: - data_type_value, _format_value = get_parameters(_info) - result_list.append("{},{}".format(data_type_value, _format_value)) - return result_list - - -def get_all_input_parameters(support_info): - result = get_all_parameters(support_info, 'inputs') - return '/'.join(result) - - -def insert_content_into_file(input_file, content): - with open(input_file, 'r+') as file: - lines = file.readlines() - for index, line in enumerate(lines): - match_result = re.search(r'"staticKey":', line) - if match_result: - count = len(line) - len(line.lstrip()) - new_content = "{}{}".format(' ' * count, content) - # 插入到前一行,防止插入最后时还需要考虑是否添加逗号 - lines.insert(index, new_content) - break - file.seek(0) - file.write(''.join(lines)) - - -def insert_simplified_keys(json_file): - contents = load_json(json_file) - # 不存在'binFileName'或者'supportInfo'字段时,非需要替换的解析json文件 - if ('binFileName' not in contents) or ('supportInfo' not in contents): - return - support_info = contents.get('supportInfo') - bin_file_name = contents.get('binFileName') - bin_suffix = contents.get('binFileSuffix') - # 'simplifiedKey'字段已经存在时,直接返回,不重复生成 - if 'simplifiedKey' in support_info: - return - op_type = bin_file_name.split('_')[0] - deterministic = str(get_deterministic_value(support_info)) - precision = str(get_precision_value(support_info)) - overflow = str(get_overflow_value(support_info)) - input_parameters = get_all_input_parameters(support_info) - key = '{}/d={},p={},o={}/{}/'.format( - op_type, - deterministic, - precision, - overflow, - input_parameters) - result = '"simplifiedKey": "' + key + '",\n' - insert_content_into_file(json_file, result) - - -def insert_all_simplified_keys(root_dir): - suffix = 'json' - all_json_files = get_specified_suffix_file(root_dir, suffix) - for _json in all_json_files: - insert_simplified_keys(_json) - - -def args_prase(): - parser = argparse.ArgumentParser() - parser.add_argument('-p', - '--path', - nargs='?', - required=True, - help='Parse the path of the json file.') - return parser.parse_args() - - -def main(): - args = args_prase() - insert_all_simplified_keys(args.path) - - -if __name__ == '__main__': - main() diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py deleted file mode 100755 index 2b77c970d..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_entry.py +++ /dev/null @@ -1,115 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - - -def gen_fun_def(title, kernel, argn, arg_type, arg_name): - entry = [] - entry.append(title) - entry.append(kernel) - entry.append('(') - args = [] - for i in range(0, argn): - args.append(arg_type + ' ' + arg_name + str(i)) - entry.append(', '.join(args)) - entry.append(')') - return ' '.join(entry) - - -def gen_batch_kernel_body(fname, argn, arg_name): - body = [] - body.append('{') - fun = [] - fun.append(fname) - fun.append('(') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - fun.append(', '.join(args)) - fun.append(');') - body.append(' '.join(fun)) - body.append('}') - return '\n'.join(body) - - -def gen_mc_kernel_body(kn, argn, arg_name, blknum): - body = [] - body.append('{') - body.append(' switch(block_idx) {') - for blk in range(0, blknum): - fun = [] - fun.append('{}_blk{:02d}'.format(kn, blk)) - fun.append('(') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - fun.append(', '.join(args)) - fun.append(')') - body.append(' case {}: {}; break;'.format(blk, ' '.join(fun))) - body.append(' default: break;') - body.append(' }') - body.append('}') - return '\n'.join(body) - - -def gen_proc_body(argn, arg_name): - body = [] - body.append('{') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - body.append('uint64_t __x = (uint64_t)' + ' + (uint64_t)'.join(args) + ';') - body.append('__asm__ ("NOP");') - body.append('__asm__ ("NOP");') - body.append('__asm__ ("NOP");') - body.append('}') - return '\n'.join(body) - - -def batch_code_gen(kn, argn, argt): - codes = [] - kernel_name = kn - proc_name = kernel_name + '_percore' - arg_num = int(argn) - data_type = argt - arg_type = '__gm__ ' + data_type + '* __restrict__' - arg_name = 'arg' - kernel_title = 'extern \"C\" __global__ __aicore__ void' - proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' - codes.append('#ifndef __aicore__') - codes.append('#define __aicore__ [aicore]') - codes.append('#endif') - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') - codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) - codes.append(gen_batch_kernel_body(proc_name, arg_num, arg_name)) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) - codes.append(gen_proc_body(arg_num, arg_name)) - return '\n'.join(codes) + '\n' - - -def mc_code_gen(kn, argn, argt, blknum): - codes = [] - kernel_name = kn - core_num = int(blknum) - arg_num = int(argn) - data_type = argt - arg_type = '__gm__ ' + data_type + '* __restrict__' - arg_name = 'arg' - kernel_title = 'extern \"C\" __global__ __aicore__ void' - proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' - codes.append('#ifndef __aicore__') - codes.append('#define __aicore__ [aicore]') - codes.append('#endif') - for i in range(0, core_num): - proc_name = '{}_blk{:02d}'.format(kernel_name, i) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') - codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) - codes.append(gen_mc_kernel_body(kernel_name, arg_num, arg_name, core_num)) - for i in range(0, core_num): - proc_name = '{}_blk{:02d}'.format(kernel_name, i) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) - codes.append(gen_proc_body(arg_num, arg_name)) - return '\n'.join(codes) + '\n' diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp b/atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp deleted file mode 100755 index 5079a1043..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/kernel_impl.temp +++ /dev/null @@ -1,10 +0,0 @@ -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#define __ASCENDC_REPLAY_CODE__ -#include "__CCE_FILE__" diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING deleted file mode 100644 index d159169d1..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/COPYING +++ /dev/null @@ -1,339 +0,0 @@ - GNU GENERAL PUBLIC LICENSE - Version 2, June 1991 - - Copyright (C) 1989, 1991 Free Software Foundation, Inc., - 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA - Everyone is permitted to copy and distribute verbatim copies - of this license document, but changing it is not allowed. - - Preamble - - The licenses for most software are designed to take away your -freedom to share and change it. By contrast, the GNU General Public -License is intended to guarantee your freedom to share and change free -software--to make sure the software is free for all its users. This -General Public License applies to most of the Free Software -Foundation's software and to any other program whose authors commit to -using it. (Some other Free Software Foundation software is covered by -the GNU Lesser General Public License instead.) You can apply it to -your programs, too. - - When we speak of free software, we are referring to freedom, not -price. Our General Public Licenses are designed to make sure that you -have the freedom to distribute copies of free software (and charge for -this service if you wish), that you receive source code or can get it -if you want it, that you can change the software or use pieces of it -in new free programs; and that you know you can do these things. - - To protect your rights, we need to make restrictions that forbid -anyone to deny you these rights or to ask you to surrender the rights. -These restrictions translate to certain responsibilities for you if you -distribute copies of the software, or if you modify it. - - For example, if you distribute copies of such a program, whether -gratis or for a fee, you must give the recipients all the rights that -you have. You must make sure that they, too, receive or can get the -source code. And you must show them these terms so they know their -rights. - - We protect your rights with two steps: (1) copyright the software, and -(2) offer you this license which gives you legal permission to copy, -distribute and/or modify the software. - - Also, for each author's protection and ours, we want to make certain -that everyone understands that there is no warranty for this free -software. If the software is modified by someone else and passed on, we -want its recipients to know that what they have is not the original, so -that any problems introduced by others will not reflect on the original -authors' reputations. - - Finally, any free program is threatened constantly by software -patents. We wish to avoid the danger that redistributors of a free -program will individually obtain patent licenses, in effect making the -program proprietary. To prevent this, we have made it clear that any -patent must be licensed for everyone's free use or not licensed at all. - - The precise terms and conditions for copying, distribution and -modification follow. - - GNU GENERAL PUBLIC LICENSE - TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION - - 0. This License applies to any program or other work which contains -a notice placed by the copyright holder saying it may be distributed -under the terms of this General Public License. The "Program", below, -refers to any such program or work, and a "work based on the Program" -means either the Program or any derivative work under copyright law: -that is to say, a work containing the Program or a portion of it, -either verbatim or with modifications and/or translated into another -language. (Hereinafter, translation is included without limitation in -the term "modification".) Each licensee is addressed as "you". - -Activities other than copying, distribution and modification are not -covered by this License; they are outside its scope. The act of -running the Program is not restricted, and the output from the Program -is covered only if its contents constitute a work based on the -Program (independent of having been made by running the Program). -Whether that is true depends on what the Program does. - - 1. You may copy and distribute verbatim copies of the Program's -source code as you receive it, in any medium, provided that you -conspicuously and appropriately publish on each copy an appropriate -copyright notice and disclaimer of warranty; keep intact all the -notices that refer to this License and to the absence of any warranty; -and give any other recipients of the Program a copy of this License -along with the Program. - -You may charge a fee for the physical act of transferring a copy, and -you may at your option offer warranty protection in exchange for a fee. - - 2. You may modify your copy or copies of the Program or any portion -of it, thus forming a work based on the Program, and copy and -distribute such modifications or work under the terms of Section 1 -above, provided that you also meet all of these conditions: - - a) You must cause the modified files to carry prominent notices - stating that you changed the files and the date of any change. - - b) You must cause any work that you distribute or publish, that in - whole or in part contains or is derived from the Program or any - part thereof, to be licensed as a whole at no charge to all third - parties under the terms of this License. - - c) If the modified program normally reads commands interactively - when run, you must cause it, when started running for such - interactive use in the most ordinary way, to print or display an - announcement including an appropriate copyright notice and a - notice that there is no warranty (or else, saying that you provide - a warranty) and that users may redistribute the program under - these conditions, and telling the user how to view a copy of this - License. (Exception: if the Program itself is interactive but - does not normally print such an announcement, your work based on - the Program is not required to print an announcement.) - -These requirements apply to the modified work as a whole. If -identifiable sections of that work are not derived from the Program, -and can be reasonably considered independent and separate works in -themselves, then this License, and its terms, do not apply to those -sections when you distribute them as separate works. But when you -distribute the same sections as part of a whole which is a work based -on the Program, the distribution of the whole must be on the terms of -this License, whose permissions for other licensees extend to the -entire whole, and thus to each and every part regardless of who wrote it. - -Thus, it is not the intent of this section to claim rights or contest -your rights to work written entirely by you; rather, the intent is to -exercise the right to control the distribution of derivative or -collective works based on the Program. - -In addition, mere aggregation of another work not based on the Program -with the Program (or with a work based on the Program) on a volume of -a storage or distribution medium does not bring the other work under -the scope of this License. - - 3. You may copy and distribute the Program (or a work based on it, -under Section 2) in object code or executable form under the terms of -Sections 1 and 2 above provided that you also do one of the following: - - a) Accompany it with the complete corresponding machine-readable - source code, which must be distributed under the terms of Sections - 1 and 2 above on a medium customarily used for software interchange; or, - - b) Accompany it with a written offer, valid for at least three - years, to give any third party, for a charge no more than your - cost of physically performing source distribution, a complete - machine-readable copy of the corresponding source code, to be - distributed under the terms of Sections 1 and 2 above on a medium - customarily used for software interchange; or, - - c) Accompany it with the information you received as to the offer - to distribute corresponding source code. (This alternative is - allowed only for noncommercial distribution and only if you - received the program in object code or executable form with such - an offer, in accord with Subsection b above.) - -The source code for a work means the preferred form of the work for -making modifications to it. For an executable work, complete source -code means all the source code for all modules it contains, plus any -associated interface definition files, plus the scripts used to -control compilation and installation of the executable. However, as a -special exception, the source code distributed need not include -anything that is normally distributed (in either source or binary -form) with the major components (compiler, kernel, and so on) of the -operating system on which the executable runs, unless that component -itself accompanies the executable. - -If distribution of executable or object code is made by offering -access to copy from a designated place, then offering equivalent -access to copy the source code from the same place counts as -distribution of the source code, even though third parties are not -compelled to copy the source along with the object code. - - 4. You may not copy, modify, sublicense, or distribute the Program -except as expressly provided under this License. Any attempt -otherwise to copy, modify, sublicense or distribute the Program is -void, and will automatically terminate your rights under this License. -However, parties who have received copies, or rights, from you under -this License will not have their licenses terminated so long as such -parties remain in full compliance. - - 5. You are not required to accept this License, since you have not -signed it. However, nothing else grants you permission to modify or -distribute the Program or its derivative works. These actions are -prohibited by law if you do not accept this License. Therefore, by -modifying or distributing the Program (or any work based on the -Program), you indicate your acceptance of this License to do so, and -all its terms and conditions for copying, distributing or modifying -the Program or works based on it. - - 6. Each time you redistribute the Program (or any work based on the -Program), the recipient automatically receives a license from the -original licensor to copy, distribute or modify the Program subject to -these terms and conditions. You may not impose any further -restrictions on the recipients' exercise of the rights granted herein. -You are not responsible for enforcing compliance by third parties to -this License. - - 7. If, as a consequence of a court judgment or allegation of patent -infringement or for any other reason (not limited to patent issues), -conditions are imposed on you (whether by court order, agreement or -otherwise) that contradict the conditions of this License, they do not -excuse you from the conditions of this License. If you cannot -distribute so as to satisfy simultaneously your obligations under this -License and any other pertinent obligations, then as a consequence you -may not distribute the Program at all. For example, if a patent -license would not permit royalty-free redistribution of the Program by -all those who receive copies directly or indirectly through you, then -the only way you could satisfy both it and this License would be to -refrain entirely from distribution of the Program. - -If any portion of this section is held invalid or unenforceable under -any particular circumstance, the balance of the section is intended to -apply and the section as a whole is intended to apply in other -circumstances. - -It is not the purpose of this section to induce you to infringe any -patents or other property right claims or to contest validity of any -such claims; this section has the sole purpose of protecting the -integrity of the free software distribution system, which is -implemented by public license practices. Many people have made -generous contributions to the wide range of software distributed -through that system in reliance on consistent application of that -system; it is up to the author/donor to decide if he or she is willing -to distribute software through any other system and a licensee cannot -impose that choice. - -This section is intended to make thoroughly clear what is believed to -be a consequence of the rest of this License. - - 8. If the distribution and/or use of the Program is restricted in -certain countries either by patents or by copyrighted interfaces, the -original copyright holder who places the Program under this License -may add an explicit geographical distribution limitation excluding -those countries, so that distribution is permitted only in or among -countries not thus excluded. In such case, this License incorporates -the limitation as if written in the body of this License. - - 9. The Free Software Foundation may publish revised and/or new versions -of the General Public License from time to time. Such new versions will -be similar in spirit to the present version, but may differ in detail to -address new problems or concerns. - -Each version is given a distinguishing version number. If the Program -specifies a version number of this License which applies to it and "any -later version", you have the option of following the terms and conditions -either of that version or of any later version published by the Free -Software Foundation. If the Program does not specify a version number of -this License, you may choose any version ever published by the Free Software -Foundation. - - 10. If you wish to incorporate parts of the Program into other free -programs whose distribution conditions are different, write to the author -to ask for permission. For software which is copyrighted by the Free -Software Foundation, write to the Free Software Foundation; we sometimes -make exceptions for this. Our decision will be guided by the two goals -of preserving the free status of all derivatives of our free software and -of promoting the sharing and reuse of software generally. - - NO WARRANTY - - 11. BECAUSE THE PROGRAM IS LICENSED FREE OF CHARGE, THERE IS NO WARRANTY -FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN -OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES -PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED -OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF -MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS -TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE -PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, -REPAIR OR CORRECTION. - - 12. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING -WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MAY MODIFY AND/OR -REDISTRIBUTE THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, -INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING -OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED -TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY -YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER -PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE -POSSIBILITY OF SUCH DAMAGES. - - END OF TERMS AND CONDITIONS - - How to Apply These Terms to Your New Programs - - If you develop a new program, and you want it to be of the greatest -possible use to the public, the best way to achieve this is to make it -free software which everyone can redistribute and change under these terms. - - To do so, attach the following notices to the program. It is safest -to attach them to the start of each source file to most effectively -convey the exclusion of warranty; and each file should have at least -the "copyright" line and a pointer to where the full notice is found. - - - Copyright (C) - - This program is free software; you can redistribute it and/or modify - it under the terms of the GNU General Public License as published by - the Free Software Foundation; either version 2 of the License, or - (at your option) any later version. - - This program is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the - GNU General Public License for more details. - - You should have received a copy of the GNU General Public License along - with this program; if not, write to the Free Software Foundation, Inc., - 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. - -Also add information on how to contact you by electronic and paper mail. - -If the program is interactive, make it output a short notice like this -when it starts in an interactive mode: - - Gnomovision version 69, Copyright (C) year name of author - Gnomovision comes with ABSOLUTELY NO WARRANTY; for details type `show w'. - This is free software, and you are welcome to redistribute it - under certain conditions; type `show c' for details. - -The hypothetical commands `show w' and `show c' should show the appropriate -parts of the General Public License. Of course, the commands you use may -be called something other than `show w' and `show c'; they could even be -mouse-clicks or menu items--whatever suits your program. - -You should also get your employer (if you work as a programmer) or your -school, if any, to sign a "copyright disclaimer" for the program, if -necessary. Here is a sample; alter the names: - - Yoyodyne, Inc., hereby disclaims all copyright interest in the program - `Gnomovision' (which makes passes at compilers) written by James Hacker. - - , 1 April 1989 - Ty Coon, President of Vice - -This General Public License does not permit incorporating your program into -proprietary programs. If your program is a subroutine library, you may -consider it more useful to permit linking proprietary applications with the -library. If this is what you want to do, use the GNU Lesser General -Public License instead of this License. diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/README.md b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/README.md deleted file mode 100644 index b41f01682..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/README.md +++ /dev/null @@ -1,246 +0,0 @@ -[![License: GPL v2](https://img.shields.io/badge/License-GPL%20v2-blue.svg)](https://www.gnu.org/licenses/old-licenses/gpl-2.0.en.html) -![Build Status](https://github.com/megastep/makeself/workflows/CI/badge.svg) - -# makeself - Make self-extractable archives on Unix - -[makeself.sh][1] is a small shell script that generates a self-extractable -compressed tar archive from a directory. The resulting file appears as a shell script -(many of those have a **.run** suffix), and can be launched as is. The archive -will then uncompress itself to a temporary directory and an optional arbitrary -command will be executed (for example an installation script). This is pretty -similar to archives generated with WinZip Self-Extractor in the Windows world. -Makeself archives also include checksums for integrity self-validation (CRC -and/or MD5/SHA256 checksums). - -The makeself.sh script itself is used only to create the archives from a -directory of files. The resultant archive is actually a compressed (using -gzip, bzip2, or compress) TAR archive, with a small shell script stub at the -beginning. This small stub performs all the steps of extracting the files, -running the embedded command, and removing the temporary files when done. -All the user has to do to install the software contained in such an -archive is to "run" the archive, i.e **sh nice-software.run**. I recommend -using the ".run" (which was introduced by some Makeself archives released by -Loki Software) or ".sh" suffix for such archives not to confuse the users, -so that they will know they are actually shell scripts (with quite a lot of binary data -attached to them though!). - -I am trying to keep the code of this script as portable as possible, i.e it is -not relying on any bash-specific features and only calls commands that are -installed on any functioning UNIX-compatible system. This script as well as -the archives it generates should run on any Unix flavor, with any compatible -Bourne shell, provided of course that the compression programs are available. - -As of version 2.1, Makeself has been rewritten and tested on the following -platforms : - - * Linux (all distributions) - * Sun Solaris (8 and above) - * HP-UX (tested on 11.0 and 11i on HPPA RISC) - * SCO OpenUnix and OpenServer - * IBM AIX 5.1L - * macOS (Darwin) - * SGI IRIX 6.5 - * FreeBSD - * UnicOS / Cray - * Cygwin (Windows) - -If you successfully run Makeself and/or archives created with it on another -system, then please [let me know][2]! - -Examples of publicly available archives made using makeself are : - - * Game patches and installers for [Id Software][3] games like Quake 3 for Linux or Return To Castle Wolfenstein ; - * All game patches released by [Loki Software][4] for the Linux version of popular games ; - * The [nVidia drivers][5] for Linux - * The installer for the Linux version of [Google Earth][6] - * The [VirtualBox][7] installers for Linux - * The [Makeself][1] distribution itself ;-) - * and countless others... - -**Important note for Apache users:** By default, most Web servers will think that Makeself archives are regular text files and thus they may show up as text in a Web browser. The correct way to prevent this is to add a MIME type for this file format, like so (in httpd.conf) : - -`AddType application/x-makeself .run` - -**Important note for certain GNU/Linux distributions:** Archives created with Makeself prior to v2.1.2 were using an old syntax for the _head_ and _tail_ Unix commands that is being progressively obsoleted in their GNU forms. Therefore you may have problems uncompressing some of these archives. A workaround for this is to set the environment variable $_POSIX2_VERSION to enable the old syntax, i.e. : - -`export _POSIX2_VERSION=199209` - -## Usage - -The syntax of makeself is the following: - -``` -makeself.sh [args] archive_dir file_name label startup_script [script_args] -``` - - * _args_ are optional options for Makeself. The available ones are : - - * **`--version`** : Prints the version number on stdout, then exits immediately - * **`--gzip`** : Use gzip for compression (the default on platforms on which gzip is commonly available, like Linux) - * **`--bzip2`** : Use bzip2 instead of gzip for better compression. The bzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. - * **`--pbzip2`** : Use pbzip2 instead of gzip for better and faster compression on machines having multiple CPUs. The pbzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. - * **`--xz`** : Use xz instead of gzip for better compression. The xz command must be available in the command path. It is recommended that the archive prefix be set to something like '.xz.run' for the archive, so that potential users know that they'll need xz to extract it. - * **`--lzo`** : Use lzop instead of gzip for better compression. The lzop command must be available in the command path. It is recommended that the archive prefix be set to something like `.lzo.run` for the archive, so that potential users know that they'll need lzop to extract it. - * **`--lz4`** : Use lz4 instead of gzip for better compression. The lz4 command must be available in the command path. It is recommended that the archive prefix be set to something like '.lz4.run' for the archive, so that potential users know that they'll need lz4 to extract it. - * **`--zstd`** : Use zstd instead of gzip for better compression. The zstd command must be available in the command path. It is recommended that the archive prefix be set to something like '.zstd.run' for the archive, so that potential users know that they'll need zstd to extract it. - * **`--pigz`** : Use pigz for compression. - * **`--base64`** : Encode the archive to ASCII in Base64 format instead of compressing (base64 command required). - * **`--gpg-encrypt`** : Encrypt the archive using `gpg -ac -z $COMPRESS_LEVEL`. This will prompt for a password to encrypt with. Assumes that potential users have `gpg` installed. - * **`--ssl-encrypt`** : Encrypt the archive using `openssl aes-256-cbc -a -salt`. This will prompt for a password to encrypt with. Assumes that the potential users have the OpenSSL tools installed. - * **`--compress`** : Use the UNIX `compress` command to compress the data. This should be the default on all platforms that don't have gzip available. - * **`--nocomp`** : Do not use any compression for the archive, which will then be an uncompressed TAR. - * **`--complevel`** : Specify the compression level for gzip, bzip2, pbzip2, zstd, xz, lzo or lz4. (defaults to 9) - * **`--threads`** : Specify the number of threads to be used by compressors that support parallelization. Omit to use compressor's default. Most useful (and required) for opting into xz's threading, usually with `--threads=0` for all available cores. pbzip2 and pigz are parallel by default, and setting this value allows limiting the number of threads they use. - * **`--notemp`** : The generated archive will not extract the files to a temporary directory, but in a new directory created in the current directory. This is better to distribute software packages that may extract and compile by themselves (i.e. launch the compilation through the embedded script). - * **`--current`** : Files will be extracted to the current directory, instead of in a subdirectory. This option implies `--notemp` above. - * **`--follow`** : Follow the symbolic links inside of the archive directory, i.e. store the files that are being pointed to instead of the links themselves. - * **`--append`** _(new in 2.1.x)_: Append data to an existing archive, instead of creating a new one. In this mode, the settings from the original archive are reused (compression type, label, embedded script), and thus don't need to be specified again on the command line. - * **`--header`** : Makeself uses a separate file to store the header stub, called `makeself-header.sh`. By default, it is assumed that it is stored in the same location as makeself.sh. This option can be used to specify its actual location if it is stored someplace else. - * **`--cleanup`** : Specify a script that is run when execution is interrupted or finishes successfully. The script is executed with the same environment and initial `script_args` as `startup_script`. - * **`--copy`** : Upon extraction, the archive will first extract itself to a temporary directory. The main application of this is to allow self-contained installers stored in a Makeself archive on a CD, when the installer program will later need to unmount the CD and allow a new one to be inserted. This prevents "Filesystem busy" errors for installers that span multiple CDs. - * **`--nox11`** : Disable the automatic spawning of a new terminal in X11. - * **`--nowait`** : When executed from a new X11 terminal, disable the user prompt at the end of the script execution. - * **`--nomd5`** and **`--nocrc`** : Disable the creation of a MD5 / CRC checksum for the archive. This speeds up the extraction process if integrity checking is not necessary. - * **`--sha256`** : Adds a SHA256 checksum for the archive. This is in addition to the MD5 / CRC checksums unless `--nomd5` is also used. - * **`--lsm` _file_** : Provide and LSM file to makeself, that will be embedded in the generated archive. LSM files are describing a software package in a way that is easily parseable. The LSM entry can then be later retrieved using the `--lsm` argument to the archive. An example of a LSM file is provided with Makeself. - * **`--tar-format opt`** : Specify the tar archive format (default is ustar); you may use any value accepted by your tar command (such as posix, v7, etc). - * **`--tar-extra opt`** : Append more options to the tar command line. - - For instance, in order to exclude the `.git` directory from the packaged archive directory using the GNU `tar`, one can use `makeself.sh --tar-extra "--exclude=.git" ...` - - * **`--keep-umask`** : Keep the umask set to shell default, rather than overriding when executing self-extracting archive. - * **`--packaging-date date`** : Use provided string as the packaging date instead of the current date. - * **`--license`** : Append a license file. - * **`--nooverwrite`** : Do not extract the archive if the specified target directory already exists. - * **`--help-header file`** : Add a header to the archive's `--help` output. - * `archive_dir` is the name of the directory that contains the files to be archived - * `file_name` is the name of the archive to be created - * `label` is an arbitrary text string describing the package. It will be displayed while extracting the files. - * `startup_script` is the command to be executed _from within_ the directory of extracted files. Thus, if you wish to execute a program contained in this directory, you must prefix your command with `./`. For example, `./program` will be fine. The `script_args` are additional arguments for this command. - -Here is an example, assuming the user has a package image stored in a **/home/joe/mysoft**, and he wants to generate a self-extracting package named -**mysoft.sh**, which will launch the "setup" script initially stored in /home/joe/mysoft : - -`makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup -` - -Here is also how I created the [makeself.run][9] archive which contains the Makeself distribution : - -`makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" ` - -Archives generated with Makeself can be passed the following arguments: - - * **`--keep`** : Prevent the files to be extracted in a temporary directory that will be removed after the embedded script's execution. The files will then be extracted in the current working directory and will stay here until you remove them. - * **`--verbose`** : Will prompt the user before executing the embedded command - * **`--target dir`** : Allows to extract the archive in an arbitrary place. - * **`--nox11`** : Do not spawn a X11 terminal. - * **`--confirm`** : Prompt the user for confirmation before running the embedded command. - * **`--info`** : Print out general information about the archive (does not extract). - * **`--lsm`** : Print out the LSM entry, if it is present. - * **`--list`** : List the files in the archive. - * **`--check`** : Check the archive for integrity using the embedded checksums. Does not extract the archive. - * **`--nochown`** : By default, a `chown -R` command is run on the target directory after extraction, so that all files belong to the current user. This is mostly needed if you are running as root, as tar will then try to recreate the initial user ownerships. You may disable this behavior with this flag. - * **`--tar`** : Run the tar command on the contents of the archive, using the following arguments as parameter for the command. - * **`--noexec`** : Do not run the embedded script after extraction. - * **`--noexec-cleanup`** : Do not run the embedded cleanup script. - * **`--nodiskspace`** : Do not check for available disk space before attempting to extract. - * **`--cleanup-args`** : Specify arguments to be passed to the cleanup script. Wrap value in quotes to specify multiple arguments. - -Any subsequent arguments to the archive will be passed as additional arguments to the embedded command. You must explicitly use the `--` special command-line construct before any such options to make sure that Makeself will not try to interpret them. - -## Startup Script - -The startup script must be a regular Shell script. - -Within the startup script, you can use the `$USER_PWD` variable to get the path of the folder from which the self-extracting script is executed. This is especially useful to access files that are located in the same folder as the script, as shown in the example below. - -`my-self-extracting-script.sh --fooBarFileParameter foo.bar` - -## Building and Testing - -Clone the git repo and execute `git submodule update --init --recursive` to obtain all submodules. - -* To make a release: `make` -* To run all tests: `make test` - -## Maven Usage - -Makeself is now supported by the following maven plugin [makeself-maven-plugin](https://github.com/hazendaz/makeself-maven-plugin). Please refer to project for usage and report any bugs in regards to maven plugin on that project. - -## License - -Makeself itself is covered by the [GNU General Public License][8] (GPL) version 2 and above. Archives generated by Makeself don't have to be placed under this license (although I encourage it ;-)), since the archive itself is merely data for Makeself. - -## Contributing - -I will gladly consider merging your pull requests on the [GitHub][10] repository. However, please keep the following in mind: - - * One of the main purposes of Makeself is portability. Do not submit patches that will break supported platforms. The more platform-agnostic, the better. - * Please explain clearly what the purpose of the patch is, and how you achieved it. - -## Download - -Get the latest official distribution [here][9] (version 2.4.2). - -The latest development version can be grabbed from [GitHub][10]. Feel free to submit any patches there through the fork and pull request process. - -## Version history - - * **v1.0:** Initial public release - * **v1.1:** The archive can be passed parameters that will be passed on to the embedded script, thanks to John C. Quillan - * **v1.2:** Cosmetic updates, support for bzip2 compression and non-temporary archives. Many ideas thanks to Francois Petitjean. - * **v1.3:** More patches from Bjarni R. Einarsson and Francois Petitjean: Support for no compression (`--nocomp`), script is no longer mandatory, automatic launch in an xterm, optional verbose output, and -target archive option to indicate where to extract the files. - * **v1.4:** Many patches from Francois Petitjean: improved UNIX compatibility, automatic integrity checking, support of LSM files to get info on the package at run time.. - * **v1.5.x:** A lot of bugfixes, and many other patches, including automatic verification through the usage of checksums. Version 1.5.5 was the stable release for a long time, even though the Web page didn't get updated ;-). Makeself was also officially made a part of the [Loki Setup installer][11], and its source is being maintained as part of this package. - * **v2.0:** Complete internal rewrite of Makeself. The command-line parsing was vastly improved, the overall maintenance of the package was greatly improved by separating the stub from makeself.sh. Also Makeself was ported and tested to a variety of Unix platforms. - * **v2.0.1:** First public release of the new 2.0 branch. Prior versions are officially obsoleted. This release introduced the `--copy` argument that was introduced in response to a need for the [UT2K3][12] Linux installer. - * **v2.1.0:** Big change : Makeself can now support multiple embedded tarballs, each stored separately with their own checksums. An existing archive can be updated with the `--append` flag. Checksums are also better managed, and the `--nochown` option for archives appeared. - * **v2.1.1:** Fixes related to the Unix compression (compress command). Some Linux distributions made the insane choice to make it unavailable, even though gzip is capable of uncompressing these files, plus some more bugfixes in the extraction and checksum code. - * **v2.1.2:** Some bug fixes. Use head -n to avoid problems with POSIX conformance. - * **v2.1.3:** Bug fixes with the command line when spawning terminals. Added `--tar`, `--noexec` for archives. Added `--nomd5` and `--nocrc` to avoid creating checksums in archives. The embedded script is now run through "eval". The `--info` output now includes the command used to create the archive. A man page was contributed by Bartosz Fenski. - * **v2.1.4:** Fixed `--info` output. Generate random directory name when extracting files to . to avoid problems. Better handling of errors with wrong permissions for the directory containing the files. Avoid some race conditions, Unset the $CDPATH variable to avoid problems if it is set. Better handling of dot files in the archive directory. - * **v2.1.5:** Made the md5sum detection consistent with the header code. Check for the presence of the archive directory. Added `--encrypt` for symmetric encryption through gpg (Eric Windisch). Added support for the digest command on Solaris 10 for MD5 checksums. Check for available disk space before extracting to the target directory (Andreas Schweitzer). Allow extraction to run asynchronously (patch by Peter Hatch). Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo). - * **v2.1.6:** Replaced one dot per file progress with a realtime progress percentage and a spinning cursor. Added `--noprogress` to prevent showing the progress during the decompression. Added `--target` dir to allow extracting directly to a target directory. (Guy Baconniere) - * **v2.2.0:** First major new release in years! Includes many bugfixes and user contributions. Please look at the [project page on Github][10] for all the details. - * **v2.3.0:** Support for archive encryption via GPG or OpenSSL. Added LZO and LZ4 compression support. Options to set the packaging date and stop the umask from being overriden. Optionally ignore check for available disk space when extracting. New option to check for root permissions before extracting. - * **v2.3.1:** Various compatibility updates. Added unit tests for Travis CI in the GitHub repo. New `--tar-extra`, `--untar-extra`, `--gpg-extra`, `--gpg-asymmetric-encrypt-sign` options. - * **v2.4.0:** Added optional support for SHA256 archive integrity checksums. - * **v2.4.2:** New --cleanup and --cleanup-args arguments for cleanup scripts. Added threading support for supported compressors. Now supports zstd compression. - * **v2.4.3:** Make explicit POSIX tar archives for increased compatibility. - * **v2.4.4:** Fixed various compatibility issues (no longer use POSIX tar archives), Github Actions to check on Solaris and FreeBSD. - * **v2.4.5:** Added `--tar-format` option to set the tar archive format (default is ustar) - -## Links - - * Check out the ["Loki Setup"][11] installer, used to install many Linux games and other applications, and of which I am the co-author. Since the demise of Loki, I am now the official maintainer of the project, and it is now being hosted here on GitHub. - * Bjarni R. Einarsson also wrote the **setup.sh** installer script, inspired by Makeself. [Check it out !][14] - -## Contact - -This script was written by [Stéphane Peter][15] (megastep at megastep.org). Any enhancements and suggestions are welcome. - -Contributions were included from John C. Quillan, Bjarni R. Einarsson, -Francois Petitjean, Ryan C. Gordon, and many contributors on GitHub. If you think I forgot -your name, don't hesitate to contact me. - -This project is now hosted on GitHub. Feel free to submit patches and bug reports on the [project page][10]. - -* * * - -[Stephane Peter][2] - - [1]: http://makeself.io/ - [2]: mailto:megastep@megastep.org - [3]: http://www.idsoftware.com/ - [4]: http://www.lokigames.com/products/myth2/updates.php3 - [5]: http://www.nvidia.com/ - [6]: http://earth.google.com/ - [7]: http://www.virtualbox.org/ - [8]: http://www.gnu.org/copyleft/gpl.html - [9]: https://github.com/megastep/makeself/releases/download/release-2.4.5/makeself-2.4.5.run - [10]: https://github.com/megastep/makeself - [11]: https://github.com/megastep/loki_setup/ - [12]: http://www.unrealtournament2003.com/ - [13]: http://www.icculus.org/ - [14]: http://bre.klaki.net/programs/setup.sh/ - [15]: https://stephanepeter.com/ diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION deleted file mode 100644 index 59aa62c1f..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/VERSION +++ /dev/null @@ -1 +0,0 @@ -2.4.5 diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh deleted file mode 100644 index b5692d490..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/make-release.sh +++ /dev/null @@ -1,9 +0,0 @@ -#!/bin/sh -# -# Create a distributable archive of the current version of Makeself - -VER=`cat VERSION` -mkdir -p /tmp/makeself-$VER release -cp -pPR makeself* test README.md COPYING VERSION .gitmodules /tmp/makeself-$VER/ -./makeself.sh --notemp /tmp/makeself-$VER release/makeself-$VER.run "Makeself v$VER" echo "Makeself has extracted itself" - diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh deleted file mode 100644 index 940903148..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself-header.sh +++ /dev/null @@ -1,660 +0,0 @@ -cat << EOF > "$archname" -#!/bin/bash -# This script was generated using Makeself $MS_VERSION -# The license covering this archive and its contents, if any, is wholly independent of the Makeself license (GPL) -# 2022.3.19-Modified the MS_Help function and some options -# Huawei Technologies Co., Ltd. - -ORIG_UMASK=\`umask\` - -CRCsum="$CRCsum" -MD5="$MD5sum" -SHA="$SHAsum" -SIGNATURE="$Signature" -TMPROOT=\${TMPDIR:="\$HOME"} -if ! test -d "\$TMPROOT"; then - TMPROOT="\$PWD" -fi -export TMPDIR="\$TMPROOT" -USER_PWD="\$PWD" -if ! test -d "\$USER_PWD"; then - exit 1 -fi -export USER_PWD -ARCHIVE_DIR=\`dirname "\$0"\` -export ARCHIVE_DIR - -name_of_file="\$0 " -pwd_of_file="\$PWD" -label="$LABEL" -script="$SCRIPT" -scriptargs="$SCRIPTARGS" -cleanup_script="${CLEANUP_SCRIPT}" -licensetxt="$LICENSE" -helpheader='$HELPHEADER' -targetdir="$archdirname" -filesizes="$filesizes" -totalsize="$totalsize" -keep="$KEEP" -nooverwrite="$NOOVERWRITE" -quiet="n" -accept="n" -nodiskspace="n" -export_conf="$EXPORT_CONF" -decrypt_cmd="$DECRYPT_CMD" -skip="$SKIP" - -print_cmd_arg="" -if type printf > /dev/null; then - print_cmd="printf" -elif test -x /usr/ucb/echo; then - print_cmd="/usr/ucb/echo" -else - print_cmd="echo" -fi - -if test -d /usr/xpg4/bin; then - PATH=/usr/xpg4/bin:\$PATH - export PATH -fi - -if test -d /usr/sfw/bin; then - PATH=\$PATH:/usr/sfw/bin - export PATH -fi - -unset CDPATH - -MS_Printf() -{ - \$print_cmd \$print_cmd_arg "\$1" -} - -MS_PrintLicense() -{ - PAGER=\${PAGER:=more} - if test x"\$licensetxt" != x; then - PAGER_PATH=\`exec <&- 2>&-; which \$PAGER || command -v \$PAGER || type \$PAGER\` - if test -x "\$PAGER_PATH"; then - echo "\$licensetxt" | \$PAGER - else - echo "\$licensetxt" - fi - if test x"\$accept" != xy; then - while true - do - MS_Printf "Please type y to accept, n otherwise: " - read yn - if test x"\$yn" = xn; then - keep=n - eval \$finish; exit 1 - break; - elif test x"\$yn" = xy; then - break; - fi - done - fi - fi -} - -MS_diskspace() -{ - ( - df -kP "\$1" | tail -1 | awk '{ if (\$4 ~ /%/) {print \$3} else {print \$4} }' - ) -} - -MS_dd() -{ - blocks=\`expr \$3 / 1024\` - bytes=\`expr \$3 % 1024\` - # Test for ibs, obs and conv feature - if dd if=/dev/zero of=/dev/null count=1 ibs=512 obs=512 conv=sync 2> /dev/null; then - dd if="\$1" ibs=\$2 skip=1 obs=1024 conv=sync 2> /dev/null | \\ - { test \$blocks -gt 0 && dd ibs=1024 obs=1024 count=\$blocks ; \\ - test \$bytes -gt 0 && dd ibs=1 obs=1024 count=\$bytes ; } 2> /dev/null - else - dd if="\$1" bs=\$2 skip=1 2> /dev/null - fi -} - -MS_dd_Progress() -{ - if test x"\$noprogress" = xy; then - MS_dd "\$@" - return \$? - fi - file="\$1" - offset=\$2 - length=\$3 - pos=0 - bsize=4194304 - while test \$bsize -gt \$length; do - bsize=\`expr \$bsize / 4\` - done - blocks=\`expr \$length / \$bsize\` - bytes=\`expr \$length % \$bsize\` - ( - dd ibs=\$offset skip=1 2>/dev/null - pos=\`expr \$pos \+ \$bsize\` - MS_Printf " 0%% " 1>&2 - if test \$blocks -gt 0; then - while test \$pos -le \$length; do - dd bs=\$bsize count=1 2>/dev/null - pcent=\`expr \$length / 100\` - pcent=\`expr \$pos / \$pcent\` - if test \$pcent -lt 100; then - MS_Printf "\b\b\b\b\b\b\b" 1>&2 - if test \$pcent -lt 10; then - MS_Printf " \$pcent%% " 1>&2 - else - MS_Printf " \$pcent%% " 1>&2 - fi - fi - pos=\`expr \$pos \+ \$bsize\` - done - fi - if test \$bytes -gt 0; then - dd bs=\$bytes count=1 2>/dev/null - fi - MS_Printf "\b\b\b\b\b\b\b" 1>&2 - MS_Printf " 100%% " 1>&2 - ) < "\$file" -} - -MS_Help() -{ - cat << EOH >&2 -Usage: \$0 [options] -Options: - --help | -h Print this message - --info Print embedded info : title, default target directory, embedded script ... - --list Print the list of files in the archive - --check Checks integrity and version dependency of the archive - --quiet Quiet install mode, skip human-computer interactions - --nox11 Do not spawn an xterm - --noexec Do not run embedded script - --extract= Extract directly to a target directory (absolute or relative) - Usually used with --noexec to just extract files without running - --tar arg1 [arg2 ...] Access the contents of the archive through the tar command -\${helpheader} -EOH -} - -MS_Verify_Sig() -{ - GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` - MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` - test -x "\$GPG_PATH" || GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` - test -x "\$MKTEMP_PATH" || MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` - offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` - temp_sig=\`mktemp -t XXXXX\` - echo \$SIGNATURE | base64 --decode > "\$temp_sig" - gpg_output=\`MS_dd "\$1" \$offset \$totalsize | LC_ALL=C "\$GPG_PATH" --verify "\$temp_sig" - 2>&1\` - gpg_res=\$? - rm -f "\$temp_sig" - if test \$gpg_res -eq 0 && test \`echo \$gpg_output | grep -c Good\` -eq 1; then - if test \`echo \$gpg_output | grep -c \$sig_key\` -eq 1; then - test x"\$quiet" = xn && echo "GPG signature is good" >&2 - else - echo "GPG Signature key does not match" >&2 - exit 2 - fi - else - test x"\$quiet" = xn && echo "GPG signature failed to verify" >&2 - exit 2 - fi -} - -MS_Check() -{ - OLD_PATH="\$PATH" - PATH=\${GUESS_MD5_PATH:-"\$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} - MD5_ARG="" - MD5_PATH=\`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum\` - test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which md5 || command -v md5 || type md5\` - test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which digest || command -v digest || type digest\` - PATH="\$OLD_PATH" - - SHA_PATH=\`exec <&- 2>&-; which shasum || command -v shasum || type shasum\` - test -x "\$SHA_PATH" || SHA_PATH=\`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum\` - - if test x"\$quiet" = xn; then - MS_Printf "Verifying archive integrity..." - fi - offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` - fsize=\`cat "\$1" | wc -c | tr -d " "\` - if test \$totalsize -ne \`expr \$fsize - \$offset\`; then - echo " Unexpected archive size." >&2 - exit 2 - fi - verb=\$2 - i=1 - for s in \$filesizes - do - crc=\`echo \$CRCsum | cut -d" " -f\$i\` - if test -x "\$SHA_PATH"; then - if test x"\`basename \$SHA_PATH\`" = xshasum; then - SHA_ARG="-a 256" - fi - sha=\`echo \$SHA | cut -d" " -f\$i\` - if test x"\$sha" = x0000000000000000000000000000000000000000000000000000000000000000; then - test x"\$verb" = xy && echo " \$1 does not contain an embedded SHA256 checksum." >&2 - else - shasum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$SHA_PATH \$SHA_ARG" | cut -b-64\`; - if test x"\$shasum" != x"\$sha"; then - echo "Error in SHA256 checksums: \$shasum is different from \$sha" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " SHA256 checksums are OK." >&2 - fi - crc="0000000000"; - fi - fi - if test -x "\$MD5_PATH"; then - if test x"\`basename \$MD5_PATH\`" = xdigest; then - MD5_ARG="-a md5" - fi - md5=\`echo \$MD5 | cut -d" " -f\$i\` - if test x"\$md5" = x00000000000000000000000000000000; then - test x"\$verb" = xy && echo " \$1 does not contain an embedded MD5 checksum." >&2 - else - md5sum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$MD5_PATH \$MD5_ARG" | cut -b-32\`; - if test x"\$md5sum" != x"\$md5"; then - echo "Error in MD5 checksums: \$md5sum is different from \$md5" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " MD5 checksums are OK." >&2 - fi - crc="0000000000"; verb=n - fi - fi - if test x"\$crc" = x0000000000; then - test x"\$verb" = xy && echo " \$1 does not contain a CRC checksum." >&2 - else - sum1=\`MS_dd_Progress "\$1" \$offset \$s | CMD_ENV=xpg4 cksum | awk '{print \$1}'\` - if test x"\$sum1" != x"\$crc"; then - echo "Error in checksums: \$sum1 is different from \$crc" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " CRC checksums are OK." >&2 - fi - fi - i=\`expr \$i + 1\` - offset=\`expr \$offset + \$s\` - done - if test x"\$quiet" = xn; then - echo " All good." - fi -} - -MS_Decompress() -{ - if test x"\$decrypt_cmd" != x""; then - { eval "\$decrypt_cmd" || echo " ... Decryption failed." >&2; } | eval "$GUNZIP_CMD" - else - eval "$GUNZIP_CMD" - fi - - if test \$? -ne 0; then - echo " ... Decompression failed." >&2 - fi -} - -UnTAR() -{ - if test x"\$quiet" = xn; then - tar \$1vf - $UNTAR_EXTRA 2>&1 || { echo " ... Extraction failed." >&2; kill -15 \$$; } - else - tar \$1f - $UNTAR_EXTRA 2>&1 || { echo Extraction failed. >&2; kill -15 \$$; } - fi -} - -MS_exec_cleanup() { - if test x"\$cleanup" = xy && test x"\$cleanup_script" != x""; then - cleanup=n - cd "\$tmpdir" - eval "\"\$cleanup_script\" \$scriptargs \$cleanupargs" - fi -} - -MS_cleanup() -{ - echo 'Signal caught, cleaning up' >&2 - MS_exec_cleanup - cd "\$TMPROOT" - rm -rf "\$tmpdir" - eval \$finish; exit 15 -} - -Script_Args_Check() -{ - script_supported_args=\$(echo \${helpheader} | grep -o -E "\-\-[^ ]+" | awk -F"=" {'print \$1'}) - arg_to_test=\$(echo \$1|awk -F"=" {'print \$1'}) - - for arg in \${script_supported_args}; - do - if test x"\$arg_to_test" = x"\$arg" ;then - return - fi - done - - MS_Help - exit 1 -} - -finish=true -xterm_loop= -noprogress=$NOPROGRESS -nox11=$NOX11 -copy=$COPY -ownership=$OWNERSHIP -verbose=n -cleanup=y -cleanupargs= -sig_key= - -initargs="\$@" - -while [ -n "\$*" ] -do - case "\$1" in - -h | --help) - MS_Help - exit 0 - ;; - -q | --quiet) - quiet=y - noprogress=y - shift - ;; - --info) - echo Identification: "\$label" - echo Target directory: "\$targetdir" - echo Uncompressed size: $USIZE KB - echo Compression: $COMPRESS - if test x"$ENCRYPT" != x""; then - echo Encryption: $ENCRYPT - fi - echo Date of packaging: $DATE - echo Built with Makeself version $MS_VERSION - echo Build command was: "$MS_COMMAND" - if test x"\$script" != x; then - echo Script run after extraction: - echo " " \$script \$scriptargs - fi - if test x"$copy" = xcopy; then - echo "Archive will copy itself to a temporary location" - fi - if test x"$NEED_ROOT" = xy; then - echo "Root permissions required for extraction" - fi - if test x"$KEEP" = xy; then - echo "directory \$targetdir is permanent" - else - echo "\$targetdir will be removed after extraction" - fi - exit 0 - ;; - --list) - echo Target directory: \$targetdir - offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - for s in \$filesizes - do - MS_dd "\$0" \$offset \$s | MS_Decompress | UnTAR t - offset=\`expr \$offset + \$s\` - done - exit 0 - ;; - --tar) - offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - arg1="\$2" - shift 2 || { MS_Help; exit 1; } - for s in \$filesizes - do - MS_dd "\$0" \$offset \$s | MS_Decompress | tar "\$arg1" - "\$@" - offset=\`expr \$offset + \$s\` - done - exit 0 - ;; - --check) - MS_Check "\$0" y - scriptargs="\$scriptargs \$1" - shift - ;; - --noexec) - script="" - cleanup_script="" - shift - ;; - --extract=*) - keep=y - targetdir=\`echo \$1 | cut -d"=" -f2 \` - if ! shift; then MS_Help; exit 1; fi - ;; - --nox11) - nox11=y - shift - ;; - --xwin) - if test "$NOWAIT" = n; then - finish="echo Press Return to close this window...; read junk" - fi - xterm_loop=1 - shift - ;; - --phase2) - copy=phase2 - shift - ;; - --repack | --repack-path=*) - Script_Args_Check \$1 - scriptargs="\$scriptargs '\$1'" - shift - if [[ ! "\$1" =~ ^-.* ]]; then - scriptargs="\$scriptargs '\$1'" - shift - fi - ;; - *) - Script_Args_Check \$1 - scriptargs="\$scriptargs '\$1'" - shift - ;; - esac -done - -quiet_para="" -if test x"\$quiet" = xy; then - quiet_para="--quiet " -fi -scriptargs="--\$name_of_file""--\"\$pwd_of_file\""" \$quiet_para""\$scriptargs" - -if test x"\$quiet" = xy -a x"\$verbose" = xy; then - echo Cannot be verbose and quiet at the same time. >&2 - exit 1 -fi - -if test x"$NEED_ROOT" = xy -a \`id -u\` -ne 0; then - echo "Administrative privileges required for this archive (use su or sudo)" >&2 - exit 1 -fi - -if test x"\$copy" \!= xphase2; then - MS_PrintLicense -fi - -case "\$copy" in -copy) - tmpdir="\$TMPROOT"/makeself.\$RANDOM.\`date +"%y%m%d%H%M%S"\`.\$\$ - mkdir "\$tmpdir" || { - echo "Could not create temporary directory \$tmpdir" >&2 - exit 1 - } - SCRIPT_COPY="\$tmpdir/makeself" - echo "Copying to a temporary location..." >&2 - cp "\$0" "\$SCRIPT_COPY" - chmod +x "\$SCRIPT_COPY" - cd "\$TMPROOT" - exec "\$SCRIPT_COPY" --phase2 -- \$initargs - ;; -phase2) - finish="\$finish ; rm -rf \`dirname \$0\`" - ;; -esac - -if test x"\$nox11" = xn; then - if tty -s; then # Do we have a terminal? - : - else - if test x"\$DISPLAY" != x -a x"\$xterm_loop" = x; then # No, but do we have X? - if xset q > /dev/null 2>&1; then # Check for valid DISPLAY variable - GUESS_XTERMS="xterm gnome-terminal rxvt dtterm eterm Eterm xfce4-terminal lxterminal kvt konsole aterm terminology" - for a in \$GUESS_XTERMS; do - if type \$a >/dev/null 2>&1; then - XTERM=\$a - break - fi - done - chmod a+x \$0 || echo Please add execution rights on \$0 - if test \`echo "\$0" | cut -c1\` = "/"; then # Spawn a terminal! - exec \$XTERM -e "\$0 --xwin \$initargs" - else - exec \$XTERM -e "./\$0 --xwin \$initargs" - fi - fi - fi - fi -fi - -if test x"\$targetdir" = x.; then - tmpdir="." -else - if test x"\$keep" = xy; then - if test x"\$nooverwrite" = xy && test -d "\$targetdir"; then - echo "Target directory \$targetdir already exists, aborting." >&2 - exit 1 - fi - if test x"\$quiet" = xn; then - echo "Creating directory \$targetdir" >&2 - fi - tmpdir="\$targetdir" - dashp="-p" - else - tmpdir="\$TMPROOT/selfgz\$\$\$RANDOM" - dashp="" - fi - mkdir \$dashp "\$tmpdir" || { - echo 'Cannot create target directory' \$tmpdir >&2 - echo 'You should try option --extract=' >&2 - eval \$finish - exit 1 - } -fi - -location="\`pwd\`" -if test x"\$SETUP_NOCHECK" != x1; then - MS_Check "\$0" -fi -offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - -if test x"\$verbose" = xy; then - MS_Printf "About to extract $USIZE KB in \$tmpdir ... Proceed ? [Y/n] " - read yn - if test x"\$yn" = xn; then - eval \$finish; exit 1 - fi -fi - -if test x"\$quiet" = xn; then - # Decrypting with openssl will ask for password, - # the prompt needs to start on new line - if test x"$ENCRYPT" = x"openssl"; then - echo "Decrypting and uncompressing \$label..." - else - MS_Printf "Uncompressing \$label" - fi -fi -res=3 -if test x"\$keep" = xn; then - trap MS_cleanup 1 2 3 15 -fi - -if test x"\$nodiskspace" = xn; then - leftspace=\`MS_diskspace "\$tmpdir"\` - if test -n "\$leftspace"; then - if test "\$leftspace" -lt $USIZE; then - echo - echo "Not enough space left in "\`dirname \$tmpdir\`" (\$leftspace KB) to decompress \$0 ($USIZE KB)" >&2 - if test x"\$keep" = xn; then - echo "Consider setting TMPDIR to a directory with more free space." - fi - eval \$finish; exit 1 - fi - fi -fi - -for s in \$filesizes -do - if MS_dd_Progress "\$0" \$offset \$s | MS_Decompress | ( cd "\$tmpdir"; umask \$ORIG_UMASK ; UnTAR xp ) 1>/dev/null; then - if test x"\$ownership" = xy; then - (cd "\$tmpdir"; chown -R \`id -u\` .; chgrp -R \`id -g\` .) - fi - else - echo >&2 - echo "Unable to decompress \$0" >&2 - eval \$finish; exit 1 - fi - offset=\`expr \$offset + \$s\` -done -if test x"\$quiet" = xn; then - echo -fi - -cd "\$tmpdir" -res=0 -if test x"\$script" != x; then - if test x"\$export_conf" = x"y"; then - MS_BUNDLE="\$0" - MS_LABEL="\$label" - MS_SCRIPT="\$script" - MS_SCRIPTARGS="\$scriptargs" - MS_ARCHDIRNAME="\$archdirname" - MS_KEEP="\$KEEP" - MS_NOOVERWRITE="\$NOOVERWRITE" - MS_COMPRESS="\$COMPRESS" - MS_CLEANUP="\$cleanup" - export MS_BUNDLE MS_LABEL MS_SCRIPT MS_SCRIPTARGS - export MS_ARCHDIRNAME MS_KEEP MS_NOOVERWRITE MS_COMPRESS - fi - - if test x"\$verbose" = x"y"; then - yn="x" - while test x"\$yn" != x -a x"\$yn" != xy -a x"\$yn" != xY -a x"\$yn" != xn -a x"\$yn" != xN - do - MS_Printf "OK to execute: \$script \$scriptargs \$* ? [Y/n] " - read yn - if test x"\$yn" = x -o x"\$yn" = xy -o x"\$yn" = xY; then - eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$?; - elif test x"\$yn" = xn -o x"\$yn" = xN; then - echo "Unable to decompress \$script ,because of aborting! ";res=\$? - else - echo "Input value is unacceptable,please try again." - fi - done - else - eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$? - fi - if test "\$res" -ne 0; then - test x"\$verbose" = xy && echo "The program '\$script' returned an error code (\$res)" >&2 - fi -fi - -MS_exec_cleanup - -if test x"\$keep" = xn; then - cd "\$TMPROOT" - rm -rf "\$tmpdir" -fi -eval \$finish; exit \$res -EOF diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 deleted file mode 100644 index 81bf6e4ff..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.1 +++ /dev/null @@ -1,110 +0,0 @@ -.TH "MAKESELF" "1" "2.4.5" -.SH "NAME" -makeself \- An utility to generate self-extractable archives. -.SH "SYNTAX" -.B makeself [\fIoptions\fP] archive_dir file_name label -.B [\fIstartup_script\fP] [\fIargs\fP] -.SH "DESCRIPTION" -This program is a free (GPL) utility designed to create self-extractable -archives from a directory. -.SH "OPTIONS" -The following options are supported. -.TP 15 -.B -v, --version -Prints out the makeself version number and exits. -.TP -.B -h, --help -Print out help information. -.TP -.B --tar-quietly -Suppress verbose output from the tar command -.TP -.B --quiet -Do not print any messages other than errors -.TP -.B --gzip -Compress using gzip (default if detected). -.TP -.B --bzip2 -Compress using bzip2. -.TP -.B --pbzip2 -Compress using pbzip2. -.TP -.B --xz -Compress using xz. -.TP -.B --lzo -Compress using lzop. -.TP -.B --lz4 -Compress using lz4. -.TP -.B --compress -Compress using the UNIX 'compress' command. -.TP -.B --nocomp -Do not compress the data. -.TP -.B --complevel lvl -Specify the compression level for gzip,bzip2,pbzui2,xz,lzo or lz4 -.TP -.B --notemp -The archive will create archive_dir in the current directory and -uncompress in ./archive_dir. -.TP -.B --copy -Upon extraction, the archive will first copy itself to a temporary directory. -.TP -.B --append -Append more files to an existing makeself archive. The label and startup scripts will then be ignored. -.TP -.B --current -Files will be extracted to the current directory. Both --current and --target dir imply --notemp. -.TP -.B --target dir -Extract directly to a target directory. Directory path can be either absolute or relative. -.TP -.B --header file -Specify location of the header script. -.TP -.B --cleanup file -Specify a cleanup script that executes on interrupt and when finished successfully. -.TP -.B --follow -Follow the symlinks in the archive. -.TP -.B --noprogress -Do not show the progress during the decompression. -.TP -.B --nox11 -Disable automatic spawn of an xterm if running in X11. -.TP -.B --nowait -Do not wait for user input after executing embedded program from an xterm. -.TP -.B --nomd5 -Do not create a MD5 checksum for the archive. -.TP -.B --nocrc -Do not create a CRC32 checksum for the archive. -.TP -.B --lsm file -LSM file describing the package. -.B --packaging-date date -Use provided string as the packaging date instead of the current date. -.SH "EXAMPLES" -Here is an example, assuming the user has a package image stored in a /home/joe/mysoft, -and he wants to generate a self-extracting package named mysoft.sh, which will launch -the "setup" script initially stored in /home/joe/mysoft: -.TP -makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup -.TP -Here is also how I created the makeself.run archive which contains the Makeself distribution: -.TP -makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" -.SH "AUTHORS" -Makeself has been written by Stéphane Peter . -.BR -This man page was originally written by Bartosz Fenski for the -Debian GNU/Linux distribution (but it may be used by others). diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm deleted file mode 100644 index 3c4cea8c1..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.lsm +++ /dev/null @@ -1,16 +0,0 @@ -Begin3 -Title: makeself.sh -Version: 2.4.5 -Description: makeself.sh is a shell script that generates a self-extractable - tar.gz archive from a directory. The resulting file appears as a shell - script, and can be launched as is. The archive will then uncompress - itself to a temporary directory and an arbitrary command will be - executed (for example an installation script). This is pretty similar - to archives generated with WinZip Self-Extractor in the Windows world. -Keywords: Installation archive tar winzip -Author: Stephane Peter (megastep@megastep.org) -Maintained-by: Stephane Peter (megastep@megastep.org) -Original-site: https://makeself.io/ -Platform: Unix -Copying-policy: GPL -End diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh deleted file mode 100755 index c8ea56597..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/makeself.sh +++ /dev/null @@ -1,822 +0,0 @@ -#!/bin/sh -# -# Makeself version 2.4.x -# by Stephane Peter -# -# Utility to create self-extracting tar.gz archives. -# The resulting archive is a file holding the tar.gz archive with -# a small Shell script stub that uncompresses the archive to a temporary -# directory and then executes a given script from withing that directory. -# -# Makeself home page: https://makeself.io/ -# -# Version 2.0 is a rewrite of version 1.0 to make the code easier to read and maintain. -# -# Version history : -# - 1.0 : Initial public release -# - 1.1 : The archive can be passed parameters that will be passed on to -# the embedded script, thanks to John C. Quillan -# - 1.2 : Package distribution, bzip2 compression, more command line options, -# support for non-temporary archives. Ideas thanks to Francois Petitjean -# - 1.3 : More patches from Bjarni R. Einarsson and Francois Petitjean: -# Support for no compression (--nocomp), script is no longer mandatory, -# automatic launch in an xterm, optional verbose output, and -target -# archive option to indicate where to extract the files. -# - 1.4 : Improved UNIX compatibility (Francois Petitjean) -# Automatic integrity checking, support of LSM files (Francois Petitjean) -# - 1.5 : Many bugfixes. Optionally disable xterm spawning. -# - 1.5.1 : More bugfixes, added archive options -list and -check. -# - 1.5.2 : Cosmetic changes to inform the user of what's going on with big -# archives (Quake III demo) -# - 1.5.3 : Check for validity of the DISPLAY variable before launching an xterm. -# More verbosity in xterms and check for embedded command's return value. -# Bugfix for Debian 2.0 systems that have a different "print" command. -# - 1.5.4 : Many bugfixes. Print out a message if the extraction failed. -# - 1.5.5 : More bugfixes. Added support for SETUP_NOCHECK environment variable to -# bypass checksum verification of archives. -# - 1.6.0 : Compute MD5 checksums with the md5sum command (patch from Ryan Gordon) -# - 2.0 : Brand new rewrite, cleaner architecture, separated header and UNIX ports. -# - 2.0.1 : Added --copy -# - 2.1.0 : Allow multiple tarballs to be stored in one archive, and incremental updates. -# Added --nochown for archives -# Stopped doing redundant checksums when not necesary -# - 2.1.1 : Work around insane behavior from certain Linux distros with no 'uncompress' command -# Cleaned up the code to handle error codes from compress. Simplified the extraction code. -# - 2.1.2 : Some bug fixes. Use head -n to avoid problems. -# - 2.1.3 : Bug fixes with command line when spawning terminals. -# Added --tar for archives, allowing to give arbitrary arguments to tar on the contents of the archive. -# Added --noexec to prevent execution of embedded scripts. -# Added --nomd5 and --nocrc to avoid creating checksums in archives. -# Added command used to create the archive in --info output. -# Run the embedded script through eval. -# - 2.1.4 : Fixed --info output. -# Generate random directory name when extracting files to . to avoid problems. (Jason Trent) -# Better handling of errors with wrong permissions for the directory containing the files. (Jason Trent) -# Avoid some race conditions (Ludwig Nussel) -# Unset the $CDPATH variable to avoid problems if it is set. (Debian) -# Better handling of dot files in the archive directory. -# - 2.1.5 : Made the md5sum detection consistent with the header code. -# Check for the presence of the archive directory -# Added --encrypt for symmetric encryption through gpg (Eric Windisch) -# Added support for the digest command on Solaris 10 for MD5 checksums -# Check for available disk space before extracting to the target directory (Andreas Schweitzer) -# Allow extraction to run asynchronously (patch by Peter Hatch) -# Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo) -# - 2.1.6 : Replaced one dot per file progress with a realtime progress percentage and a spining cursor (Guy Baconniere) -# Added --noprogress to prevent showing the progress during the decompression (Guy Baconniere) -# Added --target dir to allow extracting directly to a target directory (Guy Baconniere) -# - 2.2.0 : Many bugfixes, updates and contributions from users. Check out the project page on Github for the details. -# - 2.3.0 : Option to specify packaging date to enable byte-for-byte reproducibility. (Marc Pawlowsky) -# - 2.4.0 : Optional support for SHA256 checksums in archives. -# - 2.4.2 : Add support for threads for several compressors. (M. Limber) -# Added zstd support. -# - 2.4.3 : Make explicit POSIX tar archives for increased compatibility. -# - 2.4.5 : Added --tar-format to override ustar tar archive format -# -# (C) 1998-2021 by Stephane Peter -# -# This software is released under the terms of the GNU GPL version 2 and above -# Please read the license at http://www.gnu.org/copyleft/gpl.html -# Self-extracting archives created with this script are explictly NOT released under the term of the GPL -# - -MS_VERSION=2.4.5 -MS_COMMAND="$0" -unset CDPATH - -for f in ${1+"$@"}; do - MS_COMMAND="$MS_COMMAND \\\\ - \\\"$f\\\"" -done - -# For Solaris systems -if test -d /usr/xpg4/bin; then - PATH=/usr/xpg4/bin:$PATH - export PATH -fi - -# Procedures - -MS_Usage() -{ - echo "Usage: $0 [args] archive_dir file_name label startup_script [script_args]" - echo "args can be one or more of the following :" - echo " --version | -v : Print out Makeself version number and exit" - echo " --help | -h : Print out this help message" - echo " --tar-quietly : Suppress verbose output from the tar command" - echo " --quiet | -q : Do not print any messages other than errors." - echo " --gzip : Compress using gzip (default if detected)" - echo " --pigz : Compress with pigz" - echo " --zstd : Compress with zstd" - echo " --bzip2 : Compress using bzip2 instead of gzip" - echo " --pbzip2 : Compress using pbzip2 instead of gzip" - echo " --xz : Compress using xz instead of gzip" - echo " --lzo : Compress using lzop instead of gzip" - echo " --lz4 : Compress using lz4 instead of gzip" - echo " --compress : Compress using the UNIX 'compress' command" - echo " --complevel lvl : Compression level for gzip pigz zstd xz lzo lz4 bzip2 and pbzip2 (default 9)" - echo " --threads thds : Number of threads to be used by compressors that support parallelization." - echo " Omit to use compressor's default. Most useful (and required) for opting" - echo " into xz's threading, usually with '--threads=0' for all available cores." - echo " pbzip2 and pigz are parallel by default, and setting this value allows" - echo " limiting the number of threads they use." - echo " --base64 : Instead of compressing, encode the data using base64" - echo " --gpg-encrypt : Instead of compressing, encrypt the data using GPG" - echo " --gpg-asymmetric-encrypt-sign" - echo " : Instead of compressing, asymmetrically encrypt and sign the data using GPG" - echo " --gpg-extra opt : Append more options to the gpg command line" - echo " --ssl-encrypt : Instead of compressing, encrypt the data using OpenSSL" - echo " --ssl-passwd pass : Use the given password to encrypt the data using OpenSSL" - echo " --ssl-pass-src src : Use the given src as the source of password to encrypt the data" - echo " using OpenSSL. See \"PASS PHRASE ARGUMENTS\" in man openssl." - echo " If this option is not supplied, the user will be asked to enter" - echo " encryption password on the current terminal." - echo " --ssl-no-md : Do not use \"-md\" option not supported by older OpenSSL." - echo " --nochown : Do not give the target folder to the current user (default)" - echo " --chown : Give the target folder to the current user recursively" - echo " --nocomp : Do not compress the data" - echo " --notemp : The archive will create archive_dir in the" - echo " current directory and uncompress in ./archive_dir" - echo " --needroot : Check that the root user is extracting the archive before proceeding" - echo " --copy : Upon extraction, the archive will first copy itself to" - echo " a temporary directory" - echo " --append : Append more files to an existing Makeself archive" - echo " The label and startup scripts will then be ignored" - echo " --target dir : Extract directly to a target directory" - echo " directory path can be either absolute or relative" - echo " --nooverwrite : Do not extract the archive if the specified target directory exists" - echo " --current : Files will be extracted to the current directory" - echo " Both --current and --target imply --notemp" - echo " --tar-format opt : Specify a tar archive format (default is ustar)" - echo " --tar-extra opt : Append more options to the tar command line" - echo " --untar-extra opt : Append more options to the during the extraction of the tar archive" - echo " --nomd5 : Don't calculate an MD5 for archive" - echo " --nocrc : Don't calculate a CRC for archive" - echo " --sha256 : Compute a SHA256 checksum for the archive" - echo " --header file : Specify location of the header script" - echo " --cleanup file : Specify a cleanup script that executes on interrupt and when finished successfully." - echo " --follow : Follow the symlinks in the archive" - echo " --noprogress : Do not show the progress during the decompression" - echo " --nox11 : Disable automatic spawn of a xterm" - echo " --nowait : Do not wait for user input after executing embedded" - echo " program from an xterm" - echo " --sign passphrase : Signature private key to sign the package with" - echo " --lsm file : LSM file describing the package" - echo " --license file : Append a license file" - echo " --help-header file : Add a header to the archive's --help output" - echo " --packaging-date date" - echo " : Use provided string as the packaging date" - echo " instead of the current date." - echo - echo " --keep-umask : Keep the umask set to shell default, rather than overriding when executing self-extracting archive." - echo " --export-conf : Export configuration variables to startup_script" - echo - echo "Do not forget to give a fully qualified startup script name" - echo "(i.e. with a ./ prefix if inside the archive)." - exit 1 -} - -# Default settings -if type gzip >/dev/null 2>&1; then - COMPRESS=gzip -elif type compress >/dev/null 2>&1; then - COMPRESS=compress -else - echo "ERROR: missing commands: gzip, compress" >&2 - MS_Usage -fi -ENCRYPT=n -PASSWD="" -PASSWD_SRC="" -OPENSSL_NO_MD=n -COMPRESS_LEVEL=9 -DEFAULT_THREADS=123456 # Sentinel value -THREADS=$DEFAULT_THREADS -KEEP=n -CURRENT=n -NOX11=n -NOWAIT=n -APPEND=n -TAR_QUIETLY=n -KEEP_UMASK=n -QUIET=n -NOPROGRESS=n -COPY=none -NEED_ROOT=n -TAR_ARGS=rvf -TAR_FORMAT=ustar -TAR_EXTRA="" -GPG_EXTRA="" -DU_ARGS=-ks -HEADER=`dirname "$0"`/makeself-header.sh -SIGNATURE="" -TARGETDIR="" -NOOVERWRITE=n -DATE=`LC_ALL=C date` -EXPORT_CONF=n -SHA256=n -OWNERSHIP=n -SIGN=n -GPG_PASSPHRASE="" - -# LSM file stuff -LSM_CMD="echo No LSM. >> \"\$archname\"" - -while true -do - case "$1" in - --version | -v) - echo Makeself version $MS_VERSION - exit 0 - ;; - --pbzip2) - COMPRESS=pbzip2 - shift - ;; - --bzip2) - COMPRESS=bzip2 - shift - ;; - --gzip) - COMPRESS=gzip - shift - ;; - --pigz) - COMPRESS=pigz - shift - ;; - --zstd) - COMPRESS=zstd - shift - ;; - --xz) - COMPRESS=xz - shift - ;; - --lzo) - COMPRESS=lzo - shift - ;; - --lz4) - COMPRESS=lz4 - shift - ;; - --compress) - COMPRESS=compress - shift - ;; - --base64) - COMPRESS=base64 - shift - ;; - --gpg-encrypt) - COMPRESS=gpg - shift - ;; - --gpg-asymmetric-encrypt-sign) - COMPRESS=gpg-asymmetric - shift - ;; - --gpg-extra) - GPG_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-encrypt) - ENCRYPT=openssl - shift - ;; - --ssl-passwd) - PASSWD=$2 - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-pass-src) - PASSWD_SRC=$2 - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-no-md) - OPENSSL_NO_MD=y - shift - ;; - --nocomp) - COMPRESS=none - shift - ;; - --complevel) - COMPRESS_LEVEL="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --threads) - THREADS="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --nochown) - OWNERSHIP=n - shift - ;; - --chown) - OWNERSHIP=y - shift - ;; - --notemp) - KEEP=y - shift - ;; - --copy) - COPY=copy - shift - ;; - --current) - CURRENT=y - KEEP=y - shift - ;; - --tar-format) - TAR_FORMAT="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --tar-extra) - TAR_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --untar-extra) - UNTAR_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --target) - TARGETDIR="$2" - KEEP=y - shift 2 || { MS_Usage; exit 1; } - ;; - --sign) - SIGN=y - GPG_PASSPHRASE="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --nooverwrite) - NOOVERWRITE=y - shift - ;; - --needroot) - NEED_ROOT=y - shift - ;; - --header) - HEADER="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --cleanup) - CLEANUP_SCRIPT="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --license) - # We need to escape all characters having a special meaning in double quotes - LICENSE=$(sed 's/\\/\\\\/g; s/"/\\\"/g; s/`/\\\`/g; s/\$/\\\$/g' "$2") - shift 2 || { MS_Usage; exit 1; } - ;; - --follow) - TAR_ARGS=rvhf - DU_ARGS=-ksL - shift - ;; - --noprogress) - NOPROGRESS=y - shift - ;; - --nox11) - NOX11=y - shift - ;; - --nowait) - NOWAIT=y - shift - ;; - --nomd5) - NOMD5=y - shift - ;; - --sha256) - SHA256=y - shift - ;; - --nocrc) - NOCRC=y - shift - ;; - --append) - APPEND=y - shift - ;; - --lsm) - LSM_CMD="cat \"$2\" >> \"\$archname\"" - shift 2 || { MS_Usage; exit 1; } - ;; - --packaging-date) - DATE="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --help-header) - HELPHEADER=`sed -e "s/'/'\\\\\''/g" $2` - shift 2 || { MS_Usage; exit 1; } - [ -n "$HELPHEADER" ] && HELPHEADER="$HELPHEADER -" - ;; - --tar-quietly) - TAR_QUIETLY=y - shift - ;; - --keep-umask) - KEEP_UMASK=y - shift - ;; - --export-conf) - EXPORT_CONF=y - shift - ;; - -q | --quiet) - QUIET=y - shift - ;; - -h | --help) - MS_Usage - ;; - -*) - echo Unrecognized flag : "$1" - MS_Usage - ;; - *) - break - ;; - esac -done - -if test $# -lt 1; then - MS_Usage -else - if test -d "$1"; then - archdir="$1" - else - echo "Directory $1 does not exist." >&2 - exit 1 - fi -fi -archname="$2" - -if test "$QUIET" = "y" || test "$TAR_QUIETLY" = "y"; then - if test "$TAR_ARGS" = "rvf"; then - TAR_ARGS="rf" - elif test "$TAR_ARGS" = "rvhf"; then - TAR_ARGS="rhf" - fi -fi - -if test "$APPEND" = y; then - if test $# -lt 2; then - MS_Usage - fi - - # Gather the info from the original archive - OLDENV=`sh "$archname" --dumpconf` - if test $? -ne 0; then - echo "Unable to update archive: $archname" >&2 - exit 1 - else - eval "$OLDENV" - OLDSKIP=`expr $SKIP + 1` - fi -else - if test "$KEEP" = n -a $# = 3; then - echo "ERROR: Making a temporary archive with no embedded command does not make sense!" >&2 - echo >&2 - MS_Usage - fi - # We don't want to create an absolute directory unless a target directory is defined - if test "$CURRENT" = y; then - archdirname="." - elif test x"$TARGETDIR" != x; then - archdirname="$TARGETDIR" - else - archdirname=`basename "$1"` - fi - - if test $# -lt 3; then - MS_Usage - fi - - LABEL="$3" - SCRIPT="$4" - test "x$SCRIPT" = x || shift 1 - shift 3 - SCRIPTARGS="$*" -fi - -if test "$KEEP" = n -a "$CURRENT" = y; then - echo "ERROR: It is A VERY DANGEROUS IDEA to try to combine --notemp and --current." >&2 - exit 1 -fi - -case $COMPRESS in -gzip) - GZIP_CMD="gzip -c$COMPRESS_LEVEL" - GUNZIP_CMD="gzip -cd" - ;; -pigz) - GZIP_CMD="pigz -$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD --processes $THREADS" - fi - GUNZIP_CMD="gzip -cd" - ;; -zstd) - GZIP_CMD="zstd -$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD --threads=$THREADS" - fi - GUNZIP_CMD="zstd -cd" - ;; -pbzip2) - GZIP_CMD="pbzip2 -c$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD -p$THREADS" - fi - GUNZIP_CMD="bzip2 -d" - ;; -bzip2) - GZIP_CMD="bzip2 -$COMPRESS_LEVEL" - GUNZIP_CMD="bzip2 -d" - ;; -xz) - GZIP_CMD="xz -c$COMPRESS_LEVEL" - # Must opt-in by specifying a value since not all versions of xz support threads - if test $THREADS -ne $DEFAULT_THREADS; then - GZIP_CMD="$GZIP_CMD --threads=$THREADS" - fi - GUNZIP_CMD="xz -d" - ;; -lzo) - GZIP_CMD="lzop -c$COMPRESS_LEVEL" - GUNZIP_CMD="lzop -d" - ;; -lz4) - GZIP_CMD="lz4 -c$COMPRESS_LEVEL" - GUNZIP_CMD="lz4 -d" - ;; -base64) - GZIP_CMD="base64" - GUNZIP_CMD="base64 --decode -i -" - ;; -gpg) - GZIP_CMD="gpg $GPG_EXTRA -ac -z$COMPRESS_LEVEL" - GUNZIP_CMD="gpg -d" - ENCRYPT="gpg" - ;; -gpg-asymmetric) - GZIP_CMD="gpg $GPG_EXTRA -z$COMPRESS_LEVEL -es" - GUNZIP_CMD="gpg --yes -d" - ENCRYPT="gpg" - ;; -compress) - GZIP_CMD="compress -fc" - GUNZIP_CMD="(type compress >/dev/null 2>&1 && compress -fcd || gzip -cd)" - ;; -none) - GZIP_CMD="cat" - GUNZIP_CMD="cat" - ;; -esac - -if test x"$ENCRYPT" = x"openssl"; then - if test x"$APPEND" = x"y"; then - echo "Appending to existing archive is not compatible with OpenSSL encryption." >&2 - fi - - ENCRYPT_CMD="openssl enc -aes-256-cbc -salt" - DECRYPT_CMD="openssl enc -aes-256-cbc -d" - - if test x"$OPENSSL_NO_MD" != x"y"; then - ENCRYPT_CMD="$ENCRYPT_CMD -md sha256" - DECRYPT_CMD="$DECRYPT_CMD -md sha256" - fi - - if test -n "$PASSWD_SRC"; then - ENCRYPT_CMD="$ENCRYPT_CMD -pass $PASSWD_SRC" - elif test -n "$PASSWD"; then - ENCRYPT_CMD="$ENCRYPT_CMD -pass pass:$PASSWD" - fi -fi - -tmpfile="${TMPDIR:-/tmp}/mkself$$" - -if test -f "$HEADER"; then - oldarchname="$archname" - archname="$tmpfile" - # Generate a fake header to count its lines - SKIP=0 - . "$HEADER" - SKIP=`cat "$tmpfile" |wc -l` - # Get rid of any spaces - SKIP=`expr $SKIP` - rm -f "$tmpfile" - if test "$QUIET" = "n"; then - echo "Header is $SKIP lines long" >&2 - fi - archname="$oldarchname" -else - echo "Unable to open header file: $HEADER" >&2 - exit 1 -fi - -if test "$QUIET" = "n"; then - echo -fi - -if test "$APPEND" = n; then - if test -f "$archname"; then - echo "WARNING: Overwriting existing file: $archname" >&2 - fi -fi - -USIZE=`du $DU_ARGS "$archdir" | awk '{print $1}'` - -if test "." = "$archdirname"; then - if test "$KEEP" = n; then - archdirname="makeself-$$-`date +%Y%m%d%H%M%S`" - fi -fi - -test -d "$archdir" || { echo "Error: $archdir does not exist."; rm -f "$tmpfile"; exit 1; } -if test "$QUIET" = "n"; then - echo "About to compress $USIZE KB of data..." - echo "Adding files to archive named \"$archname\"..." -fi - -# See if we have GNU tar -TAR=`exec <&- 2>&-; which gtar || command -v gtar || type gtar` -test -x "$TAR" || TAR=tar - -tmparch="${TMPDIR:-/tmp}/mkself$$.tar" -( - if test "$APPEND" = "y"; then - tail -n "+$OLDSKIP" "$archname" | eval "$GUNZIP_CMD" > "$tmparch" - fi - cd "$archdir" - # "Determining if a directory is empty" - # https://www.etalabs.net/sh_tricks.html - find . \ - \( \ - ! -type d \ - -o \ - \( -links 2 -exec sh -c ' - is_empty () ( - cd "$1" - set -- .[!.]* ; test -f "$1" && return 1 - set -- ..?* ; test -f "$1" && return 1 - set -- * ; test -f "$1" && return 1 - return 0 - ) - is_empty "$0"' {} \; \ - \) \ - \) -print \ - | LC_ALL=C sort \ - | sed 's/./\\&/g' \ - | xargs $TAR $TAR_EXTRA --format $TAR_FORMAT -$TAR_ARGS "$tmparch" -) || { - echo "ERROR: failed to create temporary archive: $tmparch" - rm -f "$tmparch" "$tmpfile" - exit 1 -} - -USIZE=`du $DU_ARGS "$tmparch" | awk '{print $1}'` - -eval "$GZIP_CMD" <"$tmparch" >"$tmpfile" || { - echo "ERROR: failed to create temporary file: $tmpfile" - rm -f "$tmparch" "$tmpfile" - exit 1 -} -rm -f "$tmparch" - -if test x"$ENCRYPT" = x"openssl"; then - echo "About to encrypt archive \"$archname\"..." - { eval "$ENCRYPT_CMD -in $tmpfile -out ${tmpfile}.enc" && mv -f ${tmpfile}.enc $tmpfile; } || \ - { echo Aborting: could not encrypt temporary file: "$tmpfile".; rm -f "$tmpfile"; exit 1; } -fi - -fsize=`cat "$tmpfile" | wc -c | tr -d " "` - -# Compute the checksums - -shasum=0000000000000000000000000000000000000000000000000000000000000000 -md5sum=00000000000000000000000000000000 -crcsum=0000000000 - -if test "$NOCRC" = y; then - if test "$QUIET" = "n"; then - echo "skipping crc at user request" - fi -else - crcsum=`CMD_ENV=xpg4 cksum < "$tmpfile" | sed -e 's/ /Z/' -e 's/ /Z/' | cut -dZ -f1` - if test "$QUIET" = "n"; then - echo "CRC: $crcsum" - fi -fi - -if test "$SHA256" = y; then - SHA_PATH=`exec <&- 2>&-; which shasum || command -v shasum || type shasum` - if test -x "$SHA_PATH"; then - shasum=`eval "$SHA_PATH -a 256" < "$tmpfile" | cut -b-64` - else - SHA_PATH=`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum` - shasum=`eval "$SHA_PATH" < "$tmpfile" | cut -b-64` - fi - if test "$QUIET" = "n"; then - if test -x "$SHA_PATH"; then - echo "SHA256: $shasum" - else - echo "SHA256: none, SHA command not found" - fi - fi -fi -if test "$NOMD5" = y; then - if test "$QUIET" = "n"; then - echo "Skipping md5sum at user request" - fi -else - # Try to locate a MD5 binary - OLD_PATH=$PATH - PATH=${GUESS_MD5_PATH:-"$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} - MD5_ARG="" - MD5_PATH=`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum` - test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which md5 || command -v md5 || type md5` - test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which digest || command -v digest || type digest` - PATH=$OLD_PATH - if test -x "$MD5_PATH"; then - if test `basename ${MD5_PATH}`x = digestx; then - MD5_ARG="-a md5" - fi - md5sum=`eval "$MD5_PATH $MD5_ARG" < "$tmpfile" | cut -b-32` - if test "$QUIET" = "n"; then - echo "MD5: $md5sum" - fi - else - if test "$QUIET" = "n"; then - echo "MD5: none, MD5 command not found" - fi - fi -fi -if test "$SIGN" = y; then - GPG_PATH=`exec <&- 2>&-; which gpg || command -v gpg || type gpg` - if test -x "$GPG_PATH"; then - SIGNATURE=`$GPG_PATH --pinentry-mode=loopback --batch --yes --passphrase "$GPG_PASSPHRASE" --output - --detach-sig $tmpfile | base64 | tr -d \\\\n` - if test "$QUIET" = "n"; then - echo "Signature: $SIGNATURE" - fi - else - echo "Missing gpg command" >&2 - fi -fi - -totalsize=0 -for size in $fsize; -do - totalsize=`expr $totalsize + $size` -done - -if test "$APPEND" = y; then - mv "$archname" "$archname".bak || exit - - # Prepare entry for new archive - filesizes="$fsize" - CRCsum="$crcsum" - MD5sum="$md5sum" - SHAsum="$shasum" - Signature="$SIGNATURE" - # Generate the header - . "$HEADER" - # Append the new data - cat "$tmpfile" >> "$archname" - - chmod +x "$archname" - rm -f "$archname".bak - if test "$QUIET" = "n"; then - echo "Self-extractable archive \"$archname\" successfully updated." - fi -else - filesizes="$fsize" - CRCsum="$crcsum" - MD5sum="$md5sum" - SHAsum="$shasum" - Signature="$SIGNATURE" - - # Generate the header - . "$HEADER" - - # Append the compressed tar data after the stub - if test "$QUIET" = "n"; then - echo - fi - cat "$tmpfile" >> "$archname" - chmod +x "$archname" - if test "$QUIET" = "n"; then - echo Self-extractable archive \"$archname\" successfully created. - fi -fi -rm -f "$tmpfile" diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh deleted file mode 100644 index 31ee16511..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/makeself/run-tests.sh +++ /dev/null @@ -1,8 +0,0 @@ -#!/bin/sh -# Run every available test - Bash needed -cd test -for test in *test; -do - echo "Running test $test ..." - bash $test || { echo "*** ERROR: Test '$test' failed!"; exit 1; } -done diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh b/atb_operator/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh deleted file mode 100755 index a977bd51d..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/merge_aicpu_info_json.sh +++ /dev/null @@ -1,31 +0,0 @@ -#!/bin/bash - -project_path=$1 -build_path=$2 -vendor_name=customize -echo $@ -if [[ ! -d "$project_path" ]]; then - echo "[ERROR] No projcet path is provided" - exit 1 -fi - -if [[ ! -d "$build_path" ]]; then - echo "[ERROR] No build path is provided" - exit 1 -fi - -if [[ ! -d "$ASCEND_OPP_PATH" ]]; then - echo "[ERROR] No opp install path is provided" - exit 1 -fi -custom_exist_info_json=$ASCEND_OPP_PATH/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json -custom_new_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json -temp_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/temp_cust_aicpu_kernel.json - -if [[ -f "$custom_exist_info_json" ]] && [[ -f "$custom_new_info_json" ]]; then - cp -f $custom_exist_info_json $temp_info_json - chmod +w $temp_info_json - python3 ${project_path}/cmake/util/insert_op_info.py ${custom_new_info_json} ${temp_info_json} - cp -f $temp_info_json $custom_new_info_json - rm -f $temp_info_json -fi diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py deleted file mode 100755 index c38e79edc..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/opdesc_parser.py +++ /dev/null @@ -1,260 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os - - -OP_ALL = '__ALLOP__' -SOC_ALL = '__ALLSOC__' -SOC_TO_SHORT_SOC_MAP = { - "ascend910a": "ascend910", - "ascend910proa": "ascend910", - "ascend910b": "ascend910", - "ascend910prob": "ascend910", - "ascend910premiuma": "ascend910", - "ascend910b1": "ascend910b", - "ascend910b2": "ascend910b", - "ascend910b3": "ascend910b", - "ascend910b4": "ascend910b", - "ascend310p1": "ascend310p", - "ascend310p3": "ascend310p", - "ascend310p3vir01": "ascend310p", - "ascend310p3vir02": "ascend310p", - "ascend310p3vir04": "ascend310p", - "ascend310p3vir08": "ascend310p", - "ascend310b1": "ascend310b", - "bs9sx1aa": "bs9sx1a" -} - - -class OpDesc: - def __init__(self: any, op_type: str): - self.op_type = op_type - self.attr_list = [] - self.attr_val = {} - self.input_name = [] - self.input_type = [] - self.input_dtype = [] - self.input_fmt = [] - self.output_name = [] - self.output_type = [] - self.output_dtype = [] - self.output_fmt = [] - self.op_fmt_sel = False - self.op_chk_support = False - self.op_intf = '' - self.kern_name = '' - self.op_file = '' - self.op_replay_flag = False - self.op_replay_batch = False - self.input_idx = -1 - self.output_idx = -1 - self.max_block_dim = 32 - self.max_shape_size = 268435456 - self.dynamic_shape = False - self.op_range_limit = '' - self.custom_compile_options = {} - self.custom_all_compile_options = {} - - @staticmethod - def _parse_digit(conf: str) -> int: - return int(conf.split('=')[1]) - - @staticmethod - def _parse_flag(conf: str) -> bool: - if 'true' == conf.split('=')[1]: - return True - return False - - @staticmethod - def _parse_str(conf: str) -> str: - return conf.split('=')[1] - - @staticmethod - def _parse_list(conf: str) -> list: - return conf.split('=')[1].split(',') - - def parse_input(self: any, conf: str): - if conf.startswith('input{}.name'.format(int(self.input_idx) + 1)): - self.input_idx += 1 - self.input_name.append(self._parse_str(conf)) - elif conf.startswith('input{}.paramType'.format(int(self.input_idx))): - self.input_type.append(self._parse_str(conf)) - elif conf.startswith('input{}.dtype'.format(int(self.input_idx))): - self.input_dtype.append(self._parse_str(conf)) - elif conf.startswith('input{}.format'.format(int(self.input_idx))): - self.input_fmt.append(self._parse_str(conf)) - else: - return - - def parse_output(self: any, conf: str): - if conf.startswith('output{}.name'.format(int(self.output_idx) + 1)): - self.output_idx += 1 - self.output_name.append(self._parse_str(conf)) - elif conf.startswith('output{}.paramType'.format(int(self.output_idx))): - self.output_type.append(self._parse_str(conf)) - elif conf.startswith('output{}.dtype'.format(int(self.output_idx))): - self.output_dtype.append(self._parse_str(conf)) - elif conf.startswith('output{}.format'.format(int(self.output_idx))): - self.output_fmt.append(self._parse_str(conf)) - else: - return - - def parse_op_format(self: any, conf: str): - self.op_fmt_sel = self._parse_flag(conf) - - def parse_check_support(self: any, conf: str): - self.op_chk_support = self._parse_flag(conf) - - def parse_range_limit(self: any, conf: str): - self.op_range_limit = self._parse_str(conf) - - def parse_kern_name(self: any, conf: str): - self.kern_name = self._parse_str(conf) - - def parse_op_intf(self: any, conf: str): - self.op_intf = self._parse_str(conf) - - def parse_op_file(self: any, conf: str): - self.op_file = self._parse_str(conf) - - def parse_dynamic_shape(self: any, conf: str): - self.dynamic_shape = self._parse_flag(conf) - - def parse_attr_list(self: any, conf: str): - self.attr_list = self._parse_list(conf) - - def parse_attr_val(self: any, conf: str): - for attr in self.attr_list: - if self.attr_val.get(attr) is None: - self.attr_val[attr] = {} - if conf.startswith('attr_{}.type'.format(attr)): - self.attr_val.get(attr)['type'] = self._parse_str(conf) - elif conf.startswith('attr_{}.paramType'.format(attr)): - self.attr_val.get(attr)['paramType'] = self._parse_str(conf) - elif conf.startswith('attr_{}.defaultValue'.format(attr)): - self.attr_val.get(attr)['defaultValue'] = self._parse_str(conf) - - def parse_replay_val(self: any, batch_list: list, iterator_list: list): - if self.op_type in batch_list: - self.op_replay_flag = True - self.op_replay_batch = True - elif self.op_type in iterator_list: - self.op_replay_flag = True - self.op_replay_batch = False - - -def _is_op_type_in_opdesc(op_descs: list, op_type: str): - for op in op_descs: - if op_type == op.op_type: - return True - return False - - -def _set_all_options_to_opdescs(op_descs, soc_ver_compile_options): - for op in op_descs: - op.custom_all_compile_options = soc_ver_compile_options - - -def _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options): - for op in op_descs: - if op.op_type != op_type: - continue - op.custom_compile_options = soc_ver_compile_options - - -def _trans_soc_ver_to_short(soc_ver: str): - low_soc_ver = soc_ver.lower() - if low_soc_ver not in SOC_TO_SHORT_SOC_MAP: - print(f'WARNING: caution: {soc_ver} will trans into ascend910, if not your intention,' - f'use ascend910b1~4 instead') - return SOC_TO_SHORT_SOC_MAP[low_soc_ver] - - -def _get_op_custom_options(op_descs: list, auto_gen_dir: str): - if auto_gen_dir is None: - return {} - file = os.path.join(auto_gen_dir, "custom_compile_options.ini") - if not os.path.exists(file): - print(f'WARNING: cannot find {auto_gen_dir}/custom_compile_options.ini') - return {} - with open (file, 'r') as fd: - lines = fd.readlines() - for line in lines: - param_list = str.split(line.rstrip('\n'), ',') - if len(param_list) != 3: - raise Exception(f'ERROR: custom compile option {param_list} len is not 3') - op_type = param_list[0] - if op_type.upper() == 'ALL': - op_type = OP_ALL - if op_type != OP_ALL and _is_op_type_in_opdesc(op_descs, op_type) == False: - print(f'WARNING: op: {op_type} are not exists in this project') - continue - soc_ver_compile_options = {} - soc_ver = param_list[1] - options_str = param_list[2] - options = str.split(options_str, ';') - if soc_ver == '': - soc_ver_compile_options[SOC_ALL] = options - else: - soc_ver_list = str.split(soc_ver, ';') - for ver in soc_ver_list: - short_ver = _trans_soc_ver_to_short(ver) - soc_ver_compile_options[short_ver] = options - if op_type == OP_ALL: - _set_all_options_to_opdescs(op_descs, soc_ver_compile_options) - else: - _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options) - - -def get_op_desc(file: str, batch_list: list, iterator_list: list, builder: any, - op_type: list, auto_gen_dir: str = None) -> list: - op_descs = [] - op_match = False - with open (file, 'r') as fd: - lines = fd.readlines() - for line in lines: - line = line.strip() - if line.startswith('['): - name = line[1:-1] - if op_type is None or name in op_type: - op_match = True - op_desc = builder(name) - op_desc.parse_replay_val(batch_list, iterator_list) - op_descs.append(op_desc) - else: - op_match = False - if op_type is not None and len(op_descs) == len(op_type): - return op_descs - continue - if not op_match: - continue - if line.startswith('input'): - op_desc.parse_input(line) - elif line.startswith('output'): - op_desc.parse_output(line) - elif line.startswith('dynamicFormat.flag'): - op_desc.parse_op_format(line) - elif line.startswith('needCheckSupport.flag'): - op_desc.parse_check_support(line) - elif line.startswith('rangeLimit.value'): - op_desc.parse_range_limit(line) - elif line.startswith('opInterface.value'): - op_desc.parse_op_intf(line) - elif line.startswith('kernel.name'): - op_desc.parse_kern_name(line) - elif line.startswith('opFile.value'): - op_desc.parse_op_file(line) - elif line.startswith('dynamicShapeSupport.flag'): - op_desc.parse_dynamic_shape(line) - elif line.startswith('attr.list'): - op_desc.parse_attr_list(line) - elif line.startswith('attr_'): - op_desc.parse_attr_val(line) - _get_op_custom_options(op_descs, auto_gen_dir) - return op_descs diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py deleted file mode 100755 index 721465fee..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/parse_ini_to_json.py +++ /dev/null @@ -1,338 +0,0 @@ -# Copyright 2020-2021 Huawei Technologies Co., Ltd -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -""" -parser ini to json -""" - -import json -import os -import stat -import sys - - -ATTR_TYPE_LIST = ["int", "float", "bool", "str", "listInt", "listFloat", "listBool", "listStr", "listListInt", - "type", "listType", "tensor", "listTensor"] -ATTR_PARAMTYPE_LIST = ["optional", "required"] -BOOL_FLAG_KEY = ["dynamicFormat", "dynamicShapeSupport", "dynamicRankSupport", "precision_reduce", "heavyOp", - "needCheckSupport"] -BOOL_LIST = ["true", "false"] -DTYPE_LIST = ["float16", "float", "float32", "int8", "int16", "int32", "uint8", "uint16", "uint32", "bool", - "int64", "uint64", "qint8", "qint16", "qint32", "quint8", "quint16", "double", "complex64", - "complex128", "string", "resource", "dual", "dual_sub_int8", "dual_sub_uint8", "string_ref", - "int4", "bfloat16", "uint1"] -FORMAT_LIST = ["NCHW", "NHWC", "ND", "NC1HWC0", "FRACTAL_Z", "NC1C0HWPAD", "NHWC1C0", "FSR_NCHW", "FRACTAL_DECONV", - "C1HWNC0", "FRACTAL_DECONV_TRANSPOSE", "FRACTAL_DECONV_SP_STRIDE_TRANS", "NC1HWC0_C04", - "FRACTAL_Z_C04", "CHWN", "FRACTAL_DECONV_SP_STRIDE8_TRANS", "HWCN", "NC1KHKWHWC0", "BN_WEIGHT", - "FILTER_HWCK", "HASHTABLE_LOOKUP_LOOKUPS", "HASHTABLE_LOOKUP_KEYS", "HASHTABLE_LOOKUP_VALUE", - "HASHTABLE_LOOKUP_OUTPUT", "HASHTABLE_LOOKUP_HITS", "C1HWNCoC0", "MD", "NDHWC", "FRACTAL_ZZ", - "FRACTAL_NZ", "NCDHW", "DHWCN", "NDC1HWC0", "FRACTAL_Z_3D", "CN", "NC", "DHWNC", - "FRACTAL_Z_3D_TRANSPOSE", "FRACTAL_ZN_LSTM", "FRACTAL_ZN_RNN", "FRACTAL_Z_G", "NULL"] - - -def parse_ini_files(ini_files): - """ - parse ini files to json - Parameters: - ---------------- - ini_files:input file list - return:ops_info - ---------------- - """ - tbe_ops_info = {} - for ini_file in ini_files: - check_file_size(ini_file) - parse_ini_to_obj(ini_file, tbe_ops_info) - return tbe_ops_info - - -def check_file_size(input_file): - try: - file_size = os.path.getsize(input_file) - except OSError as os_error: - print('[ERROR] Failed to open "%s". %s' % (input_file, str(os_error))) - raise OSError from os_error - if file_size > 10*1024*1024: - print('[WARN] The size of %s exceeds 10MB, it may take more time to run, please wait.' % input_file) - - -def parse_ini_to_obj(ini_file, tbe_ops_info): - """ - parse ini file to json obj - Parameters: - ---------------- - ini_file:ini file path - tbe_ops_info:ops_info - ---------------- - """ - with open(ini_file) as ini_file: - lines = ini_file.readlines() - op_dict = {} - op_name = "" - find_op_type = False - for line in lines: - line = line.rstrip() - if line == "": - continue - if line.startswith("["): - if line.endswith("]"): - op_name = line[1:-1] - op_dict = {} - tbe_ops_info[op_name] = op_dict - find_op_type = True - elif "=" in line: - key1 = line[:line.index("=")] - key2 = line[line.index("=")+1:] - key1_0, key1_1 = key1.split(".") - if key1_0 not in op_dict: - op_dict[key1_0] = {} - if key1_1 in op_dict.get(key1_0): - raise RuntimeError("Op:" + op_name + " " + key1_0 + " " + - key1_1 + " is repeated!") - dic_key = op_dict.get(key1_0) - dic_key[key1_1] = key2 - else: - continue - if not find_op_type: - raise RuntimeError("Not find OpType in .ini file.") - - -def check_output_exist(op_dict, is_valid): - """ - Function Description: - Check output is exist - Parameter: op_dict - Parameter: is_valid - """ - if "output0" in op_dict: - output0_dict = op_dict.get("output0") - if output0_dict.get("name", None) is None: - is_valid = False - print("output0.name is required in .ini file!") - else: - is_valid = False - print("output0 is required in .ini file!") - return is_valid - - -def check_attr_dict(attr_dict, is_valid, attr): - """ - Function Description: - Check attr_dict - Parameter: attr_dict - Parameter: is_valid - Parameter: attr - """ - attr_type = attr_dict.get("type") - value = attr_dict.get("value") - param_type = attr_dict.get("paramType") - if attr_type is None or value is None: - is_valid = False - print("If attr.list is exist, {0}.type and {0}.value is required".format(attr)) - if param_type and param_type not in ATTR_PARAMTYPE_LIST: - is_valid = False - print("{0}.paramType only support {1}.".format(attr, ATTR_PARAMTYPE_LIST)) - if attr_type and attr_type not in ATTR_TYPE_LIST: - is_valid = False - print("{0}.type only support {1}.".format(attr, ATTR_TYPE_LIST)) - return is_valid - - -def check_attr(op_dict, is_valid): - """ - Function Description: - Check attr - Parameter: op_dict - Parameter: is_valid - """ - if "attr" in op_dict: - attr_dict = op_dict.get("attr") - attr_list_str = attr_dict.get("list", None) - if attr_list_str is None: - is_valid = False - print("attr.list is required in .ini file!") - else: - attr_list = attr_list_str.split(",") - for attr_name in attr_list: - attr = "attr_" + attr_name.strip() - attr_dict = op_dict.get(attr) - if attr_dict: - is_valid = check_attr_dict(attr_dict, is_valid, attr) - else: - is_valid = False - print("%s is required in .ini file, when attr.list is %s!" % (attr, attr_list_str)) - return is_valid - - -def check_bool_flag(op_dict, is_valid): - """ - Function Description: - check_bool_flag - Parameter: op_dict - Parameter: is_valid - """ - for key in BOOL_FLAG_KEY: - if key in op_dict: - op_bool_key = op_dict.get(key) - if op_bool_key.get("flag").strip() not in BOOL_LIST: - is_valid = False - print("{0}.flag only support {1}.".format(key, BOOL_LIST)) - return is_valid - - -def check_type_format(op_info, is_valid, op_info_key): - """ - Function Description: - Check type and format - Parameter: op_info - Parameter: is_valid - Parameter: op_info_key - """ - op_info_dtype_str = op_info.get("dtype") - op_info_dtype_num = 0 - op_info_format_num = 0 - if op_info_dtype_str: - op_info_dtype = op_info_dtype_str.split(",") - op_info_dtype_num = len(op_info_dtype) - for dtype in op_info_dtype: - if dtype.strip() not in DTYPE_LIST: - is_valid = False - print("{0}.dtype not support {1}.".format(op_info_key, dtype)) - op_info_format_str = op_info.get("format") - if op_info_format_str: - op_info_format = op_info_format_str.split(",") - op_info_format_num = len(op_info_format) - for op_format in op_info_format: - if op_format.strip() not in FORMAT_LIST: - is_valid = False - print("{0}.format not support {1}.".format(op_info_key, op_format)) - if op_info_dtype_num > 0 and op_info_format_num > 0: - if op_info_dtype_num != op_info_format_num: - is_valid = False - print("The number of {0}.dtype not match the number of {0}.format.".format(op_info_key)) - return is_valid - - -def check_op_info(tbe_ops): - """ - Function Description: - Check info. - Parameter: tbe_ops - Return Value: is_valid - """ - print("\n\n==============check valid for ops info start==============") - required_op_input_info_keys = ["paramType", "name"] - required_op_output_info_keys = ["paramType", "name"] - param_type_valid_value = ["dynamic", "optional", "required"] - is_valid = True - for op_key in tbe_ops: - op_dict = tbe_ops[op_key] - for op_info_key in op_dict: - if op_info_key.startswith("input"): - op_input_info = op_dict[op_info_key] - missing_keys = [] - for required_op_input_info_key in required_op_input_info_keys: - if required_op_input_info_key not in op_input_info: - missing_keys.append(required_op_input_info_key) - if len(missing_keys) > 0: - print("op: " + op_key + " " + op_info_key + " missing: " + - ",".join(missing_keys)) - is_valid = False - else: - if not op_input_info["paramType"] in param_type_valid_value: - print("op: " + op_key + " " + op_info_key + \ - " paramType not valid, valid key:[dynamic, " - "optional, required]") - is_valid = False - is_valid = check_type_format(op_input_info, is_valid, op_info_key) - if op_info_key.startswith("output"): - op_input_info = op_dict[op_info_key] - missing_keys = [] - for required_op_input_info_key in required_op_output_info_keys: - if required_op_input_info_key not in op_input_info: - missing_keys.append(required_op_input_info_key) - if len(missing_keys) > 0: - print("op: " + op_key + " " + op_info_key + " missing: " + - ",".join(missing_keys)) - is_valid = False - else: - if not op_input_info["paramType"] in param_type_valid_value: - print("op: " + op_key + " " + op_info_key + - " paramType not valid, valid key:[dynamic, " - "optional, required]") - is_valid = False - is_valid = check_type_format(op_input_info, is_valid, op_info_key) - is_valid = check_attr(op_dict, is_valid) - is_valid = check_bool_flag(op_dict, is_valid) - print("==============check valid for ops info end================\n\n") - return is_valid - - -def write_json_file(tbe_ops_info, json_file_path): - """ - Save info to json file - Parameters: - ---------------- - tbe_ops_info: ops_info - json_file_path: json file path - ---------------- - """ - json_file_real_path = os.path.realpath(json_file_path) - wr_flag = os.O_WRONLY | os.O_CREAT - wr_mode = stat.S_IWUSR | stat.S_IRUSR - with os.fdopen(os.open(json_file_real_path, wr_flag, wr_mode), 'w') as file_path: - # Only the owner and group have rights - os.chmod(json_file_real_path, stat.S_IWGRP + stat.S_IWUSR + stat.S_IRGRP - + stat.S_IRUSR) - json.dump(tbe_ops_info, file_path, sort_keys=True, indent=4, - separators=(',', ':')) - print("Compile op info cfg successfully.") - - -def parse_ini_to_json(ini_file_paths, outfile_path): - """ - parse ini files to json file - Parameters: - ---------------- - ini_file_paths: list of ini file path - outfile_path: output file path - ---------------- - """ - tbe_ops_info = parse_ini_files(ini_file_paths) - if not check_op_info(tbe_ops_info): - print("Compile op info cfg failed.") - return False - write_json_file(tbe_ops_info, outfile_path) - return True - - -if __name__ == '__main__': - args = sys.argv - - OUTPUT_FILE_PATH = "tbe_ops_info.json" - ini_file_path_list = [] - - for arg in args: - if arg.endswith("ini"): - ini_file_path_list.append(arg) - OUTPUT_FILE_PATH = arg.replace(".ini", ".json") - if arg.endswith("json"): - OUTPUT_FILE_PATH = arg - - if len(ini_file_path_list) == 0: - ini_file_path_list.append("tbe_ops_info.ini") - - if not parse_ini_to_json(ini_file_path_list, OUTPUT_FILE_PATH): - sys.exit(1) - sys.exit(0) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/preset_parse.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/preset_parse.py deleted file mode 100755 index 8f1124b1d..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/preset_parse.py +++ /dev/null @@ -1,23 +0,0 @@ -import json -import sys -import os - - -def get_config_opts(file): - src_dir = os.path.abspath(os.path.dirname(file)) - opts = '' - with open(file, 'r') as fd: - config = json.load(fd) - for conf in config: - if conf == 'configurePresets': - for node in config[conf]: - macros = node.get('cacheVariables') - if macros is not None: - for key in macros: - opts += '-D{}={} '.format(key, macros[key]['value']) - opts = opts.replace('${sourceDir}', src_dir) - print(opts) - - -if __name__ == "__main__": - get_config_opts(sys.argv[1]) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py deleted file mode 100755 index 1baa364ef..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_codegen.py +++ /dev/null @@ -1,105 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import os -import stat -import collections -import kernel_entry as keb -from tiling_data_def_build import gen_tiling -import code_channel_infer -import const_var - -PYF_PATH = os.path.dirname(__file__) - -ReplayCodeGenParams = collections.namedtuple('ReplayCodeGenParams',\ -['op_type', 'impl', 'tiling_file', 'kernel', 'entry', 'argn', 'op_replay_batch', 'max_block_dim', 'max_shape_size']) - - -class ReplayCodeGen: - def __init__(self, replayCodeGenParams): - self.op_type = replayCodeGenParams.op_type - self.impl = replayCodeGenParams.impl - self.tiling_file = replayCodeGenParams.tiling_file - self.tiling_data_file = '' - self.kernel = replayCodeGenParams.kernel - self.entry = replayCodeGenParams.entry - self.argn = replayCodeGenParams.argn - self.batch = False - self.outdir = '' - self.data_type = 'uint8_t' - self.blknum = 32 - self.op_replay_batch = replayCodeGenParams.op_replay_batch - self.max_block_dim = replayCodeGenParams.max_block_dim - self.max_shape_size = replayCodeGenParams.max_shape_size - - def set_batch(self, is_batch): - self.batch = is_batch - - def set_outdir(self, outdir): - self.outdir = outdir - - def gen_replay(self, ops_product: str): - kerentry = os.path.join(self.outdir, self.kernel + '_entry.cce') - kerimpl = os.path.join(self.outdir, self.kernel + '_impl.cpp') - replayimpl = os.path.join(self.outdir, self.kernel + '_replay.cpp') - if self.batch: - reptmp = os.path.join(PYF_PATH, 'batch_replay_impl.temp') - else: - reptmp = os.path.join(PYF_PATH, 'replay_impl.temp') - kertmp = os.path.join(PYF_PATH, 'kernel_impl.temp') - self._gen_kentry(kerentry) - self._gen_kimpl_code(kerimpl, kertmp) - self._gen_tiling_data_header() - self._gen_replay_code(replayimpl, reptmp, ops_product) - - def _gen_tiling_data_header(self): - self.tiling_data_file = os.path.join(self.outdir, self.kernel + '_tiling_data.h') - gen_tiling(self.tiling_file, self.tiling_data_file) - - def _gen_kimpl_code(self, src, tmpfile): - with open(tmpfile, 'r') as fd: - temp = fd.read() - temp = temp.replace('__CCE_FILE__', self.impl) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(temp) - - def _gen_replay_code(self, src, tmpfile, ops_product: str): - with open(tmpfile, 'r') as fd: - temp = fd.read() - temp = temp.replace('__ARG_NUM__', str(self.argn)) - argdef = [] - kargs = [] - for i in range(0, self.argn): - argdef.append('{} *'.format(self.data_type)) - kargs.append('({} *)GetArg({})'.format(self.data_type, i)) - temp = temp.replace('__ARGS_DEF__', ', '.join(argdef)) - temp = temp.replace('__KERNEL_ARGS__', ', '.join(kargs)) - temp = temp.replace('__KERNEL_FUN__', self.entry) - core_type_infer = 'core_type' - code_channel = code_channel_infer.infer_code_channel(code_channel_infer.InfoCodeChanelParams(self.impl,\ - self.tiling_data_file, self.kernel, self.outdir, ops_product, None)) - if code_channel == code_channel_infer.CODE_VEC: - core_type_infer = '0' - elif code_channel == code_channel_infer.CODE_CUBE: - core_type_infer = '1' - temp = temp.replace('__CORE_TYPE__', core_type_infer) - # regist function - temp = temp.replace('__OPS_PRODUCT__', ops_product) - temp = temp.replace('__OPTYPE__', self.op_type) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(temp) - - def _gen_kentry(self, src): - kf = '' - pre_alloc_str = 'A' * 256 - if self.batch: - kf += keb.batch_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str), self.argn, self.data_type) - else: - kf += keb.mc_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str),\ - self.argn, self.data_type, self.blknum) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(kf) diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp b/atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp deleted file mode 100755 index 1d30dd865..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/replay_impl.temp +++ /dev/null @@ -1,120 +0,0 @@ -#include -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#include "register/op_check.h" -#define __ASCENDC_REPLAY_CODE__ -using namespace std; -using namespace optiling; -using namespace AscendCReplay; - -extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); -extern "C" int elf_append(char *elf, uint32_t elfSize, char *jit, int kernum, int blknum[], char *atext[], - int alen[], int atlen, const char* kernelname[]); - -#define KERNEL_N 1 -#define ARG_N (__ARG_NUM__) -#define MAX_L (1024 * 1024 * 100) -#define MAX_E (1024 * 1024) - -int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) -{ - // gen type 1 : direct call codes 0: load .o file - if (param.gentype < 0 || param.gentype > 1) { - printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); - return 0; - } else if (param.gentype == 1 && param.objptr == nullptr) { - printf("Error: call replay with direct call mode, but code obj addr is null\n"); - return 0; - } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { - printf("Error: call replay with object file mode, but object file path is null\n"); - return 0; - } - // core_type 0:MIX 1:CUBE 2:VEC - if (core_type < 0 || core_type > 2) { - printf("Error: call replay core type is %d !\n", core_type); - return 0; - } - g_coreType = __CORE_TYPE__; - g_taskRation = param.task_ration; - g_tilingKey = param.tiling_key; - - unsigned char *buf, *jit; - char *kernel[KERNEL_N * 32]; - int len[KERNEL_N * 32]; - int blknum[KERNEL_N]; - int max; - block_num = param.block_dim; - g_ubBase = block_num; - uint8_t *code = (uint8_t *)malloc(MAX_L); - uint8_t *pos = code; - struct timespec tp1, tp2; - - clock_gettime(CLOCK_MONOTONIC, &tp1); - if (block_num > 32) { - printf("Error: block_num > 32\n"); - return 0; - } - //__OP_FOPEN__ - for (int i = 0; i < KERNEL_N; i++) { - for (int j = 0; j < ARG_N; j++) - AddArg(j, ARG_STEP * (j + 1)); - for (block_idx = 0; block_idx < block_num; block_idx++) { - //__OP_SET_KERNEL__ - int code_idx = i * block_num + block_idx; -#ifdef FP_CEILING - SetCtrlFloatEnable(); -#else - SetCtrlFloatDisable(); -#endif - CodeInit(pos, false); - __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); - CodeEnd(); - kernel[code_idx] = (char *)pos; - len[code_idx] = CodeLen(); - pos += len[code_idx]; - printf("kernel %d core %ld code generated len %d\n", i, block_idx, len[code_idx]); - } - blknum[i] = block_num; - } - //__OP_FCLOSE__ - clock_gettime(CLOCK_MONOTONIC, &tp2); - buf = (unsigned char *)malloc(MAX_E); - int fd = open(param.entry_file, O_RDONLY); - if (fd < 0) { - printf("[error]: cannot find entry.o : %s\n", param.entry_file); - return 0; - } - uint32_t bufSize = read(fd, buf, MAX_E); - if (bufSize <= 0) { - printf("[error]: entry.o : %s is too small ! \n", param.entry_file); - } - close(fd); - jit = (unsigned char *)malloc(MAX_L); - printf("total code generated %ld\n", pos - code); - int sz = elf_append((char *)buf, bufSize, (char *)jit, KERNEL_N, blknum, kernel, len, pos - code, ¶m.kernel_name); - if (tp1.tv_sec != tp2.tv_sec) { - printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); - } else { - printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); - } - printf("new elf size %d\n", sz); - if (param.gentype == 0) { - fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); - (void)write(fd, jit, sz); - close(fd); - free(jit); - } else if (param.gentype == 1) { - *param.objptr = (char*)jit; - } - free(buf); - free(code); - return sz; -} - -REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py b/atb_operator/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py deleted file mode 100755 index a96304261..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/cmake/util/tiling_data_def_build.py +++ /dev/null @@ -1,84 +0,0 @@ -#!/usr/bin/env python -# coding=utf-8 -""" -Function: -The replay funtion entry -Copyright Information: -Huawei Technologies Co., Ltd. All Rights Reserved © 2020 -""" - -import sys -import os -import stat -import re -import const_var - - -def gen_tiling(tiling_header_file: str, tiling_file_out: str): - if not os.path.exists(tiling_header_file): - print("warning: no userdef tiling header file: ", tiling_header_file) - return - print("generate tiling def header file: ", tiling_file_out) - tmp_name = os.path.splitext(os.path.basename(tiling_header_file))[0].upper() - tiling_source = '#ifndef __{}_H__\n'.format(tmp_name) - tiling_source += '#define __{}_H__\n\n'.format(tmp_name) - tiling_source += '#include \n' - tiling_source += '#include \n\n' - tiling_source += '#include "kernel_tiling/kernel_tiling.h"\n\n' - end_source = "" - pattern = re.compile(r'[(](.*)[)]', re.S) - with open(tiling_header_file, 'r') as fd: - lines = fd.readlines() - for line in lines: - line = line.strip() - if (line.startswith('BEGIN_TILING_DATA_DEF')): - tiling_source += '#pragma pack(1)\n' - tiling_source += 'struct ' - struct_def = re.findall(pattern, line)[0] - tiling_source += struct_def + ' {\n' - elif (line.startswith('TILING_DATA_FIELD_DEF_ARR')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {}[{}] = {{}};\n'.format(fds[0].strip(), fds[2].strip(), fds[1].strip()) - elif (line.startswith('TILING_DATA_FIELD_DEF_STRUCT')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {};\n'.format(fds[0].strip(), fds[1].strip()) - elif (line.startswith('TILING_DATA_FIELD_DEF')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {} = 0;\n'.format(fds[0].strip(), fds[1].strip()) - elif (line.startswith('END_TILING_DATA_DEF')): - tiling_source += '};\n' - tiling_source += '#pragma pack()\n\n' - tiling_source += '#ifdef __NPU_TILING__\n' - tiling_source += \ - 'inline [aicore] void Init{stru}(const __gm__ uint8_t* tiling, {stru}* const_data)\n'\ - .format(stru=struct_def) - tiling_source += '{\n' - tiling_source += ' const __gm__ uint32_t *src = (const __gm__ uint32_t *)tiling;\n' - tiling_source += ' uint32_t *dst = (uint32_t *)const_data;\n' - tiling_source += ' for (auto i = 0; i < sizeof({}) / 4; i++) *(dst + i) = *(src + i);\n'\ - .format(struct_def) - tiling_source += '}\n' - tiling_source += '#else\n' - tiling_source += 'inline void Init{stru}(uint8_t* tiling, {stru}* const_data)\n'.format(stru=struct_def) - tiling_source += '{\n' - tiling_source += ' memcpy(const_data, tiling, sizeof({}));\n'.format(struct_def) - tiling_source += '}\n' - tiling_source += '#endif\n\n' - end_source = ''' -#define GET_TILING_DATA(tiling_data, tiling_arg) \\ -{stru} tiling_data; \\ -Init{stru}(tiling_arg, &tiling_data)\n -'''.format(stru=struct_def) - tiling_source += end_source - tiling_source += '#endif' - with os.fdopen(os.open(tiling_file_out, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(tiling_source) - - -if __name__ == '__main__': - if len(sys.argv) <= 2: - raise RuntimeError('arguments must greater than 2') - gen_tiling(sys.argv[1], sys.argv[2]) diff --git a/atb_operator/AtbAddSample/AddAscendC/framework/CMakeLists.txt b/atb_operator/AtbAddSample/AddAscendC/framework/CMakeLists.txt deleted file mode 100644 index b6be9b492..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/framework/CMakeLists.txt +++ /dev/null @@ -1,11 +0,0 @@ -if(NOT EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/mindspore") - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/caffe_plugin") - add_subdirectory(caffe_plugin) - endif() - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/tf_plugin") - add_subdirectory(tf_plugin) - endif() - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/onnx_plugin") - add_subdirectory(onnx_plugin) - endif() -endif() diff --git a/atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt b/atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt deleted file mode 100644 index a6aba5c20..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/CMakeLists.txt +++ /dev/null @@ -1,14 +0,0 @@ - -aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} plugin_srcs) -add_library(cust_tf_parsers SHARED ${plugin_srcs}) -target_compile_definitions(cust_tf_parsers PRIVATE google=ascend_private) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_tf_parsers PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_tf_parsers PRIVATE intf_pub graph) -install(TARGETS cust_tf_parsers - LIBRARY DESTINATION packages/vendors/${vendor_name}/framework/tensorflow -) diff --git a/atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc b/atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc deleted file mode 100644 index 2cd837ce5..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/framework/tf_plugin/tensorflow_add_custom_plugin.cc +++ /dev/null @@ -1,23 +0,0 @@ -/* Copyright (C) 2020-2021. Huawei Technologies Co., Ltd. All -rights reserved. - * - * This program is free software; you can redistribute it and/or modify - * it under the terms of the Apache License Version 2.0. - * You may not use this file except in compliance with the License. - * - * This program is distributed in the hope that it will be useful, - * but WITHOUT ANY WARRANTY; without even the implied warranty of - * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the - * Apache License for more details at - * http://www.apache.org/licenses/LICENSE-2.0 - */ - -#include "register/register.h" - -namespace domi { -// register op info to GE -REGISTER_CUSTOM_OP("AddCustom") - .FrameworkType(TENSORFLOW) // type: CAFFE, TENSORFLOW - .OriginOpType("Add") // name in tf module - .ParseParamsByOperatorFn(AutoMappingByOpFn); -} // namespace domi diff --git a/atb_operator/AtbAddSample/AddAscendC/op_host/CMakeLists.txt b/atb_operator/AtbAddSample/AddAscendC/op_host/CMakeLists.txt deleted file mode 100644 index 40dd51cfa..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/op_host/CMakeLists.txt +++ /dev/null @@ -1,82 +0,0 @@ - -aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} ops_srcs) - -opbuild(OPS_SRC ${ops_srcs} - OUT_DIR ${ASCEND_AUTOGEN_PATH} -) - -add_library(cust_op_proto SHARED ${ops_srcs} ${ASCEND_AUTOGEN_PATH}/op_proto.cc) -target_compile_definitions(cust_op_proto PRIVATE OP_PROTO_LIB) -target_compile_options(cust_op_proto PRIVATE - -fvisibility=hidden -) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_op_proto PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_op_proto PRIVATE - intf_pub - exe_graph - register - tiling_api - -Wl,--whole-archive - rt2_registry - -Wl,--no-whole-archive -) -set_target_properties(cust_op_proto PROPERTIES OUTPUT_NAME - cust_opsproto_rt2.0 -) -add_library(cust_optiling SHARED ${ops_srcs}) -target_compile_definitions(cust_optiling PRIVATE OP_TILING_LIB) -target_compile_options(cust_optiling PRIVATE - -fvisibility=hidden -) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_optiling PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_optiling PRIVATE - intf_pub - exe_graph - register - tiling_api - -Wl,--whole-archive - rt2_registry - -Wl,--no-whole-archive -) -set_target_properties(cust_optiling PROPERTIES OUTPUT_NAME - cust_opmaster_rt2.0 -) - -file(GLOB aclnn_src ${ASCEND_AUTOGEN_PATH}/aclnn_*.cpp) -file(GLOB aclnn_inc ${ASCEND_AUTOGEN_PATH}/aclnn_*.h) -add_library(cust_opapi SHARED ${aclnn_src}) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_opapi PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_opapi PRIVATE intf_pub ascendcl nnopbase) - -add_custom_target(optiling_compat ALL - COMMAND ln -sf lib/linux/${CMAKE_SYSTEM_PROCESSOR}/$ - ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so -) - -install(TARGETS cust_op_proto - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_proto/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) -install(FILES ${ASCEND_AUTOGEN_PATH}/op_proto.h - DESTINATION packages/vendors/${vendor_name}/op_proto/inc) -install(TARGETS cust_optiling - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) -install(FILES ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so - DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling) -install(TARGETS cust_opapi - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_api/lib) -install(FILES ${aclnn_inc} - DESTINATION packages/vendors/${vendor_name}/op_api/include) diff --git a/atb_operator/AtbAddSample/AddAscendC/op_host/add_custom.cpp b/atb_operator/AtbAddSample/AddAscendC/op_host/add_custom.cpp deleted file mode 100644 index b61a67f11..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/op_host/add_custom.cpp +++ /dev/null @@ -1,117 +0,0 @@ -#include "add_custom_tiling.h" -#include "register/op_def_registry.h" -#include "graph/utils/type_utils.h" -#include "tiling/platform/platform_ascendc.h" - -namespace optiling { -const uint32_t BLOCK_SIZE = 32; -const uint32_t BUFFER_NUM = 2; -static ge::graphStatus TilingFunc(gert::TilingContext* context) -{ - TilingData tiling; - uint64_t ubSize; - auto ascendcPlatform = platform_ascendc::PlatformAscendC(context->GetPlatformInfo()); - ascendcPlatform.GetCoreMemSize(platform_ascendc::CoreMemType::UB, ubSize); - auto coreNum = ascendcPlatform.GetCoreNum(); - - // Based on the input length and the number of inputs, the number of bytes of the input data type is obtained - uint32_t inputNum = context->GetInputShape(0)->GetStorageShape().GetShapeSize(); - uint32_t typeLength = 0; - ge::TypeUtils::GetDataTypeLength(context->GetInputDesc(0)->GetDataType(), typeLength); - uint32_t inputLength = inputNum * typeLength; - uint32_t inputBytes = inputLength / inputNum; - - // There are a total of 3 shared UB spaces in the input and output. If it's int8, there are 2 more TBUFs - uint32_t ubDataNumber = (inputBytes == 1) ? 5 : 3; - // The number of 32B data blocks that can be used for each data. DOUBLE BUFFER is already counted here - uint32_t tileBlockNum = (ubSize / BLOCK_SIZE / BUFFER_NUM) / ubDataNumber; - uint32_t tileDataNum = (tileBlockNum * BLOCK_SIZE) / inputBytes; - - // Input data for 32B alignment - uint32_t inputLengthAlgin32 = (((inputLength + BLOCK_SIZE - 1) / BLOCK_SIZE) * BLOCK_SIZE); - // There is at least 32B of data on each core, satisfying several settings for several cores. The maximum number of audits is the actual number of audits - coreNum = (coreNum < inputLengthAlgin32 / BLOCK_SIZE) ? coreNum : inputLengthAlgin32 / BLOCK_SIZE; - coreNum = (coreNum >= 1) ? coreNum : 1; - uint32_t everyCoreInputBlockNum = inputLengthAlgin32 / BLOCK_SIZE / coreNum; - uint32_t tailBlockNum = (inputLengthAlgin32 / BLOCK_SIZE) % coreNum; - - // Small chunks are calculated and sliced several times using the number of data on each core - uint32_t smallCoreDataNum = everyCoreInputBlockNum * BLOCK_SIZE / inputBytes; - uint32_t smallTileNum = everyCoreInputBlockNum / tileBlockNum; - uint32_t finalSmallTileNum = (everyCoreInputBlockNum % tileBlockNum) == 0 ? smallTileNum : smallTileNum + 1; - // Tail block calculation for small chunks of data - uint32_t smallTailDataNum = smallCoreDataNum - (tileDataNum * smallTileNum); - smallTailDataNum = smallTailDataNum == 0 ? tileDataNum : smallTailDataNum; - - // The total length of a large block of data is 32B larger than that of a small block of data - everyCoreInputBlockNum += 1; - uint32_t bigCoreDataNum = everyCoreInputBlockNum * BLOCK_SIZE / inputBytes; - uint32_t bigTileNum = everyCoreInputBlockNum / tileBlockNum; - uint32_t finalBigTileNum = (everyCoreInputBlockNum % tileBlockNum) == 0 ? bigTileNum : bigTileNum + 1; - uint32_t bigTailDataNum = bigCoreDataNum - tileDataNum * bigTileNum; - bigTailDataNum = bigTailDataNum == 0 ? tileDataNum : bigTailDataNum; - - tiling.set_smallCoreDataNum(smallCoreDataNum); - tiling.set_bigCoreDataNum(bigCoreDataNum); - tiling.set_tileDataNum(tileDataNum); - tiling.set_smallTailDataNum(smallTailDataNum); - tiling.set_bigTailDataNum(bigTailDataNum); - tiling.set_finalSmallTileNum(finalSmallTileNum); - tiling.set_finalBigTileNum(finalBigTileNum); - tiling.set_tailBlockNum(tailBlockNum); - - context->SetBlockDim(coreNum); - tiling.SaveToBuffer(context->GetRawTilingData()->GetData(), context->GetRawTilingData()->GetCapacity()); - context->GetRawTilingData()->SetDataSize(tiling.GetDataSize()); - size_t *currentWorkspace = context->GetWorkspaceSizes(1); - currentWorkspace[0] = 0; - return ge::GRAPH_SUCCESS; -} -} - -namespace ge { -static ge::graphStatus InferShape(gert::InferShapeContext* context) -{ - const gert::Shape* x1_shape = context->GetInputShape(0); - gert::Shape* y_shape = context->GetOutputShape(0); - *y_shape = *x1_shape; - return GRAPH_SUCCESS; -} -static graphStatus InferDataType(gert::InferDataTypeContext* context) -{ - const auto inputDataType = context->GetInputDataType(0); - context->SetOutputDataType(0, inputDataType); - return ge::GRAPH_SUCCESS; -} -} - -namespace ops { -class AddCustom : public OpDef { -public: - explicit AddCustom(const char* name) : OpDef(name) - { - this->Input("x") - .ParamType(REQUIRED) - .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) - .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) - .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); - this->Input("y") - .ParamType(REQUIRED) - .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) - .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) - .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); - this->Output("z") - .ParamType(REQUIRED) - .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32, ge::DT_INT8}) - .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) - .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); - - this->SetInferShape(ge::InferShape).SetInferDataType(ge::InferDataType); - this->AICore() - .SetTiling(optiling::TilingFunc) - .AddConfig("ascend310b") - .AddConfig("ascend910b"); - } -}; -OP_ADD(AddCustom); -} diff --git a/atb_operator/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h b/atb_operator/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h deleted file mode 100644 index 28fd00a37..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/op_host/add_custom_tiling.h +++ /dev/null @@ -1,22 +0,0 @@ -/* - * Copyright (c) Huawei Technologies Co., Ltd. 2022-2023. All rights reserved. - */ -#ifndef ADD_CUSTOM_TILING_H -#define ADD_CUSTOM_TILING_H -#include "register/tilingdata_base.h" - -namespace optiling { -BEGIN_TILING_DATA_DEF(TilingData) - TILING_DATA_FIELD_DEF(uint32_t, smallCoreDataNum); - TILING_DATA_FIELD_DEF(uint32_t, bigCoreDataNum); - TILING_DATA_FIELD_DEF(uint32_t, finalBigTileNum); - TILING_DATA_FIELD_DEF(uint32_t, finalSmallTileNum); - TILING_DATA_FIELD_DEF(uint32_t, tileDataNum); - TILING_DATA_FIELD_DEF(uint32_t, smallTailDataNum); - TILING_DATA_FIELD_DEF(uint32_t, bigTailDataNum); - TILING_DATA_FIELD_DEF(uint32_t, tailBlockNum); -END_TILING_DATA_DEF; - -REGISTER_TILING_DATA_CLASS(AddCustom, TilingData) -} -#endif // ADD_CUSTOM_TILING_H \ No newline at end of file diff --git a/atb_operator/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt b/atb_operator/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt deleted file mode 100644 index 0d31a444c..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/op_kernel/CMakeLists.txt +++ /dev/null @@ -1,61 +0,0 @@ -# set custom compile options -if ("${CMAKE_BUILD_TYPE}x" STREQUAL "Debugx") - add_ops_compile_options(ALL OPTIONS -g -O0) -endif() - -foreach(compute_unit ${ASCEND_COMPUTE_UNIT}) - - # generate aic-${compute_unit}-ops-info.json - add_ops_info_target(TARGET ops_info_gen_${compute_unit} - OUTPUT ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core/${compute_unit}/aic-${compute_unit}-ops-info.json - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/config/${compute_unit} - ) - - # generate ascendc impl py once - if (NOT TARGET ascendc_impl_gen) - add_ops_impl_target(TARGET ascendc_impl_gen - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl - ) - endif() - - # dynamic shape binary compile - if (${ENABLE_BINARY_PACKAGE}) - add_bin_compile_target(TARGET ascendc_bin_${compute_unit} - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} - ADP_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/dynamic - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/binary/${compute_unit} - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/kernel - COMPUTE_UNIT ${compute_unit} - ) - add_dependencies(ascendc_bin_${compute_unit} ascendc_impl_gen) - endif() - -endforeach() - -# generate npu_supported_ops.json -add_npu_support_target(TARGET npu_supported_ops - OPS_INFO_DIR ${ASCEND_AUTOGEN_PATH} - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core - INSTALL_DIR packages/vendors/${vendor_name}/framework/${ASCEND_FRAMEWORK_TYPE} -) - -if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) - add_subdirectory(testcases) -endif() - -# install kernel file -if (${ENABLE_SOURCE_PACKAGE}) - file(GLOB KERNEL_FILES - ${CMAKE_CURRENT_SOURCE_DIR}/*.cpp - ${CMAKE_CURRENT_SOURCE_DIR}/*.h - ${CMAKE_CURRENT_SOURCE_DIR}/*.py - ) - install(FILES ${KERNEL_FILES} - DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl/dynamic - ) -endif() diff --git a/atb_operator/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp b/atb_operator/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp deleted file mode 100644 index 37c80a825..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/op_kernel/add_custom.cpp +++ /dev/null @@ -1,134 +0,0 @@ -/* - * Copyright (c) Huawei Technologies Co., Ltd. 2022-2023. All rights reserved. - * - * Function : z = x + y - * This sample is a very basic sample that implements vector add on Ascend plaform. - */ -#include "kernel_operator.h" -// tensor num for each queue -constexpr int32_t BUFFER_NUM = 2; - -template class KernelAdd { - using T = TYPE_X; -public: - __aicore__ inline KernelAdd() {} - __aicore__ inline void Init(GM_ADDR x, GM_ADDR y, GM_ADDR z, uint32_t smallCoreDataNum, - uint32_t bigCoreDataNum, uint32_t finalBigTileNum, - uint32_t finalSmallTileNum, uint32_t tileDataNum, - uint32_t smallTailDataNum, uint32_t bigTailDataNum, - uint32_t tailBlockNum) - { - ASSERT(AscendC::GetBlockNum() != 0 && "block dim can not be zero!"); - uint32_t coreNum = AscendC::GetBlockIdx(); - uint32_t globalBufferIndex = bigCoreDataNum * AscendC::GetBlockIdx(); - this->tileDataNum = tileDataNum; - if (coreNum < tailBlockNum) { - this->coreDataNum = bigCoreDataNum; - this->tileNum = finalBigTileNum; - this->tailDataNum = bigTailDataNum; - } - else { - this->coreDataNum = smallCoreDataNum; - this->tileNum = finalSmallTileNum; - this->tailDataNum = smallTailDataNum; - globalBufferIndex -= (bigCoreDataNum - smallCoreDataNum) * (AscendC::GetBlockIdx() - tailBlockNum); - } - xGm.SetGlobalBuffer((__gm__ TYPE_X*)x + globalBufferIndex, this->coreDataNum); - yGm.SetGlobalBuffer((__gm__ TYPE_Y*)y + globalBufferIndex, this->coreDataNum); - zGm.SetGlobalBuffer((__gm__ TYPE_Z*)z + globalBufferIndex, this->coreDataNum); - pipe.InitBuffer(inQueueX, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_X)); - pipe.InitBuffer(inQueueY, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_Y)); - pipe.InitBuffer(outQueueZ, BUFFER_NUM, this->tileDataNum * sizeof(TYPE_Z)); - pipe.InitBuffer(tmp1, this->tileDataNum * sizeof(half)); - pipe.InitBuffer(tmp2, this->tileDataNum * sizeof(half)); - } - __aicore__ inline void Process() - { - int32_t loopCount = this->tileNum; - this->processDataNum = this->tileDataNum; - for (int32_t i = 0; i < loopCount; i++) { - if (i == this->tileNum - 1) { - this->processDataNum = this->tailDataNum; - } - CopyIn(i); - Compute(i); - CopyOut(i); - } - } - -private: - __aicore__ inline void CopyIn(int32_t progress) - { - AscendC::LocalTensor xLocal = inQueueX.AllocTensor(); - AscendC::LocalTensor yLocal = inQueueY.AllocTensor(); - AscendC::DataCopy(xLocal, xGm[progress * this->tileDataNum], this->processDataNum); - AscendC::DataCopy(yLocal, yGm[progress * this->tileDataNum], this->processDataNum); - inQueueX.EnQue(xLocal); - inQueueY.EnQue(yLocal); - } - __aicore__ inline void Compute(int32_t progress) - { - AscendC::LocalTensor xLocal = inQueueX.DeQue(); - AscendC::LocalTensor yLocal = inQueueY.DeQue(); - AscendC::LocalTensor zLocal = outQueueZ.AllocTensor(); - if constexpr (std::is_same_v) { - auto p1 = tmp1.Get(); - auto p2 = tmp2.Get(); - AscendC::Cast(p1, xLocal, AscendC::RoundMode::CAST_NONE, this->processDataNum); - AscendC::Cast(p2, yLocal, AscendC::RoundMode::CAST_NONE, this->processDataNum); - AscendC::Add(p2, p1, p2, this->processDataNum); - AscendC::Cast(p1.ReinterpretCast(), p2, AscendC::RoundMode::CAST_RINT, this->processDataNum); - AscendC::ShiftLeft(p1.ReinterpretCast(), p1.ReinterpretCast(), int16_t(8), this->processDataNum); - AscendC::ShiftRight(p1.ReinterpretCast(), p1.ReinterpretCast(), int16_t(8), this->processDataNum); - AscendC::Cast(p2, p1.ReinterpretCast(), AscendC::RoundMode::CAST_NONE, this->processDataNum); - AscendC::Cast(zLocal, p2, AscendC::RoundMode::CAST_NONE, this->processDataNum); - } - else { - AscendC::Add(zLocal, xLocal, yLocal, this->processDataNum); - } - outQueueZ.EnQue(zLocal); - inQueueX.FreeTensor(xLocal); - inQueueY.FreeTensor(yLocal); - } - __aicore__ inline void CopyOut(int32_t progress) - { - AscendC::LocalTensor zLocal = outQueueZ.DeQue(); - AscendC::DataCopy(zGm[progress * this->tileDataNum], zLocal, this->processDataNum); - outQueueZ.FreeTensor(zLocal); - } - -private: - AscendC::TPipe pipe; - AscendC::TQue inQueueX, inQueueY; - AscendC::TQue outQueueZ; - AscendC::TBuf tmp1, tmp2; - AscendC::GlobalTensor xGm; - AscendC::GlobalTensor yGm; - AscendC::GlobalTensor zGm; - uint32_t coreDataNum; - uint32_t tileNum; - uint32_t tileDataNum; - uint32_t tailDataNum; - uint32_t processDataNum; -}; - -extern "C" __global__ __aicore__ void add_custom(GM_ADDR x, GM_ADDR y, GM_ADDR z, GM_ADDR workspace, GM_ADDR tiling) -{ - GET_TILING_DATA(tiling_data, tiling); - KernelAdd op; - op.Init(x, y, z, tiling_data.smallCoreDataNum, - tiling_data.bigCoreDataNum, tiling_data.finalBigTileNum, - tiling_data.finalSmallTileNum, tiling_data.tileDataNum, - tiling_data.smallTailDataNum, tiling_data.bigTailDataNum, - tiling_data.tailBlockNum); - op.Process(); -} - -#ifndef ASCENDC_CPU_DEBUG -// call of kernel function -void add_custom_do(uint32_t blockDim, void* l2ctrl, void* stream, uint8_t* x, uint8_t* y, uint8_t* z, - uint8_t* workspace, uint8_t* tiling) -{ - add_custom<<>>(x, y, z, workspace, tiling); -} -#endif diff --git a/atb_operator/AtbAddSample/AddAscendC/scripts/install.sh b/atb_operator/AtbAddSample/AddAscendC/scripts/install.sh deleted file mode 100644 index 9d45a8052..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/scripts/install.sh +++ /dev/null @@ -1,318 +0,0 @@ -#!/bin/bash -vendor_name=customize -targetdir=/usr/local/Ascend/opp -target_custom=0 - -sourcedir=$PWD/packages -vendordir=vendors/$vendor_name - -QUIET="y" - -while true -do - case $1 in - --quiet) - QUIET="y" - shift - ;; - --install-path=*) - INSTALL_PATH=$(echo $1 | cut -d"=" -f2-) - INSTALL_PATH=${INSTALL_PATH%*/} - shift - ;; - --*) - shift - ;; - *) - break - ;; - esac -done - -log() { - cur_date=`date +"%Y-%m-%d %H:%M:%S"` - echo "[runtime] [$cur_date] "$1 -} - -if [ -n "${INSTALL_PATH}" ]; then - if [[ ! "${INSTALL_PATH}" = /* ]]; then - log "[ERROR] use absolute path for --install-path argument" - exit 1 - fi - if [ ! -d ${INSTALL_PATH} ]; then - mkdir ${INSTALL_PATH} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${INSTALL_PATH} failed" - exit 1 - fi - fi - targetdir=${INSTALL_PATH} -elif [ -n "${ASCEND_CUSTOM_OPP_PATH}" ]; then - if [ ! -d ${ASCEND_CUSTOM_OPP_PATH} ]; then - mkdir -p ${ASCEND_CUSTOM_OPP_PATH} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${ASCEND_CUSTOM_OPP_PATH} failed" - fi - fi - targetdir=${ASCEND_CUSTOM_OPP_PATH} -else - if [ "x${ASCEND_OPP_PATH}" == "x" ]; then - log "[ERROR] env ASCEND_OPP_PATH no exist" - exit 1 - fi - targetdir="${ASCEND_OPP_PATH}" -fi - -if [ ! -d $targetdir ];then - log "[ERROR] $targetdir no exist" - exit 1 -fi - -upgrade() -{ - if [ ! -d ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 files" - return 0 - fi - - if [ ! -d ${targetdir}/$vendordir/$1 ];then - log "[INFO] create ${targetdir}/$vendordir/$1." - mkdir -p ${targetdir}/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/$1 failed" - return 1 - fi - else - has_same_file=-1 - for file_a in ${sourcedir}/$vendordir/$1/*; do - file_b=${file_a##*/}; - if [ "ls ${targetdir}/$vendordir/$1" = "" ]; then - log "[INFO] ${targetdir}/$vendordir/$1 is empty !!" - return 1 - fi - grep -q $file_b <<<`ls ${targetdir}/$vendordir/$1`; - if [[ $? -eq 0 ]]; then - echo -n "${file_b} " - has_same_file=0 - fi - done - if [ 0 -eq $has_same_file ]; then - if test $QUIET = "n"; then - echo "[INFO]: has old version in ${targetdir}/$vendordir/$1, \ - you want to Overlay Installation , please enter:[o]; \ - or replace directory installation , please enter: [r]; \ - or not install , please enter:[n]." - - while true - do - read orn - if [ "$orn" = n ]; then - return 0 - elif [ "$orn" = m ]; then - break; - elif [ "$0rn" = r ]; then - [ -n "${targetdir}/$vendordir/$1/" ] && rm -rf "${targetdir}/$vendordir/$1"/* - break; - else - echo "[ERROR] input error, please input again!" - fi - done - fi - fi - log "[INFO] replace or merge old ops $1 files .g....." - fi - - log "copy new ops $1 files ......" - if [ -d ${targetdir}/$vendordir/$1/ ]; then - chmod -R +w "$targetdir/$vendordir/$1/" >/dev/null 2>&1 - fi - cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 files failed" - return 1 - fi - - return 0 -} -upgrade_proto() -{ - if [ ! -f ${sourcedir}/$vendordir/custom.proto ]; then - log "[INFO] no need to upgrade custom.proto files" - return 0 - fi - if [ ! -d ${targetdir}/$vendordir/framework/caffe ];then - log "[INFO] create ${targetdir}/$vendordir/framework/caffe." - mkdir -p ${targetdir}/$vendordir/framework/caffe - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/framework/caffe failed" - return 1 - fi - else - if [ -f ${targetdir}/$vendordir/framework/caffe/custom.proto ]; then - # 有老版本,判断是否要覆盖式安装 - if test $QUIET = "n"; then - echo "[INFO] ${targetdir}/$vendordir/framework/caffe has old version"\ - "custom.proto file. Do you want to replace? [y/n] " - - while true - do - read yn - if [ "$yn" = n ]; then - return 0 - elif [ "$yn" = y ]; then - break; - else - echo "[ERROR] input error, please input again!" - fi - done - fi - fi - log "[INFO] replace old caffe.proto files ......" - fi - chmod -R +w "$targetdir/$vendordir/framework/caffe/" >/dev/null 2>&1 - cp -rf ${sourcedir}/$vendordir/custom.proto ${targetdir}/$vendordir/framework/caffe/ - if [ $? -ne 0 ];then - log "[ERROR] copy new custom.proto failed" - return 1 - fi - log "[INFO] copy custom.proto success" - - return 0 -} - -upgrade_file() -{ - if [ ! -e ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 file" - return 0 - fi - - log "copy new $1 files ......" - cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 file failed" - return 1 - fi - - return 0 -} - -delete_optiling_file() -{ - if [ ! -d ${targetdir}/vendors ];then - log "[INFO] $1 not exist, no need to uninstall" - return 0 - fi - sys_info=$(uname -m) - if [ ! -d ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/lib/linux/${sys_info} ];then - rm -rf ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/liboptiling.so - fi - return 0 -} - -log "[INFO] copy uninstall sh success" - -if [ ! -d ${targetdir}/vendors ];then - log "[INFO] create ${targetdir}/vendors." - mkdir -p ${targetdir}/vendors - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/vendors failed" - return 1 - fi -fi -chmod u+w ${targetdir}/vendors - -echo "[ops_custom]upgrade framework" -upgrade framework -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op proto" -upgrade op_proto -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade version.info" -upgrade_file version.info -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op impl" -delete_optiling_file op_impl -upgrade op_impl -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op api" -upgrade op_api -if [ $? -ne 0 ];then - exit 1 -fi - -upgrade_proto -if [ $? -ne 0 ];then - exit 1 -fi - -# set the set_env.bash -if [ -n "${INSTALL_PATH}" ] && [ -d ${INSTALL_PATH} ]; then - _ASCEND_CUSTOM_OPP_PATH=${targetdir}/${vendordir} - bin_path="${_ASCEND_CUSTOM_OPP_PATH}/bin" - set_env_variable="#!/bin/bash\nexport ASCEND_CUSTOM_OPP_PATH=${_ASCEND_CUSTOM_OPP_PATH}:\${ASCEND_CUSTOM_OPP_PATH}" - if [ ! -d ${bin_path} ]; then - mkdir -p ${bin_path} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${bin_path} failed" - exit 1 - fi - fi - echo -e ${set_env_variable} > ${bin_path}/set_env.bash - if [ $? -ne 0 ]; then - log "[ERROR] write ASCEND_CUSTOM_OPP_PATH to set_env.bash failed" - exit 1 - else - log "[INFO] using requirements: when custom module install finished or before you run the custom module, \ - execute the command [ source ${bin_path}/set_env.bash ] to set the environment path" - fi -else - config_file=${targetdir}/vendors/config.ini - if [ ! -f ${config_file} ]; then - touch ${config_file} - chmod 640 ${config_file} - echo "load_priority=$vendor_name" > ${config_file} - if [ $? -ne 0 ];then - echo "echo load_priority failed" - exit 1 - fi - else - found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" - found_vendor=$(echo $found_vendors | sed "s/\<$vendor_name\>//g" | tr ',' ' ') - vendor=$(echo $found_vendor | tr -s ' ' ',') - if [ "$vendor" != "" ]; then - sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" - fi - fi -fi - -chmod u-w ${targetdir}/vendors - -if [ -d ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/ ]; then - chmod -R 440 ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/* >/dev/null 2>&1 -fi -if [ -f ${targetdir}/ascend_install.info ]; then - chmod -R 440 ${targetdir}/ascend_install.info -fi -if [ -f ${targetdir}/scene.info ]; then - chmod -R 440 ${targetdir}/scene.info -fi -if [ -f ${targetdir}/version.info ]; then - chmod -R 440 ${targetdir}/version.info -fi - -echo "SUCCESS" -exit 0 - diff --git a/atb_operator/AtbAddSample/AddAscendC/scripts/upgrade.sh b/atb_operator/AtbAddSample/AddAscendC/scripts/upgrade.sh deleted file mode 100644 index 34801efc4..000000000 --- a/atb_operator/AtbAddSample/AddAscendC/scripts/upgrade.sh +++ /dev/null @@ -1,151 +0,0 @@ -#!/bin/bash -vendor_name=customize -targetdir=/usr/local/Ascend/opp -target_custom=0 - -sourcedir=$PWD/packages -vendordir=vendors/$vendor_name - -log() { - cur_date=`date +"%Y-%m-%d %H:%M:%S"` - echo "[runtime] [$cur_date] "$1 -} - -if [[ "x${ASCEND_OPP_PATH}" == "x" ]];then - log "[ERROR] env ASCEND_OPP_PATH no exist" - exit 1 -fi - -targetdir=${ASCEND_OPP_PATH} - -if [ ! -d $targetdir ];then - log "[ERROR] $targetdir no exist" - exit 1 -fi - -upgrade() -{ - if [ ! -d ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 files" - return 0 - fi - - if [ ! -d ${targetdir}/$vendordir/$1 ];then - log "[INFO] create ${targetdir}/$vendordir/$1." - mkdir -p ${targetdir}/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/$1 failed" - return 1 - fi - else - vendor_installed_dir=$(ls "$targetdir/vendors" 2> /dev/null) - for i in $vendor_installed_dir;do - vendor_installed_file=$(ls "$vendor_installed_dir/$vendor_name/$i" 2> /dev/null) - if [ "$i" = "$vendor_name" ] && [ "$vendor_installed_file" != "" ]; then - echo "[INFO]: $vendor_name custom opp package has been installed on the path $vendor_installed_dir, \ - you want to Overlay Installation , please enter:[o]; \ - or replace directory installation , please enter: [r]; \ - or not install , please enter:[n]." - fi - while true - do - read mrn - if [ "$mrn" = m ]; then - break - elif [ "$mrn" = r ]; then - [ -n "$vendor_installed_file"] && rm -rf "$vendor_installed_file" - break - elif [ "$mrn" = n ]; then - return 0 - else - echo "[WARNING]: Input error, please input m or r or n to choose!" - fi - done - done - log "[INFO] replace old ops $1 files ......" - fi - - log "copy new ops $1 files ......" - cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 files failed" - return 1 - fi - - return 0 -} - -upgrade_file() -{ - if [ ! -e ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 file" - return 0 - fi - - log "copy new $1 files ......" - cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 file failed" - return 1 - fi - - return 0 -} - -log "[INFO] copy uninstall sh success" - -echo "[ops_custom]upgrade framework" -upgrade framework -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op proto" -upgrade op_proto -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op impl" -upgrade op_impl -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op api" -upgrade op_api -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade version.info" -upgrade_file version.info -if [ $? -ne 0 ];then - exit 1 -fi - -config_file=${targetdir}/vendors/config.ini -found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" -found_vendor=$(echo $found_vendors | sed "s/\<$vendor_name\>//g" | tr ',' ' ') -vendor=$(echo $found_vendor | tr -s ' ' ',') -if [ "$vendor" != "" ]; then - sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" -fi - -changemode() -{ - if [ -d ${targetdir} ];then - chmod -R 550 ${targetdir}>/dev/null 2>&1 - fi - - return 0 -} -echo "[ops_custom]changemode..." -#changemode -if [ $? -ne 0 ];then - exit 1 -fi - -echo "SUCCESS" -exit 0 - diff --git a/atb_operator/AtbAddSample/readme.md b/atb_operator/AtbAddSample/readme.md index dbab64271..c1d8752db 100644 --- a/atb_operator/AtbAddSample/readme.md +++ b/atb_operator/AtbAddSample/readme.md @@ -1,8 +1,8 @@ -### 概述 +## 概述 -本样例基于AscendC自定义Add算子,开发了ATB插件并进行了插件调用测试. +本样例基于AscendC自定义[Add](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)算子,开发了ATB插件并进行了插件调用测试. -### 项目结构介绍 +## 项目结构介绍 ``` ├── AddAscendC //Add AscendC自定义算子工程 @@ -11,25 +11,13 @@ ├── AddOperationTest //AddOperation 测试代码 ``` -### 样例运行 +## 样例运行 -#### Add AscendC自定义算子部署 +### Add AscendC自定义算子部署 -- 运行生成算子run包 +参照[cann-ops](https://gitee.com/ascend/cann-ops/blob/master/README.md) - ```shell - cd AddAscendC - bash build.sh - ``` - -- 算子run包部署 - - ```shell - cd AddAscendC/build_out - ./custom_opp_hce_aarch64.run - ``` - -#### AddOperation ATB插件部署 +### AddOperation ATB插件部署 - 运行编译脚本完成部署(脚本会生成静态库.a文件,同时将头文件拷贝到/usr/include,.a文件拷贝到/usr/local/lib下) @@ -38,7 +26,7 @@ bash build.sh ``` -#### AddOperation测试 +### AddOperation测试 - 运行脚本完成算子测试 @@ -47,36 +35,36 @@ bash run.sh ``` -### AddOperation算子介绍 +## AddOperation算子介绍 -#### 功能 +### 功能 实现两个输入张量相加 -#### 定义 +### 定义 ``` struct AddParam { }; ``` -#### 参数列表 +### 参数列表 该算子参数为空 -#### 输入 +### 输入 | **参数** | **维度** | **数据类型** | **格式** | 描述 | | -------- | -------------------------- | --------------------- | -------- | ---------- | | x | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | | y | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | -#### 输出 +### 输出 | **参数** | **维度** | **数据类型** | **格式** | 描述 | | -------- | -------------------------- | --------------------- | -------- | ---------------------------------------- | | output | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输出tensor。数据类型和shape与x保持一致。 | -#### 规格约束 +### 规格约束 暂无 \ No newline at end of file diff --git a/atb_operator/AtbEyeSample/EyeOp/CMakeLists.txt b/atb_operator/AtbEyeSample/EyeOp/CMakeLists.txt deleted file mode 100755 index 584132d80..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/CMakeLists.txt +++ /dev/null @@ -1,69 +0,0 @@ -cmake_minimum_required(VERSION 3.16.0) -project(opp) -if(ENABLE_CROSS_COMPILE) - if(${CMAKE_SYSTEM_PROCESSOR} STREQUAL x86_64) - set(CROSS_COMPILE_PLATFORM aarch64) - else() - set(CROSS_COMPILE_PLATFORM x86_64) - endif() - set(PLATFORM ${CMAKE_SYSTEM_PROCESSOR}) - set(CMAKE_COMPILE_COMPILER_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/linux/${CROSS_COMPILE_PLATFORM}/) - set(CMAKE_COMPILE_RUNTIME_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/${CROSS_COMPILE_PLATFORM}/) - set(CMAKE_SYSTEM_PROCESSOR ${CROSS_COMPILE_PLATFORM}) - set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) - set(CMAKE_CXX_COMPILER ${CMAKE_CROSS_PLATFORM_COMPILER}) -else() - set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) -endif() - -include(cmake/config.cmake) -include(cmake/func.cmake) -include(cmake/intf.cmake) - -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/framework) - add_subdirectory(framework) -endif() -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_host) - add_subdirectory(op_host) -endif() -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_kernel) - add_subdirectory(op_kernel) -endif() -if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) - add_subdirectory(testcases) -endif() - -# modify vendor_name in install.sh and upgrade.sh -add_custom_command(OUTPUT ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh - COMMAND mkdir -p ${CMAKE_BINARY_DIR}/scripts - COMMAND cp -r ${CMAKE_SOURCE_DIR}/scripts/* ${CMAKE_BINARY_DIR}/scripts/ - COMMAND sed -i "s/vendor_name=customize/vendor_name=${vendor_name}/g" ${CMAKE_BINARY_DIR}/scripts/* -) -add_custom_target(modify_vendor ALL DEPENDS ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh) -install(DIRECTORY ${CMAKE_BINARY_DIR}/scripts/ DESTINATION . FILE_PERMISSIONS OWNER_EXECUTE OWNER_READ GROUP_READ) - -install(FILES ${CMAKE_SOURCE_DIR}/custom.proto DESTINATION packages OPTIONAL) - -get_system_info(SYSTEM_INFO) - -# gen version.info -add_custom_target(gen_version_info ALL - COMMAND bash ${CMAKE_CURRENT_SOURCE_DIR}/cmake/util/gen_version_info.sh ${ASCEND_CANN_PACKAGE_PATH} ${CMAKE_CURRENT_BINARY_DIR} -) - -install(FILES ${CMAKE_CURRENT_BINARY_DIR}/version.info - DESTINATION packages/vendors/${vendor_name}/) - -# CPack config -set(CPACK_PACKAGE_NAME ${CMAKE_PROJECT_NAME}) -set(CPACK_PACKAGE_VERSION ${CMAKE_PROJECT_VERSION}) -set(CPACK_PACKAGE_DESCRIPTION "CPack opp project") -set(CPACK_PACKAGE_DESCRIPTION_SUMMARY "CPack opp project") -set(CPACK_PACKAGE_DIRECTORY ${CMAKE_INSTALL_PREFIX}) -set(CPACK_PACKAGE_FILE_NAME "custom_opp_${SYSTEM_INFO}.run") -set(CPACK_GENERATOR External) -set(CPACK_CMAKE_GENERATOR "Unix Makefiles") -set(CPACK_EXTERNAL_ENABLE_STAGING TRUE) -set(CPACK_EXTERNAL_PACKAGE_SCRIPT ${CMAKE_SOURCE_DIR}/cmake/makeself.cmake) -set(CPACK_EXTERNAL_BUILT_PACKAGES ${CPACK_PACKAGE_DIRECTORY}/_CPack_Packages/Linux/External/${CPACK_PACKAGE_FILE_NAME}/${CPACK_PACKAGE_FILE_NAME}) -include(CPack) diff --git a/atb_operator/AtbEyeSample/EyeOp/CMakePresets.json b/atb_operator/AtbEyeSample/EyeOp/CMakePresets.json deleted file mode 100755 index d32bf6d9a..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/CMakePresets.json +++ /dev/null @@ -1,63 +0,0 @@ -{ - "version": 1, - "cmakeMinimumRequired": { - "major": 3, - "minor": 19, - "patch": 0 - }, - "configurePresets": [ - { - "name": "default", - "displayName": "Default Config", - "description": "Default build using Unix Makefiles generator", - "generator": "Unix Makefiles", - "binaryDir": "${sourceDir}/build_out", - "cacheVariables": { - "CMAKE_BUILD_TYPE": { - "type": "STRING", - "value": "Release" - }, - "ENABLE_SOURCE_PACKAGE": { - "type": "BOOL", - "value": "True" - }, - "ENABLE_BINARY_PACKAGE": { - "type": "BOOL", - "value": "True" - }, - "ASCEND_COMPUTE_UNIT": { - "type": "STRING", - "value": "ascend310b;ascend910b" - }, - "ENABLE_TEST": { - "type": "BOOL", - "value": "True" - }, - "vendor_name": { - "type": "STRING", - "value": "customize" - }, - "ASCEND_CANN_PACKAGE_PATH": { - "type": "PATH", - "value": "/usr/local/Ascend/ascend-toolkit/latest" - }, - "ASCEND_PYTHON_EXECUTABLE": { - "type": "STRING", - "value": "python3" - }, - "CMAKE_INSTALL_PREFIX": { - "type": "PATH", - "value": "${sourceDir}/build_out" - }, - "ENABLE_CROSS_COMPILE": { - "type": "BOOL", - "value": "False" - }, - "CMAKE_CROSS_PLATFORM_COMPILER": { - "type": "PATH", - "value": "/usr/bin/aarch64-linux-gnu-g++" - } - } - } - ] -} \ No newline at end of file diff --git a/atb_operator/AtbEyeSample/EyeOp/README.md b/atb_operator/AtbEyeSample/EyeOp/README.md deleted file mode 100644 index 5b2d1c7e6..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/README.md +++ /dev/null @@ -1,216 +0,0 @@ -# 概述 -本样例基于EyeOp算子工程,介绍了单算子工程及aclnn接口说明。 - -## 算子工程介绍 -### 算子描述 -创建batch_shape个二维矩阵 $m\times n$ ,对角元素全为1,其它元素都为0。 - -### 算子规格描述 - - - - - - - - - - - - - - - - - - - - - - - - - - -
算子类型(OpType)Eye
nameshapedata typeformat默认值
算子输入y-float32, float16, int32ND\
算子输出y-float32, float16, int32ND\
attr属性num_rows\int\\
num_columns\int\0
batch_shape\list_int\{1}
dtype\int\0
核函数名eye
- - -### 算子工程文件介绍 -其中,算子工程目录EyeOp包含算子实现的模板文件、编译脚本等,如下所示: -``` -├── EyeOp //Add自定义算子工程 -│ ├── cmake -│ ├── framework // 算子插件实现文件目录,单算子模型文件的生成不依赖算子适配插件,无需关注 -│ ├── op_host // host侧实现文件 -│ ├── op_kernel // kernel侧实现文件 -│ ├── scripts // 自定义算子工程打包相关脚本所在目录 -│ ├── build.sh // 编译入口脚本 -│ ├── CMakeLists.txt // 算子工程的CMakeLists.txt -│ └── CMakePresets.json // 编译配置项 -``` - -### 编译运行样例算子 -针对自定义算子工程,编译运行包含如下步骤: -- 编译自定义算子工程生成算子安装包; -- 安装自定义算子到算子库中; -- 调用执行自定义算子; - -详细操作如下所示。 -#### 1. 获取源码包 -编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 -#### 2. 编译算子工程 - 编译自定义算子工程,构建生成自定义算子包。 - - - 执行如下命令,切换到算子工程EyeOp目录。 - - ```bash - cd ${git_clone_path}/samples/atb_operator/AtbEyeSample/EyeOp - ``` - - - 修改CMakePresets.json中ASCEND_CANN_PACKAGE_PATH为CANN软件包安装后的实际路径。 - - - ```json - { - …… - "configurePresets": [ - { - …… - "ASCEND_CANN_PACKAGE_PATH": { - "type": "PATH", - "value": "/usr/local/Ascend/ascend-toolkit/latest" //请替换为CANN软件包安装后的实际路径。eg:/home/HwHiAiUser/Ascend/ascend-toolkit/latest - }, - …… - } - ] - } - ``` - - 在算子工程EyeOp目录下执行如下命令,进行算子工程编译。 - - ```bash - ./build.sh - ``` -编译成功后,会在当前目录下创建build_out目录,并在build_out目录下生成自定义算子安装包custom_opp_\_\.run,例如“custom_opp_ubuntu_x86_64.run”。 - -备注:如果要使用dump调试功能,需要移除op_host内和CMakeLists.txt内的Atlas 200/500 A2 推理产品的配置。 - -#### 3. 部署算子包 - -执行如下命令,在自定义算子安装包所在路径下,安装自定义算子包。 - ```bash - cd build_out - ./custom_opp__.run - ``` -命令执行成功后,自定义算子包中的相关文件将部署至当前环境的OPP算子库的vendors/customize目录中。 - -#### 4. 配置环境变量 - - 请根据当前环境上CANN开发套件包的[安装方式](https://hiascend.com/document/redirect/CannCommunityInstSoftware),选择对应配置环境变量的命令。 - - 默认路径,root用户安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=/usr/local/Ascend/ascend-toolkit/latest - ``` - - 默认路径,非root用户安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=$HOME/Ascend/ascend-toolkit/latest - ``` - - 指定路径install_path,安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=${install_path}/ascend-toolkit/latest - ``` - -## aclnn接口说明 - -自定义算子编译部署后,会自动生成单算子API,可以直接在应用程序中调用。算子API的形式一般定义为“两段式接口”,形如: - ```cpp - aclnnStatus aclnnEyeGetWorkspaceSize(const aclTensor *x, int num_rows, int num_columns, int *batch_shape, int dtype, uint64_t workspaceSize, aclOpExecutor **executor); - aclnnStatus aclnnEye(void *workspace, int64_t workspaceSize, aclOpExecutor **executor, aclrtStream stream); - ``` -其中aclnnEyeGetWorkspaceSize为第一段接口,主要用于计算本次API调用计算过程中需要多少的workspace内存。获取到本次API计算需要的workspace大小之后,按照workspaceSize大小申请Device侧内存,然后调用第二段接口aclnnEye执行计算。具体参考[AscendCL单算子调用](https://hiascend.com/document/redirect/CannCommunityAscendCInVorkSingleOp)>单算子API执行 章节。 - -### 功能描述 -* 算子功能:创建batch_shape个二维矩阵 $m\times n$ ,对角元素全为1,其它元素都为0。 - -#### 参数说明 -##### aclnnEyeGetWorkspaceSize: - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
接口参数描述
aclnnEyeGetWorkspaceSizex输入输出tensor,支持flaot/half/int32类型,ND排布格式
num_rows必选属性,生成矩阵的行数,支持int32类型数值
num_columns可选属性,生成矩阵的列数,支持int32类型数值,默认与num_rows一致
batch_shape可选属性,数组内元素乘积为生成二维矩阵的个数,支持int数组,默认为{}
dtype可选属性,输出tensor的数据类型,当前输入输出同地址,同数据类型,该参数无作用,支持int32类型数值, 默认为0
workspaceSize返回需要在Device侧申请的workspace大小
executor返回op执行器,包含了算子计算流程
- - -| 返回值 | 描述 | -| ---- | ---- | -| ACL_SUCCESS | 成功 | -| ACL_ERROR | 失败 | - -##### aclnnEye - - - - - - - - - - - - - - - - - - - - - - - -
接口参数描述
aclnnEyeworkspace输入Device侧申请的workspace内存
workspaceSizeworkspace大小
executor输入op执行器,包含了算子计算流程
stream输入执行流
- -| 返回值 | 描述 | -| ---- | ---- | -| ACL_SUCCESS | 成功 | -| ACL_ERROR | 失败 | - - -## 更新说明 -| 时间 | 更新事项 | -|----|------| -| 2025/03/27 | 新增README | \ No newline at end of file diff --git a/atb_operator/AtbEyeSample/EyeOp/build.sh b/atb_operator/AtbEyeSample/EyeOp/build.sh deleted file mode 100755 index 27d1968d1..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/build.sh +++ /dev/null @@ -1,80 +0,0 @@ -#!/bin/bash -script_path=$(realpath $(dirname $0)) - - -mkdir -p build_out -rm -rf build_out/* -cd build_out - -opts=$(python3 $script_path/cmake/util/preset_parse.py $script_path/CMakePresets.json) -ENABLE_CROSS="-DENABLE_CROSS_COMPILE=True" -ENABLE_BINARY="-DENABLE_BINARY_PACKAGE=True" -cmake_version=$(cmake --version | grep "cmake version" | awk '{print $3}') - -cmake_run_package() -{ - target=$1 - cmake --build . --target $target -j16 - if [ $? -ne 0 ]; then exit 1; fi - - if [ $target = "package" ]; then - if test -d ./op_kernel/binary ; then - ./cust*.run - if [ $? -ne 0 ]; then exit 1; fi - cmake --build . --target binary -j16 - if [ $? -ne 0 ]; then - echo "[ERROR] Kernel compile failed, the run package will not be generated." - rm -rf ./cust*.run && rm -rf ./cust*.run.json && exit 1; - fi - cmake --build . --target $target -j16 - fi - fi -} - -if [[ $opts =~ $ENABLE_CROSS ]] && [[ $opts =~ $ENABLE_BINARY ]] -then - target=package - if [ "$1"x != ""x ]; then target=$1; fi - if [ "$cmake_version" \< "3.19.0" ] ; then - cmake .. $opts -DENABLE_CROSS_COMPILE=0 - else - cmake .. --preset=default -DENABLE_CROSS_COMPILE=0 - fi - cmake_run_package $target - cp -r kernel ../ - rm -rf * - if [ "$cmake_version" \< "3.19.0" ] ; then - cmake .. $opts - else - cmake .. --preset=default - fi - - cmake --build . --target $target -j16 - if [ $? -ne 0 ]; then - echo "[ERROR] Kernel compile failed, the run package will not be generated." - rm -rf ./cust*.run && rm -rf ./cust*.run.json && exit 1; - fi - if [ $target = "package" ]; then - if test -d ./op_kernel/binary ; then - ./cust*.run - fi - fi - rm -rf ../kernel - -else - target=package - if [ "$1"x != ""x ]; then target=$1; fi - if [ "$cmake_version" \< "3.19.0" ] ; then - cmake .. $opts - else - cmake .. --preset=default - fi - cmake_run_package $target -fi - - -# for debug -# cd build_out -# make -# cpack -# verbose append -v \ No newline at end of file diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/config.cmake b/atb_operator/AtbEyeSample/EyeOp/cmake/config.cmake deleted file mode 100755 index 886119daa..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/config.cmake +++ /dev/null @@ -1,25 +0,0 @@ - -set(CMAKE_CXX_FLAGS_DEBUG "") -set(CMAKE_CXX_FLAGS_RELEASE "") - -if (NOT DEFINED vendor_name) - set(vendor_name customize CACHE STRING "") -endif() -if (NOT DEFINED ASCEND_CANN_PACKAGE_PATH) - set(ASCEND_CANN_PACKAGE_PATH /usr/local/Ascend/latest CACHE PATH "") -endif() -if (NOT DEFINED ASCEND_PYTHON_EXECUTABLE) - set(ASCEND_PYTHON_EXECUTABLE python3 CACHE STRING "") -endif() -if (NOT DEFINED ASCEND_COMPUTE_UNIT) - message(FATAL_ERROR "ASCEND_COMPUTE_UNIT not set in CMakePreset.json ! -") -endif() -set(ASCEND_TENSOR_COMPILER_PATH ${ASCEND_CANN_PACKAGE_PATH}/compiler) -set(ASCEND_CCEC_COMPILER_PATH ${ASCEND_TENSOR_COMPILER_PATH}/ccec_compiler/bin) -set(ASCEND_AUTOGEN_PATH ${CMAKE_BINARY_DIR}/autogen) -set(ASCEND_FRAMEWORK_TYPE tensorflow) -file(MAKE_DIRECTORY ${ASCEND_AUTOGEN_PATH}) -set(CUSTOM_COMPILE_OPTIONS "custom_compile_options.ini") -execute_process(COMMAND rm -rf ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} - COMMAND touch ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS}) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/func.cmake b/atb_operator/AtbEyeSample/EyeOp/cmake/func.cmake deleted file mode 100755 index 4179dfd25..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/func.cmake +++ /dev/null @@ -1,192 +0,0 @@ - -function(get_system_info SYSTEM_INFO) - if (UNIX) - execute_process(COMMAND grep -i ^id= /etc/os-release OUTPUT_VARIABLE TEMP) - string(REGEX REPLACE "\n|id=|ID=|\"" "" SYSTEM_NAME ${TEMP}) - set(${SYSTEM_INFO} ${SYSTEM_NAME}_${CMAKE_SYSTEM_PROCESSOR} PARENT_SCOPE) - elseif (WIN32) - message(STATUS "System is Windows. Only for pre-build.") - else () - message(FATAL_ERROR "${CMAKE_SYSTEM_NAME} not support.") - endif () -endfunction() - -function(opbuild) - message(STATUS "Opbuild generating sources") - cmake_parse_arguments(OPBUILD "" "OUT_DIR;PROJECT_NAME;ACCESS_PREFIX" "OPS_SRC" ${ARGN}) - execute_process(COMMAND ${CMAKE_COMPILE} -g -fPIC -shared -std=c++11 ${OPBUILD_OPS_SRC} -D_GLIBCXX_USE_CXX11_ABI=0 - -I ${ASCEND_CANN_PACKAGE_PATH}/include -L ${ASCEND_CANN_PACKAGE_PATH}/lib64 -lexe_graph -lregister -ltiling_api - -o ${OPBUILD_OUT_DIR}/libascend_all_ops.so - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("build ops lib info: ${EXEC_INFO}") - message("build ops lib error: ${EXEC_ERROR}") - message(FATAL_ERROR "opbuild run failed!") - endif() - set(proj_env "") - set(prefix_env "") - if (NOT "${OPBUILD_PROJECT_NAME}x" STREQUAL "x") - set(proj_env "OPS_PROJECT_NAME=${OPBUILD_PROJECT_NAME}") - endif() - if (NOT "${OPBUILD_ACCESS_PREFIX}x" STREQUAL "x") - set(prefix_env "OPS_DIRECT_ACCESS_PREFIX=${OPBUILD_ACCESS_PREFIX}") - endif() - execute_process(COMMAND ${proj_env} ${prefix_env} ${ASCEND_CANN_PACKAGE_PATH}/toolkit/tools/opbuild/op_build - ${OPBUILD_OUT_DIR}/libascend_all_ops.so ${OPBUILD_OUT_DIR} - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("opbuild ops info: ${EXEC_INFO}") - message("opbuild ops error: ${EXEC_ERROR}") - endif() - message(STATUS "Opbuild generating sources - done") -endfunction() - -function(add_ops_info_target) - cmake_parse_arguments(OPINFO "" "TARGET;OPS_INFO;OUTPUT;INSTALL_DIR" "" ${ARGN}) - get_filename_component(opinfo_file_path "${OPINFO_OUTPUT}" DIRECTORY) - add_custom_command(OUTPUT ${OPINFO_OUTPUT} - COMMAND mkdir -p ${opinfo_file_path} - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/parse_ini_to_json.py - ${OPINFO_OPS_INFO} ${OPINFO_OUTPUT} - ) - add_custom_target(${OPINFO_TARGET} ALL - DEPENDS ${OPINFO_OUTPUT} - ) - install(FILES ${OPINFO_OUTPUT} - DESTINATION ${OPINFO_INSTALL_DIR} - ) -endfunction() - -function(add_ops_compile_options OP_TYPE) - cmake_parse_arguments(OP_COMPILE "" "OP_TYPE" "COMPUTE_UNIT;OPTIONS" ${ARGN}) - file(APPEND ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} - "${OP_TYPE},${OP_COMPILE_COMPUTE_UNIT},${OP_COMPILE_OPTIONS}\n") -endfunction() - -function(add_ops_impl_target) - cmake_parse_arguments(OPIMPL "" "TARGET;OPS_INFO;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) - add_custom_command(OUTPUT ${OPIMPL_OUT_DIR}/.impl_timestamp - COMMAND mkdir -m 700 -p ${OPIMPL_OUT_DIR}/dynamic - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py - ${OPIMPL_OPS_INFO} - \"${OPIMPL_OPS_BATCH}\" \"${OPIMPL_OPS_ITERATE}\" - ${OPIMPL_IMPL_DIR} - ${OPIMPL_OUT_DIR}/dynamic - ${ASCEND_AUTOGEN_PATH} - - COMMAND rm -rf ${OPIMPL_OUT_DIR}/.impl_timestamp - COMMAND touch ${OPIMPL_OUT_DIR}/.impl_timestamp - DEPENDS ${OPIMPL_OPS_INFO} - ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py - ) - add_custom_target(${OPIMPL_TARGET} ALL - DEPENDS ${OPIMPL_OUT_DIR}/.impl_timestamp) - if (${ENABLE_SOURCE_PACKAGE}) - install(DIRECTORY ${OPIMPL_OUT_DIR}/dynamic - DESTINATION ${OPIMPL_INSTALL_DIR} - ) - endif() -endfunction() - -function(add_npu_support_target) - cmake_parse_arguments(NPUSUP "" "TARGET;OPS_INFO_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) - get_filename_component(npu_sup_file_path "${NPUSUP_OUT_DIR}" DIRECTORY) - add_custom_command(OUTPUT ${NPUSUP_OUT_DIR}/npu_supported_ops.json - COMMAND mkdir -p ${NPUSUP_OUT_DIR} - COMMAND ${CMAKE_SOURCE_DIR}/cmake/util/gen_ops_filter.sh - ${NPUSUP_OPS_INFO_DIR} - ${NPUSUP_OUT_DIR} - ) - add_custom_target(npu_supported_ops ALL - DEPENDS ${NPUSUP_OUT_DIR}/npu_supported_ops.json - ) - install(FILES ${NPUSUP_OUT_DIR}/npu_supported_ops.json - DESTINATION ${NPUSUP_INSTALL_DIR} - ) -endfunction() - -function(add_bin_compile_target) - cmake_parse_arguments(BINCMP "" "TARGET;OPS_INFO;COMPUTE_UNIT;IMPL_DIR;ADP_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/src) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/gen) - execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_bin_param_build.py - ${BINCMP_OPS_INFO} ${BINCMP_OUT_DIR}/gen ${BINCMP_COMPUTE_UNIT} - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("ops binary compile scripts gen info: ${EXEC_INFO}") - message("ops binary compile scripts gen error: ${EXEC_ERROR}") - message(FATAL_ERROR "ops binary compile scripts gen failed!") - endif() - if (NOT TARGET binary) - add_custom_target(binary) - endif() - add_custom_target(${BINCMP_TARGET} - COMMAND cp -r ${BINCMP_IMPL_DIR}/*.* ${BINCMP_OUT_DIR}/src - ) - add_custom_target(${BINCMP_TARGET}_gen_ops_config - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/insert_simplified_keys.py -p ${BINCMP_OUT_DIR}/bin - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_ops_config.py -p ${BINCMP_OUT_DIR}/bin - -s ${BINCMP_COMPUTE_UNIT} - ) - add_dependencies(binary ${BINCMP_TARGET}_gen_ops_config) - file(GLOB bin_scripts ${BINCMP_OUT_DIR}/gen/*.sh) - foreach(bin_script ${bin_scripts}) - get_filename_component(bin_file ${bin_script} NAME_WE) - string(REPLACE "-" ";" bin_sep ${bin_file}) - list(GET bin_sep 0 op_type) - list(GET bin_sep 1 op_file) - list(GET bin_sep 2 op_index) - if (NOT TARGET ${BINCMP_TARGET}_${op_file}_copy) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file}) - add_custom_target(${BINCMP_TARGET}_${op_file}_copy - COMMAND cp ${BINCMP_ADP_DIR}/${op_file}.py ${BINCMP_OUT_DIR}/src/${op_type}.py - ) - install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} - DESTINATION ${BINCMP_INSTALL_DIR}/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) - install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json - DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT}/ OPTIONAL - ) - endif() - add_custom_target(${BINCMP_TARGET}_${op_file}_${op_index} - COMMAND export HI_PYTHON=${ASCEND_PYTHON_EXECUTABLE} && bash ${bin_script} ${BINCMP_OUT_DIR}/src/${op_type}.py ${BINCMP_OUT_DIR}/bin/${op_file} && echo $(MAKE) - WORKING_DIRECTORY ${BINCMP_OUT_DIR} - ) - add_dependencies(${BINCMP_TARGET}_${op_file}_${op_index} ${BINCMP_TARGET} ${BINCMP_TARGET}_${op_file}_copy) - add_dependencies(${BINCMP_TARGET}_gen_ops_config ${BINCMP_TARGET}_${op_file}_${op_index}) - endforeach() - install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json - DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) - - install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} - DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/../build_out/kernel/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) - install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json - DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/../build_out/kernel/config/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) - install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json - DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/../build_out/kernel/config/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) - -endfunction() - -function(add_cross_compile_target) - cmake_parse_arguments(CROSSMP "" "TARGET;OUT_DIR;INSTALL_DIR" "" ${ARGN}) - add_custom_target(${CROSSMP_TARGET} ALL - DEPENDS ${CROSSMP_OUT_DIR} - ) - install(DIRECTORY ${CROSSMP_OUT_DIR} - DESTINATION ${CROSSMP_INSTALL_DIR} - ) -endfunction() diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/intf.cmake b/atb_operator/AtbEyeSample/EyeOp/cmake/intf.cmake deleted file mode 100755 index 2f362c396..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/intf.cmake +++ /dev/null @@ -1,26 +0,0 @@ - -add_library(intf_pub INTERFACE) -target_compile_options(intf_pub INTERFACE - -fPIC - -fvisibility=hidden - -fvisibility-inlines-hidden - $<$:-O2> - $<$:-O0 -g> - $<$:-std=c++11> - $<$,$>:-ftrapv -fstack-check> - $<$:-pthread -Wfloat-equal -Wshadow -Wformat=2 -Wno-deprecated -Wextra> - $,-fstack-protector-strong,-fstack-protector-all> -) -target_compile_definitions(intf_pub INTERFACE - _GLIBCXX_USE_CXX11_ABI=0 - $<$:_FORTIFY_SOURCE=2> -) -target_include_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/include) -target_link_options(intf_pub INTERFACE - $<$,EXECUTABLE>:-pie> - $<$:-s> - -Wl,-z,relro - -Wl,-z,now - -Wl,-z,noexecstack -) -target_link_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/lib64) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/makeself.cmake b/atb_operator/AtbEyeSample/EyeOp/cmake/makeself.cmake deleted file mode 100755 index 48c565bfb..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/makeself.cmake +++ /dev/null @@ -1,17 +0,0 @@ -execute_process(COMMAND chmod +x ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh) -execute_process(COMMAND ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh - --header ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself-header.sh - --help-header ./help.info - --gzip --complevel 4 --nomd5 --sha256 - ./ ${CPACK_PACKAGE_FILE_NAME} "version:1.0" ./install.sh - WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} - RESULT_VARIABLE EXEC_RESULT - ERROR_VARIABLE EXEC_ERROR -) -if (NOT "${EXEC_RESULT}x" STREQUAL "0x") - message(FATAL_ERROR "CPack Command error: ${EXEC_RESULT}\n${EXEC_ERROR}") -endif() -execute_process(COMMAND cp ${CPACK_EXTERNAL_BUILT_PACKAGES} ${CPACK_PACKAGE_DIRECTORY}/ - COMMAND echo "Copy ${CPACK_EXTERNAL_BUILT_PACKAGES} to ${CPACK_PACKAGE_DIRECTORY}/" - WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} -) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/__init__.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/__init__.py deleted file mode 100755 index c4ddc893a..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/__init__.py +++ /dev/null @@ -1,8 +0,0 @@ -#!/usr/bin/env python -# -*- coding: utf-8 -*- - -import sys -import os - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) -sys.path.append(PYF_PATH) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_bin_param_build.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_bin_param_build.py deleted file mode 100755 index 63b15c900..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_bin_param_build.py +++ /dev/null @@ -1,127 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import json -import hashlib -import const_var -import opdesc_parser - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - - -class BinParamBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - super().__init__(op_type) - self.soc = '' - self.out_path = '' - - def set_soc_version(self: any, soc: str): - self.soc = soc - - def set_out_path(self: any, out_path: str): - self.out_path = out_path - - def gen_input_json(self: any): - key_map = {} - count = len(self.input_dtype[0].split(',')) - for i in range(0, count): - inputs = [] - outputs = [] - attrs = [] - op_node = {} - for idx in range(0, len(self.input_name)): - idtypes = self.input_dtype[idx].split(',') - ifmts = self.input_fmt[idx].split(',') - itype = self.input_type[idx] - para = {} - para['name'] = self.input_name[idx][:-5] - para['index'] = idx - para['dtype'] = idtypes[i] - para['format'] = ifmts[i] - para['paramType'] = itype - para['shape'] = [-2] - if itype == 'dynamic': - inputs.append([para]) - else: - inputs.append(para) - for idx in range(0, len(self.output_name)): - odtypes = self.output_dtype[idx].split(',') - ofmts = self.output_fmt[idx].split(',') - otype = self.output_type[idx] - para = {} - para['name'] = self.output_name[idx][:-5] - para['index'] = idx - para['dtype'] = odtypes[i] - para['format'] = ofmts[i] - para['paramType'] = otype - para['shape'] = [-2] - if otype == 'dynamic': - outputs.append([para]) - else: - outputs.append(para) - for attr in self.attr_list: - att = {} - att['name'] = attr - atype = self.attr_val.get(attr).get('type').lower() - att['dtype'] = atype - att['value'] = const_var.ATTR_DEF_VAL.get(atype) - attrs.append(att) - op_node['bin_filename'] = '' - op_node['inputs'] = inputs - op_node['outputs'] = outputs - if len(attrs) > 0: - op_node['attrs'] = attrs - param = {} - param['op_type'] = self.op_type - param['op_list'] = [op_node] - objstr = json.dumps(param, indent=' ') - md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() - while key_map.get(md5sum) is not None: - objstr += '1' - md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() - key_map[md5sum] = md5sum - bin_file = self.op_type + '_' + md5sum - op_node['bin_filename'] = bin_file - param_file = os.path.join(self.out_path, bin_file + '_param.json') - param_file = os.path.realpath(param_file) - with os.fdopen(os.open(param_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - json.dump(param, fd, indent=' ') - self._write_buld_cmd(param_file, bin_file, i) - - def _write_buld_cmd(self: any, param_file: str, bin_file: str, index: int): - hard_soc = const_var.SOC_MAP_EXT.get(self.soc) - if not hard_soc: - hard_soc = self.soc.capitalize() - name_com = [self.op_type, self.op_file, str(index)] - compile_file = os.path.join(self.out_path, '-'.join(name_com) + '.sh') - compile_file = os.path.realpath(compile_file) - with os.fdopen(os.open(compile_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - fd.write('#!/bin/bash\n') - fd.write('echo "[{}] Generating {} ..."\n'.format(hard_soc, bin_file)) - cmd = const_var.BIN_CMD.format(fun=self.op_intf, soc=hard_soc, param=param_file, impl='""') - fd.write(cmd) - chk = const_var.CHK_CMD.format(res_file=bin_file + '.json') - fd.write(chk) - chk = const_var.CHK_CMD.format(res_file=bin_file + '.o') - fd.write(chk) - fd.write('echo "[{}] Generating {} Done"\n'.format(hard_soc, bin_file)) - - -def gen_bin_param_file(cfgfile: str, out_dir: str, soc: str): - op_descs = opdesc_parser.get_op_desc(cfgfile, [], [], BinParamBuilder, None) - for op_desc in op_descs: - op_desc.set_soc_version(soc) - op_desc.set_out_path(out_dir) - op_desc.gen_input_json() - - -if __name__ == '__main__': - if len(sys.argv) <= 3: - raise RuntimeError('arguments must greater than 3') - gen_bin_param_file(sys.argv[1], sys.argv[2], sys.argv[3]) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_impl_build.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_impl_build.py deleted file mode 100755 index 647801c07..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_impl_build.py +++ /dev/null @@ -1,501 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import re -import opdesc_parser -import const_var - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - -IMPL_HEAD = ''' -import os, sys -import ctypes -import json -import shutil -from tbe.common.platform import get_soc_spec -from tbe.common.utils import para_check -from tbe.tikcpp import compile_op, replay_op, check_op_cap, generalize_op_params, get_code_channel, OpInfo -from tbe.common.buildcfg import get_default_build_config -from impl.util.platform_adapter import tbe_register -from tbe.common.buildcfg import get_current_build_config -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - -DTYPE_MAP = {{"float32": ["DT_FLOAT", "float"], - "float16": ["DT_FLOAT16", "half"], - "int8": ["DT_INT8", "int8_t"], - "int16": ["DT_INT16", "int16_t"], - "int32": ["DT_INT32", "int32_t"], - "int64": ["DT_INT64", "int64_t"], - "uint1": ["DT_UINT1", "uint8_t"], - "uint8": ["DT_UINT8", "uint8_t"], - "uint16": ["DT_UINT16", "uint16_t"], - "uint32": ["DT_UINT32", "uint32_t"], - "uint64": ["DT_UINT64", "uint64_t"], - "bool": ["DT_BOOL", "bool"], - "double": ["DT_DOUBLE", "double"], - "dual": ["DT_DUAL", "unknown"], - "dual_sub_int8": ["DT_DUAL_SUB_INT8", "unknown"], - "dual_sub_uint8": ["DT_DUAL_SUB_UINT8", "unknown"], - "string": ["DT_STRING", "unknown"], - "complex64": ["DT_COMPLEX64", "unknown"], - "complex128": ["DT_COMPLEX128", "unknown"], - "qint8": ["DT_QINT8", "unknown"], - "qint16": ["DT_QINT16", "unknown"], - "qint32": ["DT_QINT32", "unknown"], - "quint8": ["DT_QUINT8", "unknown"], - "quint16": ["DT_QUINT16", "unknown"], - "resource": ["DT_RESOURCE", "unknown"], - "string_ref": ["DT_STRING_REF", "unknown"], - "int4": ["DT_INT4", "int4b_t"], - "bfloat16": ["DT_BF16", "bfloat16_t"]}} - -def add_dtype_fmt_option_single(x, x_n, is_ref: bool = False): - options = [] - x_fmt = x.get("format") - x_dtype = x.get("dtype") - x_n_in_kernel = x_n + '_REF' if is_ref else x_n - options.append("-DDTYPE_{{n}}={{t}}".format(n=x_n_in_kernel, t=DTYPE_MAP.get(x_dtype)[1])) - options.append("-DORIG_DTYPE_{{n}}={{ot}}".format(n=x_n_in_kernel, ot=DTYPE_MAP.get(x_dtype)[0])) - options.append("-DFORMAT_{{n}}=FORMAT_{{f}}".format(n=x_n_in_kernel, f=x_fmt)) - return options - -def get_dtype_fmt_options(__inputs__, __outputs__): - options = [] - input_names = {} - output_names = {} - unique_param_name_set = set() - for idx, x in enumerate(__inputs__): - if x is None: - continue - x_n = input_names[idx].upper() - unique_param_name_set.add(x_n) - options += add_dtype_fmt_option_single(x, x_n) - - for idx, x in enumerate(__outputs__): - if x is None: - continue - x_n = output_names[idx].upper() - if x_n in unique_param_name_set: - options += add_dtype_fmt_option_single(x, x_n, True) - else: - options += add_dtype_fmt_option_single(x, x_n) - return options - -def load_dso(so_path): - try: - ctypes.CDLL(so_path) - except OSError as error : - print(error) - raise RuntimeError("cannot open %s" %(so_path)) - else: - print("load so succ ", so_path) - -def get_shortsoc_compile_option(compile_option_list: list, shortsoc:str): - compile_options = [] - if shortsoc in compile_option_list: - compile_options = compile_option_list[shortsoc] - elif '__ALLSOC__' in compile_option_list: - compile_options = compile_option_list['__ALLSOC__'] - return compile_options - -def get_kernel_source(src_file, dir_snake, dir_ex): - src_ex = os.path.join(PYF_PATH, "..", "ascendc", dir_ex, src_file) - if os.path.exists(src_ex): - return src_ex - src = os.path.join(PYF_PATH, "..", "ascendc", dir_snake, src_file) - if os.path.exists(src): - return src - src = os.path.join(PYF_PATH, src_file) - if os.path.exists(src): - return src - return src_ex - -''' - -IMPL_API = ''' -@tbe_register.register_operator("{}", trans_bool_to_s8=False) -@para_check.check_op_params({}) -def {}({}, kernel_name="{}", impl_mode=""): - if get_current_build_config("enable_op_prebuild"): - return - __inputs__, __outputs__, __attrs__ = _build_args({}) - options = get_dtype_fmt_options(__inputs__, __outputs__) - options += ["-x", "cce"] - bisheng = shutil.which("bisheng") - if bisheng != None: - bisheng_path = os.path.dirname(bisheng) - tikcpp_path = os.path.realpath(os.path.join(bisheng_path, "..", "..", "tikcpp")) - else: - tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") - options.append("-I" + tikcpp_path) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) - options.append("-I" + os.path.join(PYF_PATH, "..", "ascendc", "common")) - if impl_mode == "high_performance": - options.append("-DHIGH_PERFORMANCE=1") - elif impl_mode == "high_precision": - options.append("-DHIGH_PRECISION=1") - if get_default_build_config("enable_deterministic_mode") == 1: - options.append("-DDETEMINISTIC_MODE=1") - - custom_compile_options = {}, - custom_all_compile_options = {}, - soc_version = get_soc_spec("SOC_VERSION") - soc_short = get_soc_spec("SHORT_SOC_VERSION").lower() - custom_compile_options_soc = get_shortsoc_compile_option(custom_compile_options[0], soc_short) - custom_all_compile_options_soc = get_shortsoc_compile_option(custom_all_compile_options[0], soc_short) - options += custom_all_compile_options_soc - options += custom_compile_options_soc - - origin_func_name = "{}" - ascendc_src_dir_ex = "{}" - ascendc_src_dir = "{}" - ascendc_src_file = "{}" - src = get_kernel_source(ascendc_src_file, ascendc_src_dir, ascendc_src_dir_ex) -''' - -REPLAY_OP_API = ''' - print("start replay Acend C Operator {}, kernel name is {}") - tikreplay_codegen_path = tikcpp_path + "/tikreplaylib/lib" - tikreplay_stub_path = tikcpp_path + "/tikreplaylib/lib/" + soc_version - print("start load libtikreplaylib_codegen.so and libtikreplaylib_stub.so") - codegen_so_path = tikreplay_codegen_path + "/libtikreplaylib_codegen.so" - replaystub_so_path = tikreplay_stub_path + "/libtikreplaylib_stub.so" - if PYF_PATH.endswith("dynamic"): - op_replay_path = os.path.join(PYF_PATH, "..", "..", "op_replay") - else: - op_replay_path = os.path.join(PYF_PATH, "..", "op_replay") - replayapi_so_path = os.path.join(op_replay_path, "libreplay_{}_" + soc_short + ".so") - load_dso(codegen_so_path) - load_dso(replaystub_so_path) - load_dso(replayapi_so_path) - op_type = "{}" - entry_obj = os.path.join(op_replay_path, "{}_entry_" + soc_short + ".o") - code_channel = get_code_channel(src, kernel_name, op_type, options) - op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ - attrs = __attrs__, impl_mode = impl_mode) - res, msg = replay_op(op_info, entry_obj, code_channel, src, options) - if not res: - print("call replay op failed for %s and get into call compile op" %(msg)) - compile_op(src, origin_func_name, op_info, options, code_channel, '{}') -''' - -COMPILE_OP_API = ''' - print("start compile Ascend C operator {}. kernel name is " + kernel_name) - op_type = "{}" - code_channel = get_code_channel(src, kernel_name, op_type, options) - op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ - attrs = __attrs__, impl_mode = impl_mode, origin_inputs=[{}], origin_outputs = [{}], param_type_dynamic = {},\\ - mc2_ctx = None, param_type_list = {}) - compile_op(src, origin_func_name, op_info, options, code_channel, '{}') -''' - -SUP_API = ''' -def {}({}, impl_mode=""): - __inputs__, __outputs__, __attrs__ = _build_args({}) - ret_str = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) - ret_dict = json.loads(ret_str) - err_code = ret_dict.get("ret_code") - sup = "Unknown" - reason = "Unknown reason" - if err_code is not None: - if err_code is 0: - sup = "True" - reason = "" - elif err_code is 1: - sup = "False" - reason = ret_dict.get("reason") - else: - sup = "Unknown" - reason = ret_dict.get("reason") - return sup, reason -''' -CAP_API = ''' -def {}({}, impl_mode=""): - __inputs__, __outputs__, __attrs__ = _build_args({}) - result = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) - return result.decode("utf-8") -''' -GLZ_API = ''' -@tbe_register.register_param_generalization("{}") -def {}_generalization({}, generalize_config=None): - __inputs__, __outputs__, __attrs__ = _build_args({}) - ret_str = generalize_op_params("{}", __inputs__, __outputs__, __attrs__, generalize_config) - return [json.loads(ret_str)] -''' - -ATTR_DEFAULT = {'bool': 'False', 'int': '0', 'float': '0.0', 'list_int': '[]', - 'list_float': '[]', 'list_bool': '[]', 'list_list_int': '[[]]', 'str': ''} - - -def _get_snake_str(s, i): - if s[i - 1] != '_': - if not s[i - 1].isupper(): - return "_" - elif s[i - 1].isupper() and (i + 1) < len(s) and s[i + 1].islower(): - return "_" - return "" - return "" - - -def optype_snake(origin_str): - temp_str = origin_str[0].lower() + origin_str[1:] - new_str = re.sub(r'([A-Z])', r'_\1', temp_str).lower() - return new_str - - -def optype_snake_ex(s): - snake_case = "" - for i, c in enumerate(s): - if i == 0: - snake_case += c.lower() - elif c.isupper(): - snake_case += _get_snake_str(s, i) - snake_case += c.lower() - else: - snake_case += c - return snake_case - - -class AdpBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - self.argsdefv = [] - self.op_compile_option:str = '{}' - super().__init__(op_type) - - - def write_adapt(self: any, impl_path, path: str, op_compile_option_all:list = None): - self._build_paradefault() - if impl_path != "": - src_file = os.path.join(impl_path, self.op_file + '.cpp') - if not os.path.exists(src_file): - return - out_path = os.path.abspath(path) - if self.dynamic_shape and not out_path.endswith('dynamic'): - out_path = os.path.join(path, 'dynamic') - os.makedirs(out_path, mode=0o700, exist_ok=True) - adpfile = os.path.join(out_path, self.op_file + '.py') - self._gen_op_compile_option(op_compile_option_all) - with os.fdopen(os.open(adpfile, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - self._write_head(fd) - self._write_argparse(fd) - self._write_impl(fd) - if self.op_chk_support: - self._write_cap('check_supported', fd) - self._write_cap('get_op_support_info', fd) - if self.op_fmt_sel: - self._write_cap('op_select_format', fd) - self._write_cap('get_op_specific_info', fd) - if self.op_range_limit == 'limited' or self.op_range_limit == 'dynamic': - self._write_glz(fd) - - - def _gen_op_compile_option(self:any, op_compile_option_all:list =None): - if op_compile_option_all is not None: - if self.op_type in op_compile_option_all: - self.op_compile_option = op_compile_option_all[self.op_type] - elif "__all__" in op_compile_option_all: - self.op_compile_option = op_compile_option_all["__all__"] - - - def _ip_argpack(self: any, default: bool = True) -> list: - args = [] - for i in range(len(self.input_name)): - arg = self.input_name[i] - if default and self.argsdefv[i] is not None: - arg += '=' + self.argsdefv[i] - args.append(arg) - return args - - def _op_argpack(self: any, default: bool = True) -> list: - args = [] - argidx = len(self.input_name) - for i in range(len(self.output_name)): - arg = self.output_name[i] - if default and self.argsdefv[i + argidx] is not None: - arg += '=' + self.argsdefv[i + argidx] - args.append(arg) - return args - - def _attr_argpack(self: any, default: bool = True) -> list: - args = [] - argidx = len(self.input_name) + len(self.output_name) - for i in range(len(self.attr_list)): - att = self.attr_list[i] - arg = att - if default and self.argsdefv[i + argidx] is not None: - if self.attr_val.get(att).get('type') == 'str': - arg += '="' + self.argsdefv[i + argidx] + '"' - elif self.attr_val.get(att).get('type') == 'bool': - arg += '=' + self.argsdefv[i + argidx].capitalize() - else: - arg += '=' + self.argsdefv[i + argidx] - args.append(arg) - return args - - def _build_paralist(self: any, default: bool = True) -> str: - args = [] - args.extend(self._ip_argpack(default)) - args.extend(self._op_argpack(default)) - args.extend(self._attr_argpack(default)) - return ', '.join(args) - - def _io_parachk(self: any, types: list, type_name: str) -> list: - chk = [] - for iot in types: - if iot == 'optional': - ptype = 'OPTION' - else: - ptype = iot.upper() - chk.append('para_check.{}_{}'.format(ptype, type_name)) - return chk - - def _attr_parachk(self: any) -> list: - chk = [] - for att in self.attr_list: - if self.attr_val.get(att).get('paramType') == 'optional': - pt = 'OPTION' - else: - pt = self.attr_val.get(att).get('paramType').upper() - att_type = self.attr_val.get(att).get('type').upper() - chk.append('para_check.{}_ATTR_{}'.format(pt, att_type)) - return chk - - def _build_parachk(self: any) -> str: - chk = [] - chk.extend(self._io_parachk(self.input_type, 'INPUT')) - chk.extend(self._io_parachk(self.output_type, 'OUTPUT')) - chk.extend(self._attr_parachk()) - chk.append('para_check.KERNEL_NAME') - return ', '.join(chk) - - def _build_paradefault(self: any): - optional = False - argtypes = [] - argtypes.extend(self.input_type) - argtypes.extend(self.output_type) - for atype in argtypes: - if atype == 'optional': - optional = True - if optional: - self.argsdefv.append('None') - else: - self.argsdefv.append(None) - for attr in self.attr_list: - atype = self.attr_val.get(attr).get('paramType') - if atype == 'optional': - optional = True - attrval = self.attr_val.get(attr).get('defaultValue') - if attrval is not None: - optional = True - if type == "bool": - attrval = attrval.capitalize() - elif type == "str": - attrval = "\"" + attrval + "\"" - self.argsdefv.append(attrval) - continue - if optional: - self.argsdefv.append(ATTR_DEFAULT.get(self.attr_val.get(attr).get('type'))) - else: - self.argsdefv.append(None) - - def _write_head(self: any, fd: object): - fd.write(IMPL_HEAD.format(self.input_ori_name, self.output_ori_name)) - - def _write_argparse(self: any, fd: object): - args = self._build_paralist(False) - fd.write('def _build_args({}):\n'.format(args)) - fd.write(' __inputs__ = []\n') - fd.write(' for arg in [{}]:\n'.format(', '.join(self.input_name))) - fd.write(' if arg != None:\n') - fd.write(' if isinstance(arg, (list, tuple)):\n') - fd.write(' if len(arg) == 0:\n') - fd.write(' continue\n') - fd.write(' __inputs__.append(arg[0])\n') - fd.write(' else:\n') - fd.write(' __inputs__.append(arg)\n') - fd.write(' else:\n') - fd.write(' __inputs__.append(arg)\n') - fd.write(' __outputs__ = []\n') - fd.write(' for arg in [{}]:\n'.format(', '.join(self.output_name))) - fd.write(' if arg != None:\n') - fd.write(' if isinstance(arg, (list, tuple)):\n') - fd.write(' if len(arg) == 0:\n') - fd.write(' continue\n') - fd.write(' __outputs__.append(arg[0])\n') - fd.write(' else:\n') - fd.write(' __outputs__.append(arg)\n') - fd.write(' else:\n') - fd.write(' __outputs__.append(arg)\n') - fd.write(' __attrs__ = []\n') - for attr in self.attr_list: - fd.write(' if {} != None:\n'.format(attr)) - fd.write(' attr = {}\n') - fd.write(' attr["name"] = "{}"\n'.format(attr)) - fd.write(' attr["dtype"] = "{}"\n'.format(self.attr_val.get(attr).get('type'))) - fd.write(' attr["value"] = {}\n'.format(attr)) - fd.write(' __attrs__.append(attr)\n') - fd.write(' return __inputs__, __outputs__, __attrs__\n') - - def _write_impl(self: any, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - pchk = self._build_parachk() - if len(self.kern_name) > 0: - kern_name = self.kern_name - else: - kern_name = self.op_intf - src = self.op_file + '.cpp' - fd.write(IMPL_API.format(self.op_type, pchk, self.op_intf, argsdef, kern_name, argsval,\ - self.custom_compile_options, self.custom_all_compile_options, self.op_intf,\ - optype_snake_ex(self.op_type), optype_snake(self.op_type), src)) - if self.op_replay_flag: - fd.write(REPLAY_OP_API.format(self.op_type, kern_name, self.op_file, self.op_type, self.op_file,\ - self.op_compile_option)) - else: - fd.write(COMPILE_OP_API.format(self.op_type, self.op_type, ', '.join(self.input_name),\ - ', '.join(self.output_name), False, self.input_type + self.output_type, self.op_compile_option)) - - def _write_cap(self: any, cap_name: str, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - if cap_name == 'check_supported': - fd.write(SUP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) - else: - fd.write(CAP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) - - def _write_glz(self: any, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - fd.write(GLZ_API.format(self.op_type, self.op_intf, argsdef, argsval, self.op_type)) - - -def write_scripts(cfgfile: str, cfgs: dict, dirs: dict, ops: list = None, op_compile_option:list = None): - batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') - iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') - file_map = {} - op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, AdpBuilder,\ - ops, dirs.get(const_var.AUTO_GEN_DIR)) - for op_desc in op_descs: - op_desc.write_adapt(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), op_compile_option) - file_map[op_desc.op_type] = op_desc.op_file - return file_map - -if __name__ == '__main__': - if len(sys.argv) <= 6: - raise RuntimeError('arguments must greater equal than 6') - rep_cfg = {} - rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] - rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] - cfg_dir = {} - cfg_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] - cfg_dir[const_var.CFG_OUT_DIR] = sys.argv[5] - cfg_dir[const_var.AUTO_GEN_DIR] = sys.argv[6] - write_scripts(cfgfile=sys.argv[1], cfgs=rep_cfg, dirs=cfg_dir) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_ops_config.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_ops_config.py deleted file mode 100755 index b138047af..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_ops_config.py +++ /dev/null @@ -1,120 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import glob -import json -import argparse -import const_var - - -def load_json(json_file: str): - with open(json_file, encoding='utf-8') as file: - json_content = json.load(file) - return json_content - - -def get_specified_suffix_file(root_dir, suffix): - specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) - all_suffix_files = glob.glob(specified_suffix, recursive=True) - return all_suffix_files - - -def add_simplified_config(op_type, key, core_type, task_ration, objfile, config): - simple_cfg = config.get('binary_info_config.json') - op_cfg = simple_cfg.get(op_type) - if not op_cfg: - op_cfg = {} - op_cfg['dynamicRankSupport'] = True - op_cfg['simplifiedKeyMode'] = 0 - op_cfg['binaryList'] = [] - simple_cfg[op_type] = op_cfg - bin_list = op_cfg.get('binaryList') - if core_type == 0 and task_ration == "tilingKey": - bin_list.append({'coreType': core_type, 'simplifiedKey': key, \ - 'multiKernelType':1, 'binPath': objfile}) - else: - bin_list.append({'coreType': core_type, 'simplifiedKey': key, 'binPath': objfile}) - - -def add_op_config(op_file, bin_info, config): - op_cfg = config.get(op_file) - if not op_cfg: - op_cfg = {} - op_cfg['binList'] = [] - config[op_file] = op_cfg - op_cfg.get('binList').append(bin_info) - - -def gen_ops_config(json_file, soc, config): - core_type_map = {'MIX': 0, 'AiCore': 1, 'VectorCore': 2, 'MIX_AICORE':3, 'MIX_VECTOR_CORE':4} - contents = load_json(json_file) - if ('binFileName' not in contents) or ('supportInfo' not in contents): - return - json_base_name = os.path.basename(json_file) - op_dir = os.path.basename(os.path.dirname(json_file)) - support_info = contents.get('supportInfo') - bin_name = contents.get('binFileName') - bin_suffix = contents.get('binFileSuffix') - task_ration = contents.get('taskRation') - core_type = core_type_map.get(contents.get("coreType")) - bin_file_name = bin_name + bin_suffix - op_type = bin_name.split('_')[0] - op_file = op_dir + '.json' - bin_info = {} - keys = support_info.get('simplifiedKey') - if keys: - bin_info['simplifiedKey'] = keys - for key in keys: - add_simplified_config(op_type, key, core_type, task_ration, os.path.join(soc, op_dir, bin_file_name), \ - config) - bin_info['staticKey'] = support_info.get('staticKey') - bin_info['int64Mode'] = support_info.get('int64Mode') - bin_info['inputs'] = support_info.get('inputs') - bin_info['outputs'] = support_info.get('outputs') - if support_info.get('attrs'): - bin_info['attrs'] = support_info.get('attrs') - bin_info['binInfo'] = {'jsonFilePath': os.path.join(soc, op_dir, json_base_name)} - add_op_config(op_file, bin_info, config) - - -def gen_all_config(root_dir, soc): - suffix = 'json' - config = {} - config['binary_info_config.json'] = {} - all_json_files = get_specified_suffix_file(root_dir, suffix) - for _json in all_json_files: - gen_ops_config(_json, soc, config) - for cfg_key in config.keys(): - cfg_file = os.path.join(root_dir, cfg_key) - with os.fdopen(os.open(cfg_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - json.dump(config.get(cfg_key), fd, indent=' ') - - -def args_prase(): - parser = argparse.ArgumentParser() - parser.add_argument('-p', - '--path', - nargs='?', - required=True, - help='Parse the path of the json file.') - parser.add_argument('-s', - '--soc', - nargs='?', - required=True, - help='Parse the soc_version of ops.') - return parser.parse_args() - - -def main(): - args = args_prase() - gen_all_config(args.path, args.soc) - - -if __name__ == '__main__': - main() diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_replay_build.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_replay_build.py deleted file mode 100755 index 1cac7d911..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/ascendc_replay_build.py +++ /dev/null @@ -1,65 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import opdesc_parser -import replay_codegen -import const_var -from replay_codegen import ReplayCodeGenParams - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - - -class ReplayBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - super().__init__(op_type) - - def gen_replay_source(self: any, impl_path: str, out_path: str, ops_product: str): - if not self.op_replay_flag: - print('{} replay not enabled'.format(self.op_type)) - return - argn = len(self.input_name) + len(self.output_name) + 1 - if self.op_replay_batch: - print('{} replay in batch mode'.format(self.op_type)) - else: - print('{} replay in normal mode'.format(self.op_type)) - if impl_path.endswith('op_kernel'): - implf = os.path.join(impl_path, self.op_file + '.cpp') - tiling_file = os.path.join(impl_path, "../op_host", self.op_file + '_tiling.h') - else: - if self.dynamic_shape: - dyn_path = 'dynamic' - else: - dyn_path = '' - implf = os.path.join(impl_path, dyn_path, self.op_file + '.cpp') - tiling_file = os.path.join(impl_path, "../../op_tiling", self.op_file + '_tiling.h') - rep_conf = replay_codegen.ReplayCodeGen(ReplayCodeGenParams(self.op_type, implf, tiling_file, self.op_file, \ - self.op_intf, argn, self.op_replay_batch, self.max_block_dim, self.max_shape_size)) - rep_conf.set_batch(self.op_replay_batch) - rep_conf.set_outdir(out_path) - rep_conf.gen_replay(ops_product) - - -def gen_replay(cfgfile: str, cfgs: dict, dirs: dict, ops_product: str, ops: list = None): - batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') - iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') - op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, ReplayBuilder, ops) - for op_desc in op_descs: - op_desc.gen_replay_source(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), ops_product) - - -if __name__ == '__main__': - if len(sys.argv) <= 6: - raise RuntimeError('arguments must greater than 6') - rep_cfg = {} - rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] - rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] - rep_dir = {} - rep_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] - rep_dir[const_var.CFG_OUT_DIR] = sys.argv[5] - gen_replay(sys.argv[1], rep_cfg, rep_dir, sys.argv[6]) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/batch_replay_impl.temp b/atb_operator/AtbEyeSample/EyeOp/cmake/util/batch_replay_impl.temp deleted file mode 100755 index 0e8834664..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/batch_replay_impl.temp +++ /dev/null @@ -1,117 +0,0 @@ -#include -#include -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#include "register/op_check.h" -#define __ASCENDC_REPLAY_CODE__ -#include - -using namespace std; -using namespace optiling; -using namespace AscendCReplay; - -extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); -extern "C" int elf_batch_append(char *elf, uint32_t elfSize, char *jit, int kernum, char *atext[], int alen[], - int atlen, const char* kernelname[]); - -#define KERNEL_N 1 -#define ARG_N (__ARG_NUM__) -#define MAX_L (1024 * 1024 * 100) -#define MAX_E (1024 * 1024) - -int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) -{ - // gen type 1 : direct call codes 0: load .o file - if (param.gentype < 0 || param.gentype > 1) { - printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); - return 0; - } else if (param.gentype == 1 && param.objptr == nullptr) { - printf("Error: call replay with direct call mode, but code obj addr is null\n"); - return 0; - } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { - printf("Error: call replay with object file mode, but object file path is null\n"); - return 0; - } - // core_type 0:MIX 1:CUBE 2:VEC - if (core_type < 0 || core_type > 2) { - printf("Error: call replay core type is %d !\n", core_type); - return 0; - } - g_coreType = __CORE_TYPE__; - g_taskRation = param.task_ration; - g_tilingKey = param.tiling_key; - - unsigned char *buf, *jit; - char *kernel[KERNEL_N]; - int len[KERNEL_N]; - block_idx = 0; - block_num = param.block_dim; - g_ubBase = block_num; - uint8_t *code = (uint8_t *)malloc(MAX_L); - uint8_t *pos = code; - struct timespec tp1, tp2; - - clock_gettime(CLOCK_MONOTONIC, &tp1); - if (block_num > 32) { - printf("Error: block_num > 32\n"); - return 0; - } - //__OP_FOPEN__ - for (int i = 0; i < KERNEL_N; i++) { - //__OP_SET_KERNEL__ - for (int j = 0; j < ARG_N; j++) - AddArg(j, ARG_STEP * (j + 1)); -#ifdef FP_CEILING - SetCtrlFloatEnable(); -#else - SetCtrlFloatDisable(); -#endif - CodeInit(pos, true); - __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); - CodeEnd(); - kernel[i] = (char *)pos; - len[i] = CodeLen(); - pos += len[i]; - } - //__OP_FCLOSE__ - clock_gettime(CLOCK_MONOTONIC, &tp2); - buf = (unsigned char *)malloc(MAX_E); - int fd = open(param.entry_file, O_RDONLY); - if (fd < 0) { - printf("[error]: cannot find entry.o : %s\n", param.entry_file); - return 0; - } - uint32_t bufSize = read(fd, buf, MAX_E); - if (bufSize <= 0) { - printf("[error]: entry.o : %s is too small ! \n", param.entry_file); - } - close(fd); - jit = (unsigned char *)malloc(MAX_L); - printf("total code generated %ld\n", pos - code); - int sz = elf_batch_append((char *)buf, bufSize, (char *)jit, KERNEL_N, kernel, len, pos - code, ¶m.kernel_name); - if (tp1.tv_sec != tp2.tv_sec) { - printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); - } else { - printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); - } - printf("new elf size %d\n", sz); - if (param.gentype == 0) { - fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); - (void)write(fd, jit, sz); - close(fd); - free(jit); - } else if (param.gentype == 1) { - *param.objptr = (char*)jit; - } - free(buf); - free(code); - return sz; -} - -REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/code_channel_infer.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/code_channel_infer.py deleted file mode 100755 index d4e6a540c..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/code_channel_infer.py +++ /dev/null @@ -1,58 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" -import os -import stat -import ctypes -import collections -import shutil -import subprocess -import copy - -"""CODE_* is used to cube/vector api is called in operator code -CODE_MIX means both cube and vector api is called -CODE_CUBE means only cube api is called -CODE_VEC means only vector api is called -""" -CODE_MIX = 0 -CODE_CUBE = 1 -CODE_VEC = 2 - - -def _is_v220(op_product: str): - """return if current soc version is V220 - - Returns: - res: True means V220 - """ - if op_product in ["ascend910b", "ascend910c"]: - return True - return False - - -InfoCodeChanelParams = collections.namedtuple('InfoCodeChanelParams',\ -['src_file', 'tiling_header', 'kernel_name', 'outdir', 'op_product', 'compile_options']) - - -def infer_code_channel(params: InfoCodeChanelParams): - """get code channel for v220, return CODE_MIX if soc version is not V220 - - Args: - src_file (str): AscendC operator code file - src_file (str): AscendC operator tiling header file - kernel_name (str): kernel function name - optype (str): operator type - compile_options (list): compile options for bisheng cmd - - Raises: - Exception: if not exist L1/L0/UB if code, it's not a aicore code - - Returns: - res (int): CODE_MIX/CODE_CUBE/CODE_VEC - """ - if not _is_v220(params.op_product): - return CODE_MIX - return CODE_VEC diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/const_var.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/const_var.py deleted file mode 100755 index 3e71f5ab8..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/const_var.py +++ /dev/null @@ -1,33 +0,0 @@ - -#!/usr/bin/env python -# coding=utf-8 -""" -Function: -The replay funtion entry -Copyright Information: -Huawei Technologies Co., Ltd. All Rights Reserved © 2020 -""" - -import os -import stat - - -REPLAY_BATCH = 'batch' -REPLAY_ITERATE = 'iterate' -CFG_IMPL_DIR = 'impl_dir' -CFG_OUT_DIR = 'out_dir' -AUTO_GEN_DIR = 'auto_gen_dir' -WFLAGS = os.O_WRONLY | os.O_CREAT | os.O_TRUNC -WMODES = stat.S_IWUSR | stat.S_IRUSR -SOC_MAP_EXT = {'ascend310p': 'Ascend310P3', 'ascend310b': 'Ascend310B1', - 'ascend910': 'Ascend910A', 'ascend910b': 'Ascend910B1', 'ascend910c': 'Ascend910C1'} -BIN_CMD = 'opc $1 --main_func={fun} --input_param={param} --soc_version={soc} \ ---output=$2 --impl_mode={impl} --simplified_key_mode=0 --op_mode=dynamic\n' -CHK_CMD = ''' -if ! test -f $2/{res_file} ; then - echo "$2/{res_file} not generated!" - exit 1 -fi -''' -ATTR_DEF_VAL = {'str' : '', 'int': 0, 'float': 0.0, 'bool': False, 'list_bool': [], - 'list_int': [], 'list_float': [], 'list_list_int': [[]]} diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_impl_and_mrege_json.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_impl_and_mrege_json.sh deleted file mode 100755 index 492034e49..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_impl_and_mrege_json.sh +++ /dev/null @@ -1,20 +0,0 @@ -#!/usr/bin/bash - -project_path=$1 -build_path=$2 -vendor_name=customize -if [[ ! -d "$project_path" ]]; then - echo "[ERROR] No projcet path is provided" - exit 1 -fi - -if [[ ! -d "$build_path" ]]; then - echo "[ERROR] No build path is provided" - exit 1 -fi - -# copy aicpu kernel so operators -if [[ -d "${project_path}/cpukernel/aicpu_kernel_lib" ]]; then - cp -f ${project_path}/cpukernel/aicpu_kernel_lib/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/cpu/aicpu_kernel/impl - rm -rf ${project_path}/cpukernel/aicpu_kernel_lib -fi diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_ops_filter.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_ops_filter.sh deleted file mode 100755 index d4c27d17f..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_ops_filter.sh +++ /dev/null @@ -1,62 +0,0 @@ -#!/bin/bash -# Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -# Description: Generate npu_supported_ops.json -# ============================================================================== - -if [[ -z "$1" ]]; then - echo -e "[ERROR] No source dir provided" - exit 1 -fi - -if [[ -z "$2" ]]; then - echo -e "[ERROR] No destination dir provided" - exit 1 -fi - -src=$1 -dest_file=$2/npu_supported_ops.json - -if [ -f "$dest_file" ];then - chmod u+w $dest_file -fi - -echo $* - -add_ops() { - name=$1 - isHeavy=$2 - file=$3 - grep -w "\"$name\"" ${file} >/dev/null - if [ $? == 0 ];then - return - fi - echo " \"${name}\": {" >> ${file} - echo " \"isGray\": false," >> ${file} - echo " \"isHeavy\": ${isHeavy}" >> ${file} - echo " }," >> ${file} -} - -echo "{" > ${dest_file} -ini_files=$(find ${src} -name "*.ini") -for file in ${ini_files} ; do - name=$(grep '^\[' ${file} | sed 's/\[//g' | sed 's/]//g' | sed 's/\r//g') - grep 'heavyOp.flag' ${file} >/dev/null - if [ $? == 0 ];then - isHeavy=$(grep 'heavyOp.flag' ${file} | awk -F= '{print $2}') - else - isHeavy="false" - fi - for op in ${name} ; do - add_ops ${op} "false" ${dest_file} - done -done -echo "}" >> ${dest_file} -file_count=$(cat ${dest_file} | wc -l) -line=$(($file_count-1)) -sed -i "${line}{s/,//g}" ${dest_file} - -chmod 640 "${dest_file}" -echo -e "[INFO] Succed generated ${dest_file}" - -exit 0 - diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_version_info.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_version_info.sh deleted file mode 100755 index a06cfc78d..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/gen_version_info.sh +++ /dev/null @@ -1,6 +0,0 @@ -ascend_install_dir=$1 -gen_file_dir=$2 - -# create version.info -compiler_version=$(grep "Version" -w ${ascend_install_dir}/compiler/version.info | awk -F = '{print $2}') -echo "custom_opp_compiler_version=${compiler_version}" > ${gen_file_dir}/version.info \ No newline at end of file diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_op_info.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_op_info.py deleted file mode 100755 index 28ba08757..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_op_info.py +++ /dev/null @@ -1,36 +0,0 @@ -# -*- coding: utf-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" -import json -import os -import sys -import stat -import const_var - - -if __name__ == '__main__': - if len(sys.argv) != 3: - print(sys.argv) - print('argv error, inert_op_info.py your_op_file lib_op_file') - sys.exit(2) - - with open(sys.argv[1], 'r') as load_f: - insert_operator = json.load(load_f) - - all_operators = {} - if os.path.exists(sys.argv[2]): - if os.path.getsize(sys.argv[2]) != 0: - with open(sys.argv[2], 'r') as load_f: - all_operators = json.load(load_f) - - for k in insert_operator.keys(): - if k in all_operators.keys(): - print('replace op:[', k, '] success') - else: - print('insert op:[', k, '] success') - all_operators[k] = insert_operator[k] - - with os.fdopen(os.open(sys.argv[2], const_var.WFLAGS, const_var.WMODES), 'w') as json_file: - json_file.write(json.dumps(all_operators, indent=4)) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_simplified_keys.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_simplified_keys.py deleted file mode 100755 index 7e8dc2c55..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/insert_simplified_keys.py +++ /dev/null @@ -1,242 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import os -import re -import glob -import json -import argparse - - -DATA_TPYE_DICT = { - 'float32': 0, - 'float16': 1, - 'int8': 2, - 'int16': 6, - 'uint16': 7, - 'uint8': 4, - 'int32': 3, - 'int64': 9, - 'uint32': 8, - 'uint64': 10, - 'bool': 12, - 'double': 11, - 'complex64': 16, - 'complex128': 17, - 'qint8': 18, - 'qint16': 19, - 'qint32': 20, - 'quint8': 21, - 'quint16': 22, - 'resource': 23, - 'string': 24, - 'dual': 25, - 'variant': 26, - 'bf16': 27, - 'bfloat16': 27, - 'undefined': 28, - 'int4': 29, - 'uint1': 30, - 'int2': 31 -} - -FORMAT_DICT = { - 'NCHW': 0, - 'NHWC': 1, - 'ND': 2, - 'NC1HWC0': 3, - 'FRACTAL_Z': 4, - 'NC1C0HWPAD': 5, - 'NHWC1C0': 6, - 'FSR_NCHW': 7, - 'FRACTAL_DECONV': 8, - 'C1HWNC0': 9, - 'FRACTAL_DECONV_TRANSPOSE': 10, - 'FRACTAL_DECONV_SP_STRIDE_TRANS': 11, - 'NC1HWC0_C04': 12, - 'FRACTAL_Z_C04': 13, - 'CHWN': 14, - 'FRACTAL_DECONV_SP_STRIDE8_TRANS': 15, - 'HWCN': 16, - 'NC1KHKWHWC0': 17, - 'BN_WEIGHT': 18, - 'FILTER_HWCK': 19, - 'HASHTABLE_LOOKUP_LOOKUPS': 20, - 'HASHTABLE_LOOKUP_KEYS': 21, - 'HASHTABLE_LOOKUP_VALUE': 22, - 'HASHTABLE_LOOKUP_OUTPUT': 23, - 'HASHTABLE_LOOKUP_HITS': 24, - 'C1HWNCoC0': 25, - 'MD': 26, - 'NDHWC': 27, - 'FRACTAL_ZZ': 28, - 'FRACTAL_NZ': 29, - 'NCDHW': 30, - 'DHWCN': 31, - 'NDC1HWC0': 32, - 'FRACTAL_Z_3D': 33, - 'CN': 34, - 'NC': 35, - 'DHWNC': 36, - 'FRACTAL_Z_3D_TRANSPOSE': 37, - 'FRACTAL_ZN_LSTM': 38, - 'FRACTAL_Z_G': 39, - 'RESERVED': 40, - 'ALL': 41, - 'NULL': 42, - 'ND_RNN_BIAS': 43, - 'FRACTAL_ZN_RNN': 44, - 'NYUV': 45, - 'NYUV_A': 46 -} - - -def load_json(json_file: str): - with open(json_file, encoding='utf-8') as file: - json_content = json.load(file) - return json_content - - -def get_specified_suffix_file(root_dir, suffix): - specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) - all_suffix_files = glob.glob(specified_suffix, recursive=True) - return all_suffix_files - - -def get_deterministic_value(support_info): - deterministic_key = 'deterministic' - if deterministic_key not in support_info: - return 0 - deterministic_value = support_info.get(deterministic_key) - if deterministic_value == 'true': - return 1 - else: - return 0 - - -def get_precision_value(support_info): - precision_key = 'implMode' - precision_value = support_info.get(precision_key) - if precision_value == 'high_performance': - _value = 1 - elif precision_value == 'high_precision': - _value = 2 - else: - _value = 0 - return _value - - -def get_overflow_value(support_info): - return 0 - - -def get_parameters(info): - if info: - if 'dtype' in info: - data_type = info['dtype'] - data_type_value = DATA_TPYE_DICT.get(data_type) - else: - data_type_value = 0 - if 'format' in info: - _format = info['format'] - _format_value = FORMAT_DICT.get(_format) - else: - _format_value = 0 - else: - data_type_value = 0 - _format_value = 0 - return str(data_type_value), str(_format_value) - - -def get_dynamic_parameters(info): - # 动态输入时只需获取第一个参数 - return get_parameters(info[0]) - - -def get_all_parameters(support_info, _type): - result_list = list() - info_lists = support_info.get(_type) - if info_lists: - for _info in info_lists: - # 输入为列表时是动态输入 - if isinstance(_info, (list, tuple)): - data_type_value, _format_value = get_dynamic_parameters(_info) - else: - data_type_value, _format_value = get_parameters(_info) - result_list.append("{},{}".format(data_type_value, _format_value)) - return result_list - - -def get_all_input_parameters(support_info): - result = get_all_parameters(support_info, 'inputs') - return '/'.join(result) - - -def insert_content_into_file(input_file, content): - with open(input_file, 'r+') as file: - lines = file.readlines() - for index, line in enumerate(lines): - match_result = re.search(r'"staticKey":', line) - if match_result: - count = len(line) - len(line.lstrip()) - new_content = "{}{}".format(' ' * count, content) - # 插入到前一行,防止插入最后时还需要考虑是否添加逗号 - lines.insert(index, new_content) - break - file.seek(0) - file.write(''.join(lines)) - - -def insert_simplified_keys(json_file): - contents = load_json(json_file) - # 不存在'binFileName'或者'supportInfo'字段时,非需要替换的解析json文件 - if ('binFileName' not in contents) or ('supportInfo' not in contents): - return - support_info = contents.get('supportInfo') - bin_file_name = contents.get('binFileName') - # 'simplifiedKey'字段已经存在时,直接返回,不重复生成 - if 'simplifiedKey' in support_info: - return - op_type = bin_file_name.split('_')[0] - deterministic = str(get_deterministic_value(support_info)) - precision = str(get_precision_value(support_info)) - overflow = str(get_overflow_value(support_info)) - input_parameters = get_all_input_parameters(support_info) - key = '{}/d={},p={},o={}/{}/'.format( - op_type, - deterministic, - precision, - overflow, - input_parameters) - result = '"simplifiedKey": "' + key + '",\n' - insert_content_into_file(json_file, result) - - -def insert_all_simplified_keys(root_dir): - suffix = 'json' - all_json_files = get_specified_suffix_file(root_dir, suffix) - for _json in all_json_files: - insert_simplified_keys(_json) - - -def args_prase(): - parser = argparse.ArgumentParser() - parser.add_argument('-p', - '--path', - nargs='?', - required=True, - help='Parse the path of the json file.') - return parser.parse_args() - - -def main(): - args = args_prase() - insert_all_simplified_keys(args.path) - - -if __name__ == '__main__': - main() diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_entry.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_entry.py deleted file mode 100755 index 2b77c970d..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_entry.py +++ /dev/null @@ -1,115 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - - -def gen_fun_def(title, kernel, argn, arg_type, arg_name): - entry = [] - entry.append(title) - entry.append(kernel) - entry.append('(') - args = [] - for i in range(0, argn): - args.append(arg_type + ' ' + arg_name + str(i)) - entry.append(', '.join(args)) - entry.append(')') - return ' '.join(entry) - - -def gen_batch_kernel_body(fname, argn, arg_name): - body = [] - body.append('{') - fun = [] - fun.append(fname) - fun.append('(') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - fun.append(', '.join(args)) - fun.append(');') - body.append(' '.join(fun)) - body.append('}') - return '\n'.join(body) - - -def gen_mc_kernel_body(kn, argn, arg_name, blknum): - body = [] - body.append('{') - body.append(' switch(block_idx) {') - for blk in range(0, blknum): - fun = [] - fun.append('{}_blk{:02d}'.format(kn, blk)) - fun.append('(') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - fun.append(', '.join(args)) - fun.append(')') - body.append(' case {}: {}; break;'.format(blk, ' '.join(fun))) - body.append(' default: break;') - body.append(' }') - body.append('}') - return '\n'.join(body) - - -def gen_proc_body(argn, arg_name): - body = [] - body.append('{') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - body.append('uint64_t __x = (uint64_t)' + ' + (uint64_t)'.join(args) + ';') - body.append('__asm__ ("NOP");') - body.append('__asm__ ("NOP");') - body.append('__asm__ ("NOP");') - body.append('}') - return '\n'.join(body) - - -def batch_code_gen(kn, argn, argt): - codes = [] - kernel_name = kn - proc_name = kernel_name + '_percore' - arg_num = int(argn) - data_type = argt - arg_type = '__gm__ ' + data_type + '* __restrict__' - arg_name = 'arg' - kernel_title = 'extern \"C\" __global__ __aicore__ void' - proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' - codes.append('#ifndef __aicore__') - codes.append('#define __aicore__ [aicore]') - codes.append('#endif') - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') - codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) - codes.append(gen_batch_kernel_body(proc_name, arg_num, arg_name)) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) - codes.append(gen_proc_body(arg_num, arg_name)) - return '\n'.join(codes) + '\n' - - -def mc_code_gen(kn, argn, argt, blknum): - codes = [] - kernel_name = kn - core_num = int(blknum) - arg_num = int(argn) - data_type = argt - arg_type = '__gm__ ' + data_type + '* __restrict__' - arg_name = 'arg' - kernel_title = 'extern \"C\" __global__ __aicore__ void' - proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' - codes.append('#ifndef __aicore__') - codes.append('#define __aicore__ [aicore]') - codes.append('#endif') - for i in range(0, core_num): - proc_name = '{}_blk{:02d}'.format(kernel_name, i) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') - codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) - codes.append(gen_mc_kernel_body(kernel_name, arg_num, arg_name, core_num)) - for i in range(0, core_num): - proc_name = '{}_blk{:02d}'.format(kernel_name, i) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) - codes.append(gen_proc_body(arg_num, arg_name)) - return '\n'.join(codes) + '\n' diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_impl.temp b/atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_impl.temp deleted file mode 100755 index 5079a1043..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/kernel_impl.temp +++ /dev/null @@ -1,10 +0,0 @@ -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#define __ASCENDC_REPLAY_CODE__ -#include "__CCE_FILE__" diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/COPYING b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/COPYING deleted file mode 100755 index d159169d1..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/COPYING +++ /dev/null @@ -1,339 +0,0 @@ - GNU GENERAL PUBLIC LICENSE - Version 2, June 1991 - - Copyright (C) 1989, 1991 Free Software Foundation, Inc., - 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA - Everyone is permitted to copy and distribute verbatim copies - of this license document, but changing it is not allowed. - - Preamble - - The licenses for most software are designed to take away your -freedom to share and change it. By contrast, the GNU General Public -License is intended to guarantee your freedom to share and change free -software--to make sure the software is free for all its users. This -General Public License applies to most of the Free Software -Foundation's software and to any other program whose authors commit to -using it. (Some other Free Software Foundation software is covered by -the GNU Lesser General Public License instead.) You can apply it to -your programs, too. - - When we speak of free software, we are referring to freedom, not -price. Our General Public Licenses are designed to make sure that you -have the freedom to distribute copies of free software (and charge for -this service if you wish), that you receive source code or can get it -if you want it, that you can change the software or use pieces of it -in new free programs; and that you know you can do these things. - - To protect your rights, we need to make restrictions that forbid -anyone to deny you these rights or to ask you to surrender the rights. -These restrictions translate to certain responsibilities for you if you -distribute copies of the software, or if you modify it. - - For example, if you distribute copies of such a program, whether -gratis or for a fee, you must give the recipients all the rights that -you have. You must make sure that they, too, receive or can get the -source code. And you must show them these terms so they know their -rights. - - We protect your rights with two steps: (1) copyright the software, and -(2) offer you this license which gives you legal permission to copy, -distribute and/or modify the software. - - Also, for each author's protection and ours, we want to make certain -that everyone understands that there is no warranty for this free -software. If the software is modified by someone else and passed on, we -want its recipients to know that what they have is not the original, so -that any problems introduced by others will not reflect on the original -authors' reputations. - - Finally, any free program is threatened constantly by software -patents. We wish to avoid the danger that redistributors of a free -program will individually obtain patent licenses, in effect making the -program proprietary. To prevent this, we have made it clear that any -patent must be licensed for everyone's free use or not licensed at all. - - The precise terms and conditions for copying, distribution and -modification follow. - - GNU GENERAL PUBLIC LICENSE - TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION - - 0. This License applies to any program or other work which contains -a notice placed by the copyright holder saying it may be distributed -under the terms of this General Public License. The "Program", below, -refers to any such program or work, and a "work based on the Program" -means either the Program or any derivative work under copyright law: -that is to say, a work containing the Program or a portion of it, -either verbatim or with modifications and/or translated into another -language. (Hereinafter, translation is included without limitation in -the term "modification".) Each licensee is addressed as "you". - -Activities other than copying, distribution and modification are not -covered by this License; they are outside its scope. The act of -running the Program is not restricted, and the output from the Program -is covered only if its contents constitute a work based on the -Program (independent of having been made by running the Program). -Whether that is true depends on what the Program does. - - 1. You may copy and distribute verbatim copies of the Program's -source code as you receive it, in any medium, provided that you -conspicuously and appropriately publish on each copy an appropriate -copyright notice and disclaimer of warranty; keep intact all the -notices that refer to this License and to the absence of any warranty; -and give any other recipients of the Program a copy of this License -along with the Program. - -You may charge a fee for the physical act of transferring a copy, and -you may at your option offer warranty protection in exchange for a fee. - - 2. You may modify your copy or copies of the Program or any portion -of it, thus forming a work based on the Program, and copy and -distribute such modifications or work under the terms of Section 1 -above, provided that you also meet all of these conditions: - - a) You must cause the modified files to carry prominent notices - stating that you changed the files and the date of any change. - - b) You must cause any work that you distribute or publish, that in - whole or in part contains or is derived from the Program or any - part thereof, to be licensed as a whole at no charge to all third - parties under the terms of this License. - - c) If the modified program normally reads commands interactively - when run, you must cause it, when started running for such - interactive use in the most ordinary way, to print or display an - announcement including an appropriate copyright notice and a - notice that there is no warranty (or else, saying that you provide - a warranty) and that users may redistribute the program under - these conditions, and telling the user how to view a copy of this - License. (Exception: if the Program itself is interactive but - does not normally print such an announcement, your work based on - the Program is not required to print an announcement.) - -These requirements apply to the modified work as a whole. If -identifiable sections of that work are not derived from the Program, -and can be reasonably considered independent and separate works in -themselves, then this License, and its terms, do not apply to those -sections when you distribute them as separate works. But when you -distribute the same sections as part of a whole which is a work based -on the Program, the distribution of the whole must be on the terms of -this License, whose permissions for other licensees extend to the -entire whole, and thus to each and every part regardless of who wrote it. - -Thus, it is not the intent of this section to claim rights or contest -your rights to work written entirely by you; rather, the intent is to -exercise the right to control the distribution of derivative or -collective works based on the Program. - -In addition, mere aggregation of another work not based on the Program -with the Program (or with a work based on the Program) on a volume of -a storage or distribution medium does not bring the other work under -the scope of this License. - - 3. You may copy and distribute the Program (or a work based on it, -under Section 2) in object code or executable form under the terms of -Sections 1 and 2 above provided that you also do one of the following: - - a) Accompany it with the complete corresponding machine-readable - source code, which must be distributed under the terms of Sections - 1 and 2 above on a medium customarily used for software interchange; or, - - b) Accompany it with a written offer, valid for at least three - years, to give any third party, for a charge no more than your - cost of physically performing source distribution, a complete - machine-readable copy of the corresponding source code, to be - distributed under the terms of Sections 1 and 2 above on a medium - customarily used for software interchange; or, - - c) Accompany it with the information you received as to the offer - to distribute corresponding source code. (This alternative is - allowed only for noncommercial distribution and only if you - received the program in object code or executable form with such - an offer, in accord with Subsection b above.) - -The source code for a work means the preferred form of the work for -making modifications to it. For an executable work, complete source -code means all the source code for all modules it contains, plus any -associated interface definition files, plus the scripts used to -control compilation and installation of the executable. However, as a -special exception, the source code distributed need not include -anything that is normally distributed (in either source or binary -form) with the major components (compiler, kernel, and so on) of the -operating system on which the executable runs, unless that component -itself accompanies the executable. - -If distribution of executable or object code is made by offering -access to copy from a designated place, then offering equivalent -access to copy the source code from the same place counts as -distribution of the source code, even though third parties are not -compelled to copy the source along with the object code. - - 4. You may not copy, modify, sublicense, or distribute the Program -except as expressly provided under this License. Any attempt -otherwise to copy, modify, sublicense or distribute the Program is -void, and will automatically terminate your rights under this License. -However, parties who have received copies, or rights, from you under -this License will not have their licenses terminated so long as such -parties remain in full compliance. - - 5. You are not required to accept this License, since you have not -signed it. However, nothing else grants you permission to modify or -distribute the Program or its derivative works. These actions are -prohibited by law if you do not accept this License. Therefore, by -modifying or distributing the Program (or any work based on the -Program), you indicate your acceptance of this License to do so, and -all its terms and conditions for copying, distributing or modifying -the Program or works based on it. - - 6. Each time you redistribute the Program (or any work based on the -Program), the recipient automatically receives a license from the -original licensor to copy, distribute or modify the Program subject to -these terms and conditions. You may not impose any further -restrictions on the recipients' exercise of the rights granted herein. -You are not responsible for enforcing compliance by third parties to -this License. - - 7. If, as a consequence of a court judgment or allegation of patent -infringement or for any other reason (not limited to patent issues), -conditions are imposed on you (whether by court order, agreement or -otherwise) that contradict the conditions of this License, they do not -excuse you from the conditions of this License. If you cannot -distribute so as to satisfy simultaneously your obligations under this -License and any other pertinent obligations, then as a consequence you -may not distribute the Program at all. For example, if a patent -license would not permit royalty-free redistribution of the Program by -all those who receive copies directly or indirectly through you, then -the only way you could satisfy both it and this License would be to -refrain entirely from distribution of the Program. - -If any portion of this section is held invalid or unenforceable under -any particular circumstance, the balance of the section is intended to -apply and the section as a whole is intended to apply in other -circumstances. - -It is not the purpose of this section to induce you to infringe any -patents or other property right claims or to contest validity of any -such claims; this section has the sole purpose of protecting the -integrity of the free software distribution system, which is -implemented by public license practices. Many people have made -generous contributions to the wide range of software distributed -through that system in reliance on consistent application of that -system; it is up to the author/donor to decide if he or she is willing -to distribute software through any other system and a licensee cannot -impose that choice. - -This section is intended to make thoroughly clear what is believed to -be a consequence of the rest of this License. - - 8. If the distribution and/or use of the Program is restricted in -certain countries either by patents or by copyrighted interfaces, the -original copyright holder who places the Program under this License -may add an explicit geographical distribution limitation excluding -those countries, so that distribution is permitted only in or among -countries not thus excluded. In such case, this License incorporates -the limitation as if written in the body of this License. - - 9. The Free Software Foundation may publish revised and/or new versions -of the General Public License from time to time. Such new versions will -be similar in spirit to the present version, but may differ in detail to -address new problems or concerns. - -Each version is given a distinguishing version number. If the Program -specifies a version number of this License which applies to it and "any -later version", you have the option of following the terms and conditions -either of that version or of any later version published by the Free -Software Foundation. If the Program does not specify a version number of -this License, you may choose any version ever published by the Free Software -Foundation. - - 10. If you wish to incorporate parts of the Program into other free -programs whose distribution conditions are different, write to the author -to ask for permission. For software which is copyrighted by the Free -Software Foundation, write to the Free Software Foundation; we sometimes -make exceptions for this. Our decision will be guided by the two goals -of preserving the free status of all derivatives of our free software and -of promoting the sharing and reuse of software generally. - - NO WARRANTY - - 11. BECAUSE THE PROGRAM IS LICENSED FREE OF CHARGE, THERE IS NO WARRANTY -FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN -OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES -PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED -OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF -MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS -TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE -PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, -REPAIR OR CORRECTION. - - 12. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING -WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MAY MODIFY AND/OR -REDISTRIBUTE THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, -INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING -OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED -TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY -YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER -PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE -POSSIBILITY OF SUCH DAMAGES. - - END OF TERMS AND CONDITIONS - - How to Apply These Terms to Your New Programs - - If you develop a new program, and you want it to be of the greatest -possible use to the public, the best way to achieve this is to make it -free software which everyone can redistribute and change under these terms. - - To do so, attach the following notices to the program. It is safest -to attach them to the start of each source file to most effectively -convey the exclusion of warranty; and each file should have at least -the "copyright" line and a pointer to where the full notice is found. - - - Copyright (C) - - This program is free software; you can redistribute it and/or modify - it under the terms of the GNU General Public License as published by - the Free Software Foundation; either version 2 of the License, or - (at your option) any later version. - - This program is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the - GNU General Public License for more details. - - You should have received a copy of the GNU General Public License along - with this program; if not, write to the Free Software Foundation, Inc., - 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. - -Also add information on how to contact you by electronic and paper mail. - -If the program is interactive, make it output a short notice like this -when it starts in an interactive mode: - - Gnomovision version 69, Copyright (C) year name of author - Gnomovision comes with ABSOLUTELY NO WARRANTY; for details type `show w'. - This is free software, and you are welcome to redistribute it - under certain conditions; type `show c' for details. - -The hypothetical commands `show w' and `show c' should show the appropriate -parts of the General Public License. Of course, the commands you use may -be called something other than `show w' and `show c'; they could even be -mouse-clicks or menu items--whatever suits your program. - -You should also get your employer (if you work as a programmer) or your -school, if any, to sign a "copyright disclaimer" for the program, if -necessary. Here is a sample; alter the names: - - Yoyodyne, Inc., hereby disclaims all copyright interest in the program - `Gnomovision' (which makes passes at compilers) written by James Hacker. - - , 1 April 1989 - Ty Coon, President of Vice - -This General Public License does not permit incorporating your program into -proprietary programs. If your program is a subroutine library, you may -consider it more useful to permit linking proprietary applications with the -library. If this is what you want to do, use the GNU Lesser General -Public License instead of this License. diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/README.md b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/README.md deleted file mode 100755 index b41f01682..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/README.md +++ /dev/null @@ -1,246 +0,0 @@ -[![License: GPL v2](https://img.shields.io/badge/License-GPL%20v2-blue.svg)](https://www.gnu.org/licenses/old-licenses/gpl-2.0.en.html) -![Build Status](https://github.com/megastep/makeself/workflows/CI/badge.svg) - -# makeself - Make self-extractable archives on Unix - -[makeself.sh][1] is a small shell script that generates a self-extractable -compressed tar archive from a directory. The resulting file appears as a shell script -(many of those have a **.run** suffix), and can be launched as is. The archive -will then uncompress itself to a temporary directory and an optional arbitrary -command will be executed (for example an installation script). This is pretty -similar to archives generated with WinZip Self-Extractor in the Windows world. -Makeself archives also include checksums for integrity self-validation (CRC -and/or MD5/SHA256 checksums). - -The makeself.sh script itself is used only to create the archives from a -directory of files. The resultant archive is actually a compressed (using -gzip, bzip2, or compress) TAR archive, with a small shell script stub at the -beginning. This small stub performs all the steps of extracting the files, -running the embedded command, and removing the temporary files when done. -All the user has to do to install the software contained in such an -archive is to "run" the archive, i.e **sh nice-software.run**. I recommend -using the ".run" (which was introduced by some Makeself archives released by -Loki Software) or ".sh" suffix for such archives not to confuse the users, -so that they will know they are actually shell scripts (with quite a lot of binary data -attached to them though!). - -I am trying to keep the code of this script as portable as possible, i.e it is -not relying on any bash-specific features and only calls commands that are -installed on any functioning UNIX-compatible system. This script as well as -the archives it generates should run on any Unix flavor, with any compatible -Bourne shell, provided of course that the compression programs are available. - -As of version 2.1, Makeself has been rewritten and tested on the following -platforms : - - * Linux (all distributions) - * Sun Solaris (8 and above) - * HP-UX (tested on 11.0 and 11i on HPPA RISC) - * SCO OpenUnix and OpenServer - * IBM AIX 5.1L - * macOS (Darwin) - * SGI IRIX 6.5 - * FreeBSD - * UnicOS / Cray - * Cygwin (Windows) - -If you successfully run Makeself and/or archives created with it on another -system, then please [let me know][2]! - -Examples of publicly available archives made using makeself are : - - * Game patches and installers for [Id Software][3] games like Quake 3 for Linux or Return To Castle Wolfenstein ; - * All game patches released by [Loki Software][4] for the Linux version of popular games ; - * The [nVidia drivers][5] for Linux - * The installer for the Linux version of [Google Earth][6] - * The [VirtualBox][7] installers for Linux - * The [Makeself][1] distribution itself ;-) - * and countless others... - -**Important note for Apache users:** By default, most Web servers will think that Makeself archives are regular text files and thus they may show up as text in a Web browser. The correct way to prevent this is to add a MIME type for this file format, like so (in httpd.conf) : - -`AddType application/x-makeself .run` - -**Important note for certain GNU/Linux distributions:** Archives created with Makeself prior to v2.1.2 were using an old syntax for the _head_ and _tail_ Unix commands that is being progressively obsoleted in their GNU forms. Therefore you may have problems uncompressing some of these archives. A workaround for this is to set the environment variable $_POSIX2_VERSION to enable the old syntax, i.e. : - -`export _POSIX2_VERSION=199209` - -## Usage - -The syntax of makeself is the following: - -``` -makeself.sh [args] archive_dir file_name label startup_script [script_args] -``` - - * _args_ are optional options for Makeself. The available ones are : - - * **`--version`** : Prints the version number on stdout, then exits immediately - * **`--gzip`** : Use gzip for compression (the default on platforms on which gzip is commonly available, like Linux) - * **`--bzip2`** : Use bzip2 instead of gzip for better compression. The bzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. - * **`--pbzip2`** : Use pbzip2 instead of gzip for better and faster compression on machines having multiple CPUs. The pbzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. - * **`--xz`** : Use xz instead of gzip for better compression. The xz command must be available in the command path. It is recommended that the archive prefix be set to something like '.xz.run' for the archive, so that potential users know that they'll need xz to extract it. - * **`--lzo`** : Use lzop instead of gzip for better compression. The lzop command must be available in the command path. It is recommended that the archive prefix be set to something like `.lzo.run` for the archive, so that potential users know that they'll need lzop to extract it. - * **`--lz4`** : Use lz4 instead of gzip for better compression. The lz4 command must be available in the command path. It is recommended that the archive prefix be set to something like '.lz4.run' for the archive, so that potential users know that they'll need lz4 to extract it. - * **`--zstd`** : Use zstd instead of gzip for better compression. The zstd command must be available in the command path. It is recommended that the archive prefix be set to something like '.zstd.run' for the archive, so that potential users know that they'll need zstd to extract it. - * **`--pigz`** : Use pigz for compression. - * **`--base64`** : Encode the archive to ASCII in Base64 format instead of compressing (base64 command required). - * **`--gpg-encrypt`** : Encrypt the archive using `gpg -ac -z $COMPRESS_LEVEL`. This will prompt for a password to encrypt with. Assumes that potential users have `gpg` installed. - * **`--ssl-encrypt`** : Encrypt the archive using `openssl aes-256-cbc -a -salt`. This will prompt for a password to encrypt with. Assumes that the potential users have the OpenSSL tools installed. - * **`--compress`** : Use the UNIX `compress` command to compress the data. This should be the default on all platforms that don't have gzip available. - * **`--nocomp`** : Do not use any compression for the archive, which will then be an uncompressed TAR. - * **`--complevel`** : Specify the compression level for gzip, bzip2, pbzip2, zstd, xz, lzo or lz4. (defaults to 9) - * **`--threads`** : Specify the number of threads to be used by compressors that support parallelization. Omit to use compressor's default. Most useful (and required) for opting into xz's threading, usually with `--threads=0` for all available cores. pbzip2 and pigz are parallel by default, and setting this value allows limiting the number of threads they use. - * **`--notemp`** : The generated archive will not extract the files to a temporary directory, but in a new directory created in the current directory. This is better to distribute software packages that may extract and compile by themselves (i.e. launch the compilation through the embedded script). - * **`--current`** : Files will be extracted to the current directory, instead of in a subdirectory. This option implies `--notemp` above. - * **`--follow`** : Follow the symbolic links inside of the archive directory, i.e. store the files that are being pointed to instead of the links themselves. - * **`--append`** _(new in 2.1.x)_: Append data to an existing archive, instead of creating a new one. In this mode, the settings from the original archive are reused (compression type, label, embedded script), and thus don't need to be specified again on the command line. - * **`--header`** : Makeself uses a separate file to store the header stub, called `makeself-header.sh`. By default, it is assumed that it is stored in the same location as makeself.sh. This option can be used to specify its actual location if it is stored someplace else. - * **`--cleanup`** : Specify a script that is run when execution is interrupted or finishes successfully. The script is executed with the same environment and initial `script_args` as `startup_script`. - * **`--copy`** : Upon extraction, the archive will first extract itself to a temporary directory. The main application of this is to allow self-contained installers stored in a Makeself archive on a CD, when the installer program will later need to unmount the CD and allow a new one to be inserted. This prevents "Filesystem busy" errors for installers that span multiple CDs. - * **`--nox11`** : Disable the automatic spawning of a new terminal in X11. - * **`--nowait`** : When executed from a new X11 terminal, disable the user prompt at the end of the script execution. - * **`--nomd5`** and **`--nocrc`** : Disable the creation of a MD5 / CRC checksum for the archive. This speeds up the extraction process if integrity checking is not necessary. - * **`--sha256`** : Adds a SHA256 checksum for the archive. This is in addition to the MD5 / CRC checksums unless `--nomd5` is also used. - * **`--lsm` _file_** : Provide and LSM file to makeself, that will be embedded in the generated archive. LSM files are describing a software package in a way that is easily parseable. The LSM entry can then be later retrieved using the `--lsm` argument to the archive. An example of a LSM file is provided with Makeself. - * **`--tar-format opt`** : Specify the tar archive format (default is ustar); you may use any value accepted by your tar command (such as posix, v7, etc). - * **`--tar-extra opt`** : Append more options to the tar command line. - - For instance, in order to exclude the `.git` directory from the packaged archive directory using the GNU `tar`, one can use `makeself.sh --tar-extra "--exclude=.git" ...` - - * **`--keep-umask`** : Keep the umask set to shell default, rather than overriding when executing self-extracting archive. - * **`--packaging-date date`** : Use provided string as the packaging date instead of the current date. - * **`--license`** : Append a license file. - * **`--nooverwrite`** : Do not extract the archive if the specified target directory already exists. - * **`--help-header file`** : Add a header to the archive's `--help` output. - * `archive_dir` is the name of the directory that contains the files to be archived - * `file_name` is the name of the archive to be created - * `label` is an arbitrary text string describing the package. It will be displayed while extracting the files. - * `startup_script` is the command to be executed _from within_ the directory of extracted files. Thus, if you wish to execute a program contained in this directory, you must prefix your command with `./`. For example, `./program` will be fine. The `script_args` are additional arguments for this command. - -Here is an example, assuming the user has a package image stored in a **/home/joe/mysoft**, and he wants to generate a self-extracting package named -**mysoft.sh**, which will launch the "setup" script initially stored in /home/joe/mysoft : - -`makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup -` - -Here is also how I created the [makeself.run][9] archive which contains the Makeself distribution : - -`makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" ` - -Archives generated with Makeself can be passed the following arguments: - - * **`--keep`** : Prevent the files to be extracted in a temporary directory that will be removed after the embedded script's execution. The files will then be extracted in the current working directory and will stay here until you remove them. - * **`--verbose`** : Will prompt the user before executing the embedded command - * **`--target dir`** : Allows to extract the archive in an arbitrary place. - * **`--nox11`** : Do not spawn a X11 terminal. - * **`--confirm`** : Prompt the user for confirmation before running the embedded command. - * **`--info`** : Print out general information about the archive (does not extract). - * **`--lsm`** : Print out the LSM entry, if it is present. - * **`--list`** : List the files in the archive. - * **`--check`** : Check the archive for integrity using the embedded checksums. Does not extract the archive. - * **`--nochown`** : By default, a `chown -R` command is run on the target directory after extraction, so that all files belong to the current user. This is mostly needed if you are running as root, as tar will then try to recreate the initial user ownerships. You may disable this behavior with this flag. - * **`--tar`** : Run the tar command on the contents of the archive, using the following arguments as parameter for the command. - * **`--noexec`** : Do not run the embedded script after extraction. - * **`--noexec-cleanup`** : Do not run the embedded cleanup script. - * **`--nodiskspace`** : Do not check for available disk space before attempting to extract. - * **`--cleanup-args`** : Specify arguments to be passed to the cleanup script. Wrap value in quotes to specify multiple arguments. - -Any subsequent arguments to the archive will be passed as additional arguments to the embedded command. You must explicitly use the `--` special command-line construct before any such options to make sure that Makeself will not try to interpret them. - -## Startup Script - -The startup script must be a regular Shell script. - -Within the startup script, you can use the `$USER_PWD` variable to get the path of the folder from which the self-extracting script is executed. This is especially useful to access files that are located in the same folder as the script, as shown in the example below. - -`my-self-extracting-script.sh --fooBarFileParameter foo.bar` - -## Building and Testing - -Clone the git repo and execute `git submodule update --init --recursive` to obtain all submodules. - -* To make a release: `make` -* To run all tests: `make test` - -## Maven Usage - -Makeself is now supported by the following maven plugin [makeself-maven-plugin](https://github.com/hazendaz/makeself-maven-plugin). Please refer to project for usage and report any bugs in regards to maven plugin on that project. - -## License - -Makeself itself is covered by the [GNU General Public License][8] (GPL) version 2 and above. Archives generated by Makeself don't have to be placed under this license (although I encourage it ;-)), since the archive itself is merely data for Makeself. - -## Contributing - -I will gladly consider merging your pull requests on the [GitHub][10] repository. However, please keep the following in mind: - - * One of the main purposes of Makeself is portability. Do not submit patches that will break supported platforms. The more platform-agnostic, the better. - * Please explain clearly what the purpose of the patch is, and how you achieved it. - -## Download - -Get the latest official distribution [here][9] (version 2.4.2). - -The latest development version can be grabbed from [GitHub][10]. Feel free to submit any patches there through the fork and pull request process. - -## Version history - - * **v1.0:** Initial public release - * **v1.1:** The archive can be passed parameters that will be passed on to the embedded script, thanks to John C. Quillan - * **v1.2:** Cosmetic updates, support for bzip2 compression and non-temporary archives. Many ideas thanks to Francois Petitjean. - * **v1.3:** More patches from Bjarni R. Einarsson and Francois Petitjean: Support for no compression (`--nocomp`), script is no longer mandatory, automatic launch in an xterm, optional verbose output, and -target archive option to indicate where to extract the files. - * **v1.4:** Many patches from Francois Petitjean: improved UNIX compatibility, automatic integrity checking, support of LSM files to get info on the package at run time.. - * **v1.5.x:** A lot of bugfixes, and many other patches, including automatic verification through the usage of checksums. Version 1.5.5 was the stable release for a long time, even though the Web page didn't get updated ;-). Makeself was also officially made a part of the [Loki Setup installer][11], and its source is being maintained as part of this package. - * **v2.0:** Complete internal rewrite of Makeself. The command-line parsing was vastly improved, the overall maintenance of the package was greatly improved by separating the stub from makeself.sh. Also Makeself was ported and tested to a variety of Unix platforms. - * **v2.0.1:** First public release of the new 2.0 branch. Prior versions are officially obsoleted. This release introduced the `--copy` argument that was introduced in response to a need for the [UT2K3][12] Linux installer. - * **v2.1.0:** Big change : Makeself can now support multiple embedded tarballs, each stored separately with their own checksums. An existing archive can be updated with the `--append` flag. Checksums are also better managed, and the `--nochown` option for archives appeared. - * **v2.1.1:** Fixes related to the Unix compression (compress command). Some Linux distributions made the insane choice to make it unavailable, even though gzip is capable of uncompressing these files, plus some more bugfixes in the extraction and checksum code. - * **v2.1.2:** Some bug fixes. Use head -n to avoid problems with POSIX conformance. - * **v2.1.3:** Bug fixes with the command line when spawning terminals. Added `--tar`, `--noexec` for archives. Added `--nomd5` and `--nocrc` to avoid creating checksums in archives. The embedded script is now run through "eval". The `--info` output now includes the command used to create the archive. A man page was contributed by Bartosz Fenski. - * **v2.1.4:** Fixed `--info` output. Generate random directory name when extracting files to . to avoid problems. Better handling of errors with wrong permissions for the directory containing the files. Avoid some race conditions, Unset the $CDPATH variable to avoid problems if it is set. Better handling of dot files in the archive directory. - * **v2.1.5:** Made the md5sum detection consistent with the header code. Check for the presence of the archive directory. Added `--encrypt` for symmetric encryption through gpg (Eric Windisch). Added support for the digest command on Solaris 10 for MD5 checksums. Check for available disk space before extracting to the target directory (Andreas Schweitzer). Allow extraction to run asynchronously (patch by Peter Hatch). Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo). - * **v2.1.6:** Replaced one dot per file progress with a realtime progress percentage and a spinning cursor. Added `--noprogress` to prevent showing the progress during the decompression. Added `--target` dir to allow extracting directly to a target directory. (Guy Baconniere) - * **v2.2.0:** First major new release in years! Includes many bugfixes and user contributions. Please look at the [project page on Github][10] for all the details. - * **v2.3.0:** Support for archive encryption via GPG or OpenSSL. Added LZO and LZ4 compression support. Options to set the packaging date and stop the umask from being overriden. Optionally ignore check for available disk space when extracting. New option to check for root permissions before extracting. - * **v2.3.1:** Various compatibility updates. Added unit tests for Travis CI in the GitHub repo. New `--tar-extra`, `--untar-extra`, `--gpg-extra`, `--gpg-asymmetric-encrypt-sign` options. - * **v2.4.0:** Added optional support for SHA256 archive integrity checksums. - * **v2.4.2:** New --cleanup and --cleanup-args arguments for cleanup scripts. Added threading support for supported compressors. Now supports zstd compression. - * **v2.4.3:** Make explicit POSIX tar archives for increased compatibility. - * **v2.4.4:** Fixed various compatibility issues (no longer use POSIX tar archives), Github Actions to check on Solaris and FreeBSD. - * **v2.4.5:** Added `--tar-format` option to set the tar archive format (default is ustar) - -## Links - - * Check out the ["Loki Setup"][11] installer, used to install many Linux games and other applications, and of which I am the co-author. Since the demise of Loki, I am now the official maintainer of the project, and it is now being hosted here on GitHub. - * Bjarni R. Einarsson also wrote the **setup.sh** installer script, inspired by Makeself. [Check it out !][14] - -## Contact - -This script was written by [Stéphane Peter][15] (megastep at megastep.org). Any enhancements and suggestions are welcome. - -Contributions were included from John C. Quillan, Bjarni R. Einarsson, -Francois Petitjean, Ryan C. Gordon, and many contributors on GitHub. If you think I forgot -your name, don't hesitate to contact me. - -This project is now hosted on GitHub. Feel free to submit patches and bug reports on the [project page][10]. - -* * * - -[Stephane Peter][2] - - [1]: http://makeself.io/ - [2]: mailto:megastep@megastep.org - [3]: http://www.idsoftware.com/ - [4]: http://www.lokigames.com/products/myth2/updates.php3 - [5]: http://www.nvidia.com/ - [6]: http://earth.google.com/ - [7]: http://www.virtualbox.org/ - [8]: http://www.gnu.org/copyleft/gpl.html - [9]: https://github.com/megastep/makeself/releases/download/release-2.4.5/makeself-2.4.5.run - [10]: https://github.com/megastep/makeself - [11]: https://github.com/megastep/loki_setup/ - [12]: http://www.unrealtournament2003.com/ - [13]: http://www.icculus.org/ - [14]: http://bre.klaki.net/programs/setup.sh/ - [15]: https://stephanepeter.com/ diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/VERSION b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/VERSION deleted file mode 100755 index 59aa62c1f..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/VERSION +++ /dev/null @@ -1 +0,0 @@ -2.4.5 diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/make-release.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/make-release.sh deleted file mode 100755 index b5692d490..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/make-release.sh +++ /dev/null @@ -1,9 +0,0 @@ -#!/bin/sh -# -# Create a distributable archive of the current version of Makeself - -VER=`cat VERSION` -mkdir -p /tmp/makeself-$VER release -cp -pPR makeself* test README.md COPYING VERSION .gitmodules /tmp/makeself-$VER/ -./makeself.sh --notemp /tmp/makeself-$VER release/makeself-$VER.run "Makeself v$VER" echo "Makeself has extracted itself" - diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself-header.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself-header.sh deleted file mode 100755 index 940903148..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself-header.sh +++ /dev/null @@ -1,660 +0,0 @@ -cat << EOF > "$archname" -#!/bin/bash -# This script was generated using Makeself $MS_VERSION -# The license covering this archive and its contents, if any, is wholly independent of the Makeself license (GPL) -# 2022.3.19-Modified the MS_Help function and some options -# Huawei Technologies Co., Ltd. - -ORIG_UMASK=\`umask\` - -CRCsum="$CRCsum" -MD5="$MD5sum" -SHA="$SHAsum" -SIGNATURE="$Signature" -TMPROOT=\${TMPDIR:="\$HOME"} -if ! test -d "\$TMPROOT"; then - TMPROOT="\$PWD" -fi -export TMPDIR="\$TMPROOT" -USER_PWD="\$PWD" -if ! test -d "\$USER_PWD"; then - exit 1 -fi -export USER_PWD -ARCHIVE_DIR=\`dirname "\$0"\` -export ARCHIVE_DIR - -name_of_file="\$0 " -pwd_of_file="\$PWD" -label="$LABEL" -script="$SCRIPT" -scriptargs="$SCRIPTARGS" -cleanup_script="${CLEANUP_SCRIPT}" -licensetxt="$LICENSE" -helpheader='$HELPHEADER' -targetdir="$archdirname" -filesizes="$filesizes" -totalsize="$totalsize" -keep="$KEEP" -nooverwrite="$NOOVERWRITE" -quiet="n" -accept="n" -nodiskspace="n" -export_conf="$EXPORT_CONF" -decrypt_cmd="$DECRYPT_CMD" -skip="$SKIP" - -print_cmd_arg="" -if type printf > /dev/null; then - print_cmd="printf" -elif test -x /usr/ucb/echo; then - print_cmd="/usr/ucb/echo" -else - print_cmd="echo" -fi - -if test -d /usr/xpg4/bin; then - PATH=/usr/xpg4/bin:\$PATH - export PATH -fi - -if test -d /usr/sfw/bin; then - PATH=\$PATH:/usr/sfw/bin - export PATH -fi - -unset CDPATH - -MS_Printf() -{ - \$print_cmd \$print_cmd_arg "\$1" -} - -MS_PrintLicense() -{ - PAGER=\${PAGER:=more} - if test x"\$licensetxt" != x; then - PAGER_PATH=\`exec <&- 2>&-; which \$PAGER || command -v \$PAGER || type \$PAGER\` - if test -x "\$PAGER_PATH"; then - echo "\$licensetxt" | \$PAGER - else - echo "\$licensetxt" - fi - if test x"\$accept" != xy; then - while true - do - MS_Printf "Please type y to accept, n otherwise: " - read yn - if test x"\$yn" = xn; then - keep=n - eval \$finish; exit 1 - break; - elif test x"\$yn" = xy; then - break; - fi - done - fi - fi -} - -MS_diskspace() -{ - ( - df -kP "\$1" | tail -1 | awk '{ if (\$4 ~ /%/) {print \$3} else {print \$4} }' - ) -} - -MS_dd() -{ - blocks=\`expr \$3 / 1024\` - bytes=\`expr \$3 % 1024\` - # Test for ibs, obs and conv feature - if dd if=/dev/zero of=/dev/null count=1 ibs=512 obs=512 conv=sync 2> /dev/null; then - dd if="\$1" ibs=\$2 skip=1 obs=1024 conv=sync 2> /dev/null | \\ - { test \$blocks -gt 0 && dd ibs=1024 obs=1024 count=\$blocks ; \\ - test \$bytes -gt 0 && dd ibs=1 obs=1024 count=\$bytes ; } 2> /dev/null - else - dd if="\$1" bs=\$2 skip=1 2> /dev/null - fi -} - -MS_dd_Progress() -{ - if test x"\$noprogress" = xy; then - MS_dd "\$@" - return \$? - fi - file="\$1" - offset=\$2 - length=\$3 - pos=0 - bsize=4194304 - while test \$bsize -gt \$length; do - bsize=\`expr \$bsize / 4\` - done - blocks=\`expr \$length / \$bsize\` - bytes=\`expr \$length % \$bsize\` - ( - dd ibs=\$offset skip=1 2>/dev/null - pos=\`expr \$pos \+ \$bsize\` - MS_Printf " 0%% " 1>&2 - if test \$blocks -gt 0; then - while test \$pos -le \$length; do - dd bs=\$bsize count=1 2>/dev/null - pcent=\`expr \$length / 100\` - pcent=\`expr \$pos / \$pcent\` - if test \$pcent -lt 100; then - MS_Printf "\b\b\b\b\b\b\b" 1>&2 - if test \$pcent -lt 10; then - MS_Printf " \$pcent%% " 1>&2 - else - MS_Printf " \$pcent%% " 1>&2 - fi - fi - pos=\`expr \$pos \+ \$bsize\` - done - fi - if test \$bytes -gt 0; then - dd bs=\$bytes count=1 2>/dev/null - fi - MS_Printf "\b\b\b\b\b\b\b" 1>&2 - MS_Printf " 100%% " 1>&2 - ) < "\$file" -} - -MS_Help() -{ - cat << EOH >&2 -Usage: \$0 [options] -Options: - --help | -h Print this message - --info Print embedded info : title, default target directory, embedded script ... - --list Print the list of files in the archive - --check Checks integrity and version dependency of the archive - --quiet Quiet install mode, skip human-computer interactions - --nox11 Do not spawn an xterm - --noexec Do not run embedded script - --extract= Extract directly to a target directory (absolute or relative) - Usually used with --noexec to just extract files without running - --tar arg1 [arg2 ...] Access the contents of the archive through the tar command -\${helpheader} -EOH -} - -MS_Verify_Sig() -{ - GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` - MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` - test -x "\$GPG_PATH" || GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` - test -x "\$MKTEMP_PATH" || MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` - offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` - temp_sig=\`mktemp -t XXXXX\` - echo \$SIGNATURE | base64 --decode > "\$temp_sig" - gpg_output=\`MS_dd "\$1" \$offset \$totalsize | LC_ALL=C "\$GPG_PATH" --verify "\$temp_sig" - 2>&1\` - gpg_res=\$? - rm -f "\$temp_sig" - if test \$gpg_res -eq 0 && test \`echo \$gpg_output | grep -c Good\` -eq 1; then - if test \`echo \$gpg_output | grep -c \$sig_key\` -eq 1; then - test x"\$quiet" = xn && echo "GPG signature is good" >&2 - else - echo "GPG Signature key does not match" >&2 - exit 2 - fi - else - test x"\$quiet" = xn && echo "GPG signature failed to verify" >&2 - exit 2 - fi -} - -MS_Check() -{ - OLD_PATH="\$PATH" - PATH=\${GUESS_MD5_PATH:-"\$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} - MD5_ARG="" - MD5_PATH=\`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum\` - test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which md5 || command -v md5 || type md5\` - test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which digest || command -v digest || type digest\` - PATH="\$OLD_PATH" - - SHA_PATH=\`exec <&- 2>&-; which shasum || command -v shasum || type shasum\` - test -x "\$SHA_PATH" || SHA_PATH=\`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum\` - - if test x"\$quiet" = xn; then - MS_Printf "Verifying archive integrity..." - fi - offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` - fsize=\`cat "\$1" | wc -c | tr -d " "\` - if test \$totalsize -ne \`expr \$fsize - \$offset\`; then - echo " Unexpected archive size." >&2 - exit 2 - fi - verb=\$2 - i=1 - for s in \$filesizes - do - crc=\`echo \$CRCsum | cut -d" " -f\$i\` - if test -x "\$SHA_PATH"; then - if test x"\`basename \$SHA_PATH\`" = xshasum; then - SHA_ARG="-a 256" - fi - sha=\`echo \$SHA | cut -d" " -f\$i\` - if test x"\$sha" = x0000000000000000000000000000000000000000000000000000000000000000; then - test x"\$verb" = xy && echo " \$1 does not contain an embedded SHA256 checksum." >&2 - else - shasum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$SHA_PATH \$SHA_ARG" | cut -b-64\`; - if test x"\$shasum" != x"\$sha"; then - echo "Error in SHA256 checksums: \$shasum is different from \$sha" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " SHA256 checksums are OK." >&2 - fi - crc="0000000000"; - fi - fi - if test -x "\$MD5_PATH"; then - if test x"\`basename \$MD5_PATH\`" = xdigest; then - MD5_ARG="-a md5" - fi - md5=\`echo \$MD5 | cut -d" " -f\$i\` - if test x"\$md5" = x00000000000000000000000000000000; then - test x"\$verb" = xy && echo " \$1 does not contain an embedded MD5 checksum." >&2 - else - md5sum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$MD5_PATH \$MD5_ARG" | cut -b-32\`; - if test x"\$md5sum" != x"\$md5"; then - echo "Error in MD5 checksums: \$md5sum is different from \$md5" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " MD5 checksums are OK." >&2 - fi - crc="0000000000"; verb=n - fi - fi - if test x"\$crc" = x0000000000; then - test x"\$verb" = xy && echo " \$1 does not contain a CRC checksum." >&2 - else - sum1=\`MS_dd_Progress "\$1" \$offset \$s | CMD_ENV=xpg4 cksum | awk '{print \$1}'\` - if test x"\$sum1" != x"\$crc"; then - echo "Error in checksums: \$sum1 is different from \$crc" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " CRC checksums are OK." >&2 - fi - fi - i=\`expr \$i + 1\` - offset=\`expr \$offset + \$s\` - done - if test x"\$quiet" = xn; then - echo " All good." - fi -} - -MS_Decompress() -{ - if test x"\$decrypt_cmd" != x""; then - { eval "\$decrypt_cmd" || echo " ... Decryption failed." >&2; } | eval "$GUNZIP_CMD" - else - eval "$GUNZIP_CMD" - fi - - if test \$? -ne 0; then - echo " ... Decompression failed." >&2 - fi -} - -UnTAR() -{ - if test x"\$quiet" = xn; then - tar \$1vf - $UNTAR_EXTRA 2>&1 || { echo " ... Extraction failed." >&2; kill -15 \$$; } - else - tar \$1f - $UNTAR_EXTRA 2>&1 || { echo Extraction failed. >&2; kill -15 \$$; } - fi -} - -MS_exec_cleanup() { - if test x"\$cleanup" = xy && test x"\$cleanup_script" != x""; then - cleanup=n - cd "\$tmpdir" - eval "\"\$cleanup_script\" \$scriptargs \$cleanupargs" - fi -} - -MS_cleanup() -{ - echo 'Signal caught, cleaning up' >&2 - MS_exec_cleanup - cd "\$TMPROOT" - rm -rf "\$tmpdir" - eval \$finish; exit 15 -} - -Script_Args_Check() -{ - script_supported_args=\$(echo \${helpheader} | grep -o -E "\-\-[^ ]+" | awk -F"=" {'print \$1'}) - arg_to_test=\$(echo \$1|awk -F"=" {'print \$1'}) - - for arg in \${script_supported_args}; - do - if test x"\$arg_to_test" = x"\$arg" ;then - return - fi - done - - MS_Help - exit 1 -} - -finish=true -xterm_loop= -noprogress=$NOPROGRESS -nox11=$NOX11 -copy=$COPY -ownership=$OWNERSHIP -verbose=n -cleanup=y -cleanupargs= -sig_key= - -initargs="\$@" - -while [ -n "\$*" ] -do - case "\$1" in - -h | --help) - MS_Help - exit 0 - ;; - -q | --quiet) - quiet=y - noprogress=y - shift - ;; - --info) - echo Identification: "\$label" - echo Target directory: "\$targetdir" - echo Uncompressed size: $USIZE KB - echo Compression: $COMPRESS - if test x"$ENCRYPT" != x""; then - echo Encryption: $ENCRYPT - fi - echo Date of packaging: $DATE - echo Built with Makeself version $MS_VERSION - echo Build command was: "$MS_COMMAND" - if test x"\$script" != x; then - echo Script run after extraction: - echo " " \$script \$scriptargs - fi - if test x"$copy" = xcopy; then - echo "Archive will copy itself to a temporary location" - fi - if test x"$NEED_ROOT" = xy; then - echo "Root permissions required for extraction" - fi - if test x"$KEEP" = xy; then - echo "directory \$targetdir is permanent" - else - echo "\$targetdir will be removed after extraction" - fi - exit 0 - ;; - --list) - echo Target directory: \$targetdir - offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - for s in \$filesizes - do - MS_dd "\$0" \$offset \$s | MS_Decompress | UnTAR t - offset=\`expr \$offset + \$s\` - done - exit 0 - ;; - --tar) - offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - arg1="\$2" - shift 2 || { MS_Help; exit 1; } - for s in \$filesizes - do - MS_dd "\$0" \$offset \$s | MS_Decompress | tar "\$arg1" - "\$@" - offset=\`expr \$offset + \$s\` - done - exit 0 - ;; - --check) - MS_Check "\$0" y - scriptargs="\$scriptargs \$1" - shift - ;; - --noexec) - script="" - cleanup_script="" - shift - ;; - --extract=*) - keep=y - targetdir=\`echo \$1 | cut -d"=" -f2 \` - if ! shift; then MS_Help; exit 1; fi - ;; - --nox11) - nox11=y - shift - ;; - --xwin) - if test "$NOWAIT" = n; then - finish="echo Press Return to close this window...; read junk" - fi - xterm_loop=1 - shift - ;; - --phase2) - copy=phase2 - shift - ;; - --repack | --repack-path=*) - Script_Args_Check \$1 - scriptargs="\$scriptargs '\$1'" - shift - if [[ ! "\$1" =~ ^-.* ]]; then - scriptargs="\$scriptargs '\$1'" - shift - fi - ;; - *) - Script_Args_Check \$1 - scriptargs="\$scriptargs '\$1'" - shift - ;; - esac -done - -quiet_para="" -if test x"\$quiet" = xy; then - quiet_para="--quiet " -fi -scriptargs="--\$name_of_file""--\"\$pwd_of_file\""" \$quiet_para""\$scriptargs" - -if test x"\$quiet" = xy -a x"\$verbose" = xy; then - echo Cannot be verbose and quiet at the same time. >&2 - exit 1 -fi - -if test x"$NEED_ROOT" = xy -a \`id -u\` -ne 0; then - echo "Administrative privileges required for this archive (use su or sudo)" >&2 - exit 1 -fi - -if test x"\$copy" \!= xphase2; then - MS_PrintLicense -fi - -case "\$copy" in -copy) - tmpdir="\$TMPROOT"/makeself.\$RANDOM.\`date +"%y%m%d%H%M%S"\`.\$\$ - mkdir "\$tmpdir" || { - echo "Could not create temporary directory \$tmpdir" >&2 - exit 1 - } - SCRIPT_COPY="\$tmpdir/makeself" - echo "Copying to a temporary location..." >&2 - cp "\$0" "\$SCRIPT_COPY" - chmod +x "\$SCRIPT_COPY" - cd "\$TMPROOT" - exec "\$SCRIPT_COPY" --phase2 -- \$initargs - ;; -phase2) - finish="\$finish ; rm -rf \`dirname \$0\`" - ;; -esac - -if test x"\$nox11" = xn; then - if tty -s; then # Do we have a terminal? - : - else - if test x"\$DISPLAY" != x -a x"\$xterm_loop" = x; then # No, but do we have X? - if xset q > /dev/null 2>&1; then # Check for valid DISPLAY variable - GUESS_XTERMS="xterm gnome-terminal rxvt dtterm eterm Eterm xfce4-terminal lxterminal kvt konsole aterm terminology" - for a in \$GUESS_XTERMS; do - if type \$a >/dev/null 2>&1; then - XTERM=\$a - break - fi - done - chmod a+x \$0 || echo Please add execution rights on \$0 - if test \`echo "\$0" | cut -c1\` = "/"; then # Spawn a terminal! - exec \$XTERM -e "\$0 --xwin \$initargs" - else - exec \$XTERM -e "./\$0 --xwin \$initargs" - fi - fi - fi - fi -fi - -if test x"\$targetdir" = x.; then - tmpdir="." -else - if test x"\$keep" = xy; then - if test x"\$nooverwrite" = xy && test -d "\$targetdir"; then - echo "Target directory \$targetdir already exists, aborting." >&2 - exit 1 - fi - if test x"\$quiet" = xn; then - echo "Creating directory \$targetdir" >&2 - fi - tmpdir="\$targetdir" - dashp="-p" - else - tmpdir="\$TMPROOT/selfgz\$\$\$RANDOM" - dashp="" - fi - mkdir \$dashp "\$tmpdir" || { - echo 'Cannot create target directory' \$tmpdir >&2 - echo 'You should try option --extract=' >&2 - eval \$finish - exit 1 - } -fi - -location="\`pwd\`" -if test x"\$SETUP_NOCHECK" != x1; then - MS_Check "\$0" -fi -offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - -if test x"\$verbose" = xy; then - MS_Printf "About to extract $USIZE KB in \$tmpdir ... Proceed ? [Y/n] " - read yn - if test x"\$yn" = xn; then - eval \$finish; exit 1 - fi -fi - -if test x"\$quiet" = xn; then - # Decrypting with openssl will ask for password, - # the prompt needs to start on new line - if test x"$ENCRYPT" = x"openssl"; then - echo "Decrypting and uncompressing \$label..." - else - MS_Printf "Uncompressing \$label" - fi -fi -res=3 -if test x"\$keep" = xn; then - trap MS_cleanup 1 2 3 15 -fi - -if test x"\$nodiskspace" = xn; then - leftspace=\`MS_diskspace "\$tmpdir"\` - if test -n "\$leftspace"; then - if test "\$leftspace" -lt $USIZE; then - echo - echo "Not enough space left in "\`dirname \$tmpdir\`" (\$leftspace KB) to decompress \$0 ($USIZE KB)" >&2 - if test x"\$keep" = xn; then - echo "Consider setting TMPDIR to a directory with more free space." - fi - eval \$finish; exit 1 - fi - fi -fi - -for s in \$filesizes -do - if MS_dd_Progress "\$0" \$offset \$s | MS_Decompress | ( cd "\$tmpdir"; umask \$ORIG_UMASK ; UnTAR xp ) 1>/dev/null; then - if test x"\$ownership" = xy; then - (cd "\$tmpdir"; chown -R \`id -u\` .; chgrp -R \`id -g\` .) - fi - else - echo >&2 - echo "Unable to decompress \$0" >&2 - eval \$finish; exit 1 - fi - offset=\`expr \$offset + \$s\` -done -if test x"\$quiet" = xn; then - echo -fi - -cd "\$tmpdir" -res=0 -if test x"\$script" != x; then - if test x"\$export_conf" = x"y"; then - MS_BUNDLE="\$0" - MS_LABEL="\$label" - MS_SCRIPT="\$script" - MS_SCRIPTARGS="\$scriptargs" - MS_ARCHDIRNAME="\$archdirname" - MS_KEEP="\$KEEP" - MS_NOOVERWRITE="\$NOOVERWRITE" - MS_COMPRESS="\$COMPRESS" - MS_CLEANUP="\$cleanup" - export MS_BUNDLE MS_LABEL MS_SCRIPT MS_SCRIPTARGS - export MS_ARCHDIRNAME MS_KEEP MS_NOOVERWRITE MS_COMPRESS - fi - - if test x"\$verbose" = x"y"; then - yn="x" - while test x"\$yn" != x -a x"\$yn" != xy -a x"\$yn" != xY -a x"\$yn" != xn -a x"\$yn" != xN - do - MS_Printf "OK to execute: \$script \$scriptargs \$* ? [Y/n] " - read yn - if test x"\$yn" = x -o x"\$yn" = xy -o x"\$yn" = xY; then - eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$?; - elif test x"\$yn" = xn -o x"\$yn" = xN; then - echo "Unable to decompress \$script ,because of aborting! ";res=\$? - else - echo "Input value is unacceptable,please try again." - fi - done - else - eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$? - fi - if test "\$res" -ne 0; then - test x"\$verbose" = xy && echo "The program '\$script' returned an error code (\$res)" >&2 - fi -fi - -MS_exec_cleanup - -if test x"\$keep" = xn; then - cd "\$TMPROOT" - rm -rf "\$tmpdir" -fi -eval \$finish; exit \$res -EOF diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.1 b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.1 deleted file mode 100755 index 81bf6e4ff..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.1 +++ /dev/null @@ -1,110 +0,0 @@ -.TH "MAKESELF" "1" "2.4.5" -.SH "NAME" -makeself \- An utility to generate self-extractable archives. -.SH "SYNTAX" -.B makeself [\fIoptions\fP] archive_dir file_name label -.B [\fIstartup_script\fP] [\fIargs\fP] -.SH "DESCRIPTION" -This program is a free (GPL) utility designed to create self-extractable -archives from a directory. -.SH "OPTIONS" -The following options are supported. -.TP 15 -.B -v, --version -Prints out the makeself version number and exits. -.TP -.B -h, --help -Print out help information. -.TP -.B --tar-quietly -Suppress verbose output from the tar command -.TP -.B --quiet -Do not print any messages other than errors -.TP -.B --gzip -Compress using gzip (default if detected). -.TP -.B --bzip2 -Compress using bzip2. -.TP -.B --pbzip2 -Compress using pbzip2. -.TP -.B --xz -Compress using xz. -.TP -.B --lzo -Compress using lzop. -.TP -.B --lz4 -Compress using lz4. -.TP -.B --compress -Compress using the UNIX 'compress' command. -.TP -.B --nocomp -Do not compress the data. -.TP -.B --complevel lvl -Specify the compression level for gzip,bzip2,pbzui2,xz,lzo or lz4 -.TP -.B --notemp -The archive will create archive_dir in the current directory and -uncompress in ./archive_dir. -.TP -.B --copy -Upon extraction, the archive will first copy itself to a temporary directory. -.TP -.B --append -Append more files to an existing makeself archive. The label and startup scripts will then be ignored. -.TP -.B --current -Files will be extracted to the current directory. Both --current and --target dir imply --notemp. -.TP -.B --target dir -Extract directly to a target directory. Directory path can be either absolute or relative. -.TP -.B --header file -Specify location of the header script. -.TP -.B --cleanup file -Specify a cleanup script that executes on interrupt and when finished successfully. -.TP -.B --follow -Follow the symlinks in the archive. -.TP -.B --noprogress -Do not show the progress during the decompression. -.TP -.B --nox11 -Disable automatic spawn of an xterm if running in X11. -.TP -.B --nowait -Do not wait for user input after executing embedded program from an xterm. -.TP -.B --nomd5 -Do not create a MD5 checksum for the archive. -.TP -.B --nocrc -Do not create a CRC32 checksum for the archive. -.TP -.B --lsm file -LSM file describing the package. -.B --packaging-date date -Use provided string as the packaging date instead of the current date. -.SH "EXAMPLES" -Here is an example, assuming the user has a package image stored in a /home/joe/mysoft, -and he wants to generate a self-extracting package named mysoft.sh, which will launch -the "setup" script initially stored in /home/joe/mysoft: -.TP -makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup -.TP -Here is also how I created the makeself.run archive which contains the Makeself distribution: -.TP -makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" -.SH "AUTHORS" -Makeself has been written by Stéphane Peter . -.BR -This man page was originally written by Bartosz Fenski for the -Debian GNU/Linux distribution (but it may be used by others). diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.lsm b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.lsm deleted file mode 100755 index 3c4cea8c1..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.lsm +++ /dev/null @@ -1,16 +0,0 @@ -Begin3 -Title: makeself.sh -Version: 2.4.5 -Description: makeself.sh is a shell script that generates a self-extractable - tar.gz archive from a directory. The resulting file appears as a shell - script, and can be launched as is. The archive will then uncompress - itself to a temporary directory and an arbitrary command will be - executed (for example an installation script). This is pretty similar - to archives generated with WinZip Self-Extractor in the Windows world. -Keywords: Installation archive tar winzip -Author: Stephane Peter (megastep@megastep.org) -Maintained-by: Stephane Peter (megastep@megastep.org) -Original-site: https://makeself.io/ -Platform: Unix -Copying-policy: GPL -End diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.sh deleted file mode 100755 index c8ea56597..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/makeself.sh +++ /dev/null @@ -1,822 +0,0 @@ -#!/bin/sh -# -# Makeself version 2.4.x -# by Stephane Peter -# -# Utility to create self-extracting tar.gz archives. -# The resulting archive is a file holding the tar.gz archive with -# a small Shell script stub that uncompresses the archive to a temporary -# directory and then executes a given script from withing that directory. -# -# Makeself home page: https://makeself.io/ -# -# Version 2.0 is a rewrite of version 1.0 to make the code easier to read and maintain. -# -# Version history : -# - 1.0 : Initial public release -# - 1.1 : The archive can be passed parameters that will be passed on to -# the embedded script, thanks to John C. Quillan -# - 1.2 : Package distribution, bzip2 compression, more command line options, -# support for non-temporary archives. Ideas thanks to Francois Petitjean -# - 1.3 : More patches from Bjarni R. Einarsson and Francois Petitjean: -# Support for no compression (--nocomp), script is no longer mandatory, -# automatic launch in an xterm, optional verbose output, and -target -# archive option to indicate where to extract the files. -# - 1.4 : Improved UNIX compatibility (Francois Petitjean) -# Automatic integrity checking, support of LSM files (Francois Petitjean) -# - 1.5 : Many bugfixes. Optionally disable xterm spawning. -# - 1.5.1 : More bugfixes, added archive options -list and -check. -# - 1.5.2 : Cosmetic changes to inform the user of what's going on with big -# archives (Quake III demo) -# - 1.5.3 : Check for validity of the DISPLAY variable before launching an xterm. -# More verbosity in xterms and check for embedded command's return value. -# Bugfix for Debian 2.0 systems that have a different "print" command. -# - 1.5.4 : Many bugfixes. Print out a message if the extraction failed. -# - 1.5.5 : More bugfixes. Added support for SETUP_NOCHECK environment variable to -# bypass checksum verification of archives. -# - 1.6.0 : Compute MD5 checksums with the md5sum command (patch from Ryan Gordon) -# - 2.0 : Brand new rewrite, cleaner architecture, separated header and UNIX ports. -# - 2.0.1 : Added --copy -# - 2.1.0 : Allow multiple tarballs to be stored in one archive, and incremental updates. -# Added --nochown for archives -# Stopped doing redundant checksums when not necesary -# - 2.1.1 : Work around insane behavior from certain Linux distros with no 'uncompress' command -# Cleaned up the code to handle error codes from compress. Simplified the extraction code. -# - 2.1.2 : Some bug fixes. Use head -n to avoid problems. -# - 2.1.3 : Bug fixes with command line when spawning terminals. -# Added --tar for archives, allowing to give arbitrary arguments to tar on the contents of the archive. -# Added --noexec to prevent execution of embedded scripts. -# Added --nomd5 and --nocrc to avoid creating checksums in archives. -# Added command used to create the archive in --info output. -# Run the embedded script through eval. -# - 2.1.4 : Fixed --info output. -# Generate random directory name when extracting files to . to avoid problems. (Jason Trent) -# Better handling of errors with wrong permissions for the directory containing the files. (Jason Trent) -# Avoid some race conditions (Ludwig Nussel) -# Unset the $CDPATH variable to avoid problems if it is set. (Debian) -# Better handling of dot files in the archive directory. -# - 2.1.5 : Made the md5sum detection consistent with the header code. -# Check for the presence of the archive directory -# Added --encrypt for symmetric encryption through gpg (Eric Windisch) -# Added support for the digest command on Solaris 10 for MD5 checksums -# Check for available disk space before extracting to the target directory (Andreas Schweitzer) -# Allow extraction to run asynchronously (patch by Peter Hatch) -# Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo) -# - 2.1.6 : Replaced one dot per file progress with a realtime progress percentage and a spining cursor (Guy Baconniere) -# Added --noprogress to prevent showing the progress during the decompression (Guy Baconniere) -# Added --target dir to allow extracting directly to a target directory (Guy Baconniere) -# - 2.2.0 : Many bugfixes, updates and contributions from users. Check out the project page on Github for the details. -# - 2.3.0 : Option to specify packaging date to enable byte-for-byte reproducibility. (Marc Pawlowsky) -# - 2.4.0 : Optional support for SHA256 checksums in archives. -# - 2.4.2 : Add support for threads for several compressors. (M. Limber) -# Added zstd support. -# - 2.4.3 : Make explicit POSIX tar archives for increased compatibility. -# - 2.4.5 : Added --tar-format to override ustar tar archive format -# -# (C) 1998-2021 by Stephane Peter -# -# This software is released under the terms of the GNU GPL version 2 and above -# Please read the license at http://www.gnu.org/copyleft/gpl.html -# Self-extracting archives created with this script are explictly NOT released under the term of the GPL -# - -MS_VERSION=2.4.5 -MS_COMMAND="$0" -unset CDPATH - -for f in ${1+"$@"}; do - MS_COMMAND="$MS_COMMAND \\\\ - \\\"$f\\\"" -done - -# For Solaris systems -if test -d /usr/xpg4/bin; then - PATH=/usr/xpg4/bin:$PATH - export PATH -fi - -# Procedures - -MS_Usage() -{ - echo "Usage: $0 [args] archive_dir file_name label startup_script [script_args]" - echo "args can be one or more of the following :" - echo " --version | -v : Print out Makeself version number and exit" - echo " --help | -h : Print out this help message" - echo " --tar-quietly : Suppress verbose output from the tar command" - echo " --quiet | -q : Do not print any messages other than errors." - echo " --gzip : Compress using gzip (default if detected)" - echo " --pigz : Compress with pigz" - echo " --zstd : Compress with zstd" - echo " --bzip2 : Compress using bzip2 instead of gzip" - echo " --pbzip2 : Compress using pbzip2 instead of gzip" - echo " --xz : Compress using xz instead of gzip" - echo " --lzo : Compress using lzop instead of gzip" - echo " --lz4 : Compress using lz4 instead of gzip" - echo " --compress : Compress using the UNIX 'compress' command" - echo " --complevel lvl : Compression level for gzip pigz zstd xz lzo lz4 bzip2 and pbzip2 (default 9)" - echo " --threads thds : Number of threads to be used by compressors that support parallelization." - echo " Omit to use compressor's default. Most useful (and required) for opting" - echo " into xz's threading, usually with '--threads=0' for all available cores." - echo " pbzip2 and pigz are parallel by default, and setting this value allows" - echo " limiting the number of threads they use." - echo " --base64 : Instead of compressing, encode the data using base64" - echo " --gpg-encrypt : Instead of compressing, encrypt the data using GPG" - echo " --gpg-asymmetric-encrypt-sign" - echo " : Instead of compressing, asymmetrically encrypt and sign the data using GPG" - echo " --gpg-extra opt : Append more options to the gpg command line" - echo " --ssl-encrypt : Instead of compressing, encrypt the data using OpenSSL" - echo " --ssl-passwd pass : Use the given password to encrypt the data using OpenSSL" - echo " --ssl-pass-src src : Use the given src as the source of password to encrypt the data" - echo " using OpenSSL. See \"PASS PHRASE ARGUMENTS\" in man openssl." - echo " If this option is not supplied, the user will be asked to enter" - echo " encryption password on the current terminal." - echo " --ssl-no-md : Do not use \"-md\" option not supported by older OpenSSL." - echo " --nochown : Do not give the target folder to the current user (default)" - echo " --chown : Give the target folder to the current user recursively" - echo " --nocomp : Do not compress the data" - echo " --notemp : The archive will create archive_dir in the" - echo " current directory and uncompress in ./archive_dir" - echo " --needroot : Check that the root user is extracting the archive before proceeding" - echo " --copy : Upon extraction, the archive will first copy itself to" - echo " a temporary directory" - echo " --append : Append more files to an existing Makeself archive" - echo " The label and startup scripts will then be ignored" - echo " --target dir : Extract directly to a target directory" - echo " directory path can be either absolute or relative" - echo " --nooverwrite : Do not extract the archive if the specified target directory exists" - echo " --current : Files will be extracted to the current directory" - echo " Both --current and --target imply --notemp" - echo " --tar-format opt : Specify a tar archive format (default is ustar)" - echo " --tar-extra opt : Append more options to the tar command line" - echo " --untar-extra opt : Append more options to the during the extraction of the tar archive" - echo " --nomd5 : Don't calculate an MD5 for archive" - echo " --nocrc : Don't calculate a CRC for archive" - echo " --sha256 : Compute a SHA256 checksum for the archive" - echo " --header file : Specify location of the header script" - echo " --cleanup file : Specify a cleanup script that executes on interrupt and when finished successfully." - echo " --follow : Follow the symlinks in the archive" - echo " --noprogress : Do not show the progress during the decompression" - echo " --nox11 : Disable automatic spawn of a xterm" - echo " --nowait : Do not wait for user input after executing embedded" - echo " program from an xterm" - echo " --sign passphrase : Signature private key to sign the package with" - echo " --lsm file : LSM file describing the package" - echo " --license file : Append a license file" - echo " --help-header file : Add a header to the archive's --help output" - echo " --packaging-date date" - echo " : Use provided string as the packaging date" - echo " instead of the current date." - echo - echo " --keep-umask : Keep the umask set to shell default, rather than overriding when executing self-extracting archive." - echo " --export-conf : Export configuration variables to startup_script" - echo - echo "Do not forget to give a fully qualified startup script name" - echo "(i.e. with a ./ prefix if inside the archive)." - exit 1 -} - -# Default settings -if type gzip >/dev/null 2>&1; then - COMPRESS=gzip -elif type compress >/dev/null 2>&1; then - COMPRESS=compress -else - echo "ERROR: missing commands: gzip, compress" >&2 - MS_Usage -fi -ENCRYPT=n -PASSWD="" -PASSWD_SRC="" -OPENSSL_NO_MD=n -COMPRESS_LEVEL=9 -DEFAULT_THREADS=123456 # Sentinel value -THREADS=$DEFAULT_THREADS -KEEP=n -CURRENT=n -NOX11=n -NOWAIT=n -APPEND=n -TAR_QUIETLY=n -KEEP_UMASK=n -QUIET=n -NOPROGRESS=n -COPY=none -NEED_ROOT=n -TAR_ARGS=rvf -TAR_FORMAT=ustar -TAR_EXTRA="" -GPG_EXTRA="" -DU_ARGS=-ks -HEADER=`dirname "$0"`/makeself-header.sh -SIGNATURE="" -TARGETDIR="" -NOOVERWRITE=n -DATE=`LC_ALL=C date` -EXPORT_CONF=n -SHA256=n -OWNERSHIP=n -SIGN=n -GPG_PASSPHRASE="" - -# LSM file stuff -LSM_CMD="echo No LSM. >> \"\$archname\"" - -while true -do - case "$1" in - --version | -v) - echo Makeself version $MS_VERSION - exit 0 - ;; - --pbzip2) - COMPRESS=pbzip2 - shift - ;; - --bzip2) - COMPRESS=bzip2 - shift - ;; - --gzip) - COMPRESS=gzip - shift - ;; - --pigz) - COMPRESS=pigz - shift - ;; - --zstd) - COMPRESS=zstd - shift - ;; - --xz) - COMPRESS=xz - shift - ;; - --lzo) - COMPRESS=lzo - shift - ;; - --lz4) - COMPRESS=lz4 - shift - ;; - --compress) - COMPRESS=compress - shift - ;; - --base64) - COMPRESS=base64 - shift - ;; - --gpg-encrypt) - COMPRESS=gpg - shift - ;; - --gpg-asymmetric-encrypt-sign) - COMPRESS=gpg-asymmetric - shift - ;; - --gpg-extra) - GPG_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-encrypt) - ENCRYPT=openssl - shift - ;; - --ssl-passwd) - PASSWD=$2 - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-pass-src) - PASSWD_SRC=$2 - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-no-md) - OPENSSL_NO_MD=y - shift - ;; - --nocomp) - COMPRESS=none - shift - ;; - --complevel) - COMPRESS_LEVEL="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --threads) - THREADS="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --nochown) - OWNERSHIP=n - shift - ;; - --chown) - OWNERSHIP=y - shift - ;; - --notemp) - KEEP=y - shift - ;; - --copy) - COPY=copy - shift - ;; - --current) - CURRENT=y - KEEP=y - shift - ;; - --tar-format) - TAR_FORMAT="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --tar-extra) - TAR_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --untar-extra) - UNTAR_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --target) - TARGETDIR="$2" - KEEP=y - shift 2 || { MS_Usage; exit 1; } - ;; - --sign) - SIGN=y - GPG_PASSPHRASE="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --nooverwrite) - NOOVERWRITE=y - shift - ;; - --needroot) - NEED_ROOT=y - shift - ;; - --header) - HEADER="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --cleanup) - CLEANUP_SCRIPT="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --license) - # We need to escape all characters having a special meaning in double quotes - LICENSE=$(sed 's/\\/\\\\/g; s/"/\\\"/g; s/`/\\\`/g; s/\$/\\\$/g' "$2") - shift 2 || { MS_Usage; exit 1; } - ;; - --follow) - TAR_ARGS=rvhf - DU_ARGS=-ksL - shift - ;; - --noprogress) - NOPROGRESS=y - shift - ;; - --nox11) - NOX11=y - shift - ;; - --nowait) - NOWAIT=y - shift - ;; - --nomd5) - NOMD5=y - shift - ;; - --sha256) - SHA256=y - shift - ;; - --nocrc) - NOCRC=y - shift - ;; - --append) - APPEND=y - shift - ;; - --lsm) - LSM_CMD="cat \"$2\" >> \"\$archname\"" - shift 2 || { MS_Usage; exit 1; } - ;; - --packaging-date) - DATE="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --help-header) - HELPHEADER=`sed -e "s/'/'\\\\\''/g" $2` - shift 2 || { MS_Usage; exit 1; } - [ -n "$HELPHEADER" ] && HELPHEADER="$HELPHEADER -" - ;; - --tar-quietly) - TAR_QUIETLY=y - shift - ;; - --keep-umask) - KEEP_UMASK=y - shift - ;; - --export-conf) - EXPORT_CONF=y - shift - ;; - -q | --quiet) - QUIET=y - shift - ;; - -h | --help) - MS_Usage - ;; - -*) - echo Unrecognized flag : "$1" - MS_Usage - ;; - *) - break - ;; - esac -done - -if test $# -lt 1; then - MS_Usage -else - if test -d "$1"; then - archdir="$1" - else - echo "Directory $1 does not exist." >&2 - exit 1 - fi -fi -archname="$2" - -if test "$QUIET" = "y" || test "$TAR_QUIETLY" = "y"; then - if test "$TAR_ARGS" = "rvf"; then - TAR_ARGS="rf" - elif test "$TAR_ARGS" = "rvhf"; then - TAR_ARGS="rhf" - fi -fi - -if test "$APPEND" = y; then - if test $# -lt 2; then - MS_Usage - fi - - # Gather the info from the original archive - OLDENV=`sh "$archname" --dumpconf` - if test $? -ne 0; then - echo "Unable to update archive: $archname" >&2 - exit 1 - else - eval "$OLDENV" - OLDSKIP=`expr $SKIP + 1` - fi -else - if test "$KEEP" = n -a $# = 3; then - echo "ERROR: Making a temporary archive with no embedded command does not make sense!" >&2 - echo >&2 - MS_Usage - fi - # We don't want to create an absolute directory unless a target directory is defined - if test "$CURRENT" = y; then - archdirname="." - elif test x"$TARGETDIR" != x; then - archdirname="$TARGETDIR" - else - archdirname=`basename "$1"` - fi - - if test $# -lt 3; then - MS_Usage - fi - - LABEL="$3" - SCRIPT="$4" - test "x$SCRIPT" = x || shift 1 - shift 3 - SCRIPTARGS="$*" -fi - -if test "$KEEP" = n -a "$CURRENT" = y; then - echo "ERROR: It is A VERY DANGEROUS IDEA to try to combine --notemp and --current." >&2 - exit 1 -fi - -case $COMPRESS in -gzip) - GZIP_CMD="gzip -c$COMPRESS_LEVEL" - GUNZIP_CMD="gzip -cd" - ;; -pigz) - GZIP_CMD="pigz -$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD --processes $THREADS" - fi - GUNZIP_CMD="gzip -cd" - ;; -zstd) - GZIP_CMD="zstd -$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD --threads=$THREADS" - fi - GUNZIP_CMD="zstd -cd" - ;; -pbzip2) - GZIP_CMD="pbzip2 -c$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD -p$THREADS" - fi - GUNZIP_CMD="bzip2 -d" - ;; -bzip2) - GZIP_CMD="bzip2 -$COMPRESS_LEVEL" - GUNZIP_CMD="bzip2 -d" - ;; -xz) - GZIP_CMD="xz -c$COMPRESS_LEVEL" - # Must opt-in by specifying a value since not all versions of xz support threads - if test $THREADS -ne $DEFAULT_THREADS; then - GZIP_CMD="$GZIP_CMD --threads=$THREADS" - fi - GUNZIP_CMD="xz -d" - ;; -lzo) - GZIP_CMD="lzop -c$COMPRESS_LEVEL" - GUNZIP_CMD="lzop -d" - ;; -lz4) - GZIP_CMD="lz4 -c$COMPRESS_LEVEL" - GUNZIP_CMD="lz4 -d" - ;; -base64) - GZIP_CMD="base64" - GUNZIP_CMD="base64 --decode -i -" - ;; -gpg) - GZIP_CMD="gpg $GPG_EXTRA -ac -z$COMPRESS_LEVEL" - GUNZIP_CMD="gpg -d" - ENCRYPT="gpg" - ;; -gpg-asymmetric) - GZIP_CMD="gpg $GPG_EXTRA -z$COMPRESS_LEVEL -es" - GUNZIP_CMD="gpg --yes -d" - ENCRYPT="gpg" - ;; -compress) - GZIP_CMD="compress -fc" - GUNZIP_CMD="(type compress >/dev/null 2>&1 && compress -fcd || gzip -cd)" - ;; -none) - GZIP_CMD="cat" - GUNZIP_CMD="cat" - ;; -esac - -if test x"$ENCRYPT" = x"openssl"; then - if test x"$APPEND" = x"y"; then - echo "Appending to existing archive is not compatible with OpenSSL encryption." >&2 - fi - - ENCRYPT_CMD="openssl enc -aes-256-cbc -salt" - DECRYPT_CMD="openssl enc -aes-256-cbc -d" - - if test x"$OPENSSL_NO_MD" != x"y"; then - ENCRYPT_CMD="$ENCRYPT_CMD -md sha256" - DECRYPT_CMD="$DECRYPT_CMD -md sha256" - fi - - if test -n "$PASSWD_SRC"; then - ENCRYPT_CMD="$ENCRYPT_CMD -pass $PASSWD_SRC" - elif test -n "$PASSWD"; then - ENCRYPT_CMD="$ENCRYPT_CMD -pass pass:$PASSWD" - fi -fi - -tmpfile="${TMPDIR:-/tmp}/mkself$$" - -if test -f "$HEADER"; then - oldarchname="$archname" - archname="$tmpfile" - # Generate a fake header to count its lines - SKIP=0 - . "$HEADER" - SKIP=`cat "$tmpfile" |wc -l` - # Get rid of any spaces - SKIP=`expr $SKIP` - rm -f "$tmpfile" - if test "$QUIET" = "n"; then - echo "Header is $SKIP lines long" >&2 - fi - archname="$oldarchname" -else - echo "Unable to open header file: $HEADER" >&2 - exit 1 -fi - -if test "$QUIET" = "n"; then - echo -fi - -if test "$APPEND" = n; then - if test -f "$archname"; then - echo "WARNING: Overwriting existing file: $archname" >&2 - fi -fi - -USIZE=`du $DU_ARGS "$archdir" | awk '{print $1}'` - -if test "." = "$archdirname"; then - if test "$KEEP" = n; then - archdirname="makeself-$$-`date +%Y%m%d%H%M%S`" - fi -fi - -test -d "$archdir" || { echo "Error: $archdir does not exist."; rm -f "$tmpfile"; exit 1; } -if test "$QUIET" = "n"; then - echo "About to compress $USIZE KB of data..." - echo "Adding files to archive named \"$archname\"..." -fi - -# See if we have GNU tar -TAR=`exec <&- 2>&-; which gtar || command -v gtar || type gtar` -test -x "$TAR" || TAR=tar - -tmparch="${TMPDIR:-/tmp}/mkself$$.tar" -( - if test "$APPEND" = "y"; then - tail -n "+$OLDSKIP" "$archname" | eval "$GUNZIP_CMD" > "$tmparch" - fi - cd "$archdir" - # "Determining if a directory is empty" - # https://www.etalabs.net/sh_tricks.html - find . \ - \( \ - ! -type d \ - -o \ - \( -links 2 -exec sh -c ' - is_empty () ( - cd "$1" - set -- .[!.]* ; test -f "$1" && return 1 - set -- ..?* ; test -f "$1" && return 1 - set -- * ; test -f "$1" && return 1 - return 0 - ) - is_empty "$0"' {} \; \ - \) \ - \) -print \ - | LC_ALL=C sort \ - | sed 's/./\\&/g' \ - | xargs $TAR $TAR_EXTRA --format $TAR_FORMAT -$TAR_ARGS "$tmparch" -) || { - echo "ERROR: failed to create temporary archive: $tmparch" - rm -f "$tmparch" "$tmpfile" - exit 1 -} - -USIZE=`du $DU_ARGS "$tmparch" | awk '{print $1}'` - -eval "$GZIP_CMD" <"$tmparch" >"$tmpfile" || { - echo "ERROR: failed to create temporary file: $tmpfile" - rm -f "$tmparch" "$tmpfile" - exit 1 -} -rm -f "$tmparch" - -if test x"$ENCRYPT" = x"openssl"; then - echo "About to encrypt archive \"$archname\"..." - { eval "$ENCRYPT_CMD -in $tmpfile -out ${tmpfile}.enc" && mv -f ${tmpfile}.enc $tmpfile; } || \ - { echo Aborting: could not encrypt temporary file: "$tmpfile".; rm -f "$tmpfile"; exit 1; } -fi - -fsize=`cat "$tmpfile" | wc -c | tr -d " "` - -# Compute the checksums - -shasum=0000000000000000000000000000000000000000000000000000000000000000 -md5sum=00000000000000000000000000000000 -crcsum=0000000000 - -if test "$NOCRC" = y; then - if test "$QUIET" = "n"; then - echo "skipping crc at user request" - fi -else - crcsum=`CMD_ENV=xpg4 cksum < "$tmpfile" | sed -e 's/ /Z/' -e 's/ /Z/' | cut -dZ -f1` - if test "$QUIET" = "n"; then - echo "CRC: $crcsum" - fi -fi - -if test "$SHA256" = y; then - SHA_PATH=`exec <&- 2>&-; which shasum || command -v shasum || type shasum` - if test -x "$SHA_PATH"; then - shasum=`eval "$SHA_PATH -a 256" < "$tmpfile" | cut -b-64` - else - SHA_PATH=`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum` - shasum=`eval "$SHA_PATH" < "$tmpfile" | cut -b-64` - fi - if test "$QUIET" = "n"; then - if test -x "$SHA_PATH"; then - echo "SHA256: $shasum" - else - echo "SHA256: none, SHA command not found" - fi - fi -fi -if test "$NOMD5" = y; then - if test "$QUIET" = "n"; then - echo "Skipping md5sum at user request" - fi -else - # Try to locate a MD5 binary - OLD_PATH=$PATH - PATH=${GUESS_MD5_PATH:-"$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} - MD5_ARG="" - MD5_PATH=`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum` - test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which md5 || command -v md5 || type md5` - test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which digest || command -v digest || type digest` - PATH=$OLD_PATH - if test -x "$MD5_PATH"; then - if test `basename ${MD5_PATH}`x = digestx; then - MD5_ARG="-a md5" - fi - md5sum=`eval "$MD5_PATH $MD5_ARG" < "$tmpfile" | cut -b-32` - if test "$QUIET" = "n"; then - echo "MD5: $md5sum" - fi - else - if test "$QUIET" = "n"; then - echo "MD5: none, MD5 command not found" - fi - fi -fi -if test "$SIGN" = y; then - GPG_PATH=`exec <&- 2>&-; which gpg || command -v gpg || type gpg` - if test -x "$GPG_PATH"; then - SIGNATURE=`$GPG_PATH --pinentry-mode=loopback --batch --yes --passphrase "$GPG_PASSPHRASE" --output - --detach-sig $tmpfile | base64 | tr -d \\\\n` - if test "$QUIET" = "n"; then - echo "Signature: $SIGNATURE" - fi - else - echo "Missing gpg command" >&2 - fi -fi - -totalsize=0 -for size in $fsize; -do - totalsize=`expr $totalsize + $size` -done - -if test "$APPEND" = y; then - mv "$archname" "$archname".bak || exit - - # Prepare entry for new archive - filesizes="$fsize" - CRCsum="$crcsum" - MD5sum="$md5sum" - SHAsum="$shasum" - Signature="$SIGNATURE" - # Generate the header - . "$HEADER" - # Append the new data - cat "$tmpfile" >> "$archname" - - chmod +x "$archname" - rm -f "$archname".bak - if test "$QUIET" = "n"; then - echo "Self-extractable archive \"$archname\" successfully updated." - fi -else - filesizes="$fsize" - CRCsum="$crcsum" - MD5sum="$md5sum" - SHAsum="$shasum" - Signature="$SIGNATURE" - - # Generate the header - . "$HEADER" - - # Append the compressed tar data after the stub - if test "$QUIET" = "n"; then - echo - fi - cat "$tmpfile" >> "$archname" - chmod +x "$archname" - if test "$QUIET" = "n"; then - echo Self-extractable archive \"$archname\" successfully created. - fi -fi -rm -f "$tmpfile" diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/run-tests.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/run-tests.sh deleted file mode 100755 index 31ee16511..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/makeself/run-tests.sh +++ /dev/null @@ -1,8 +0,0 @@ -#!/bin/sh -# Run every available test - Bash needed -cd test -for test in *test; -do - echo "Running test $test ..." - bash $test || { echo "*** ERROR: Test '$test' failed!"; exit 1; } -done diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/merge_aicpu_info_json.sh b/atb_operator/AtbEyeSample/EyeOp/cmake/util/merge_aicpu_info_json.sh deleted file mode 100755 index a977bd51d..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/merge_aicpu_info_json.sh +++ /dev/null @@ -1,31 +0,0 @@ -#!/bin/bash - -project_path=$1 -build_path=$2 -vendor_name=customize -echo $@ -if [[ ! -d "$project_path" ]]; then - echo "[ERROR] No projcet path is provided" - exit 1 -fi - -if [[ ! -d "$build_path" ]]; then - echo "[ERROR] No build path is provided" - exit 1 -fi - -if [[ ! -d "$ASCEND_OPP_PATH" ]]; then - echo "[ERROR] No opp install path is provided" - exit 1 -fi -custom_exist_info_json=$ASCEND_OPP_PATH/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json -custom_new_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json -temp_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/temp_cust_aicpu_kernel.json - -if [[ -f "$custom_exist_info_json" ]] && [[ -f "$custom_new_info_json" ]]; then - cp -f $custom_exist_info_json $temp_info_json - chmod +w $temp_info_json - python3 ${project_path}/cmake/util/insert_op_info.py ${custom_new_info_json} ${temp_info_json} - cp -f $temp_info_json $custom_new_info_json - rm -f $temp_info_json -fi diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/opdesc_parser.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/opdesc_parser.py deleted file mode 100755 index 0d3221602..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/opdesc_parser.py +++ /dev/null @@ -1,298 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os - - -OP_ALL = '__ALLOP__' -SOC_ALL = '__ALLSOC__' -SOC_TO_SHORT_SOC_MAP = { - "ascend910a": "ascend910", - "ascend910proa": "ascend910", - "ascend910b": "ascend910", - "ascend910prob": "ascend910", - "ascend910premiuma": "ascend910", - "ascend910b1": "ascend910b", - "ascend910b2": "ascend910b", - "ascend910b2c": "ascend910b", - "ascend910b3": "ascend910b", - "ascend910b4": "ascend910b", - "ascend910c1": "ascend910c", - "ascend910c2": "ascend910c", - "ascend910c3": "ascend910c", - "ascend910c4": "ascend910c", - "ascend310p1": "ascend310p", - "ascend310p3": "ascend310p", - "ascend310p3vir01": "ascend310p", - "ascend310p3vir02": "ascend310p", - "ascend310p3vir04": "ascend310p", - "ascend310p3vir08": "ascend310p", - "ascend310b1": "ascend310b", - "bs9sx1aa": "bs9sx1a" -} -CONFLICT_KEYWORDS = { - "and", "as", "assert", "break", "class", "continue", "def", "del", "elif", "else", - "except", "finally", "for", "from", "global", "if", "import", "in", "is", "lambda", - "not", "or", "pass", "raise", "return", "try", "while", "with", "yield", "False", - "None", "True", "nonlocal", "arg", "__inputs__", "__outputs__", "options", "bisheng", - "bisheng_path", "tikcpp_path", "impl_mode", "custom_compile_options", - "custom_all_compile_options", "soc_version", "soc_short", "custom_compile_options_soc", - "custom_all_compile_options_soc", "origin_func_name", "ascendc_src_dir_ex", - "ascendc_src_dir", "ascendc_src_file", "src", "op_type", "code_channel", "op_info", - "compile_op", "get_code_channel", "result", "__attrs__", "isinstance", "attr", - "get_current_build_config", "_build_args", "get_dtype_fmt_options", "shutil", "os", - "get_kernel_source" -} - - -class OpDesc: - def __init__(self: any, op_type: str): - self.op_type = op_type - self.attr_list = [] - self.attr_val = {} - self.input_name = [] - self.input_ori_name = [] - self.input_type = [] - self.input_dtype = [] - self.input_fmt = [] - self.output_name = [] - self.output_ori_name = [] - self.output_type = [] - self.output_dtype = [] - self.output_fmt = [] - self.op_fmt_sel = False - self.op_chk_support = False - self.op_intf = '' - self.kern_name = '' - self.op_file = '' - self.op_replay_flag = False - self.op_replay_batch = False - self.input_idx = -1 - self.output_idx = -1 - self.max_block_dim = 32 - self.max_shape_size = 268435456 - self.dynamic_shape = False - self.op_range_limit = '' - self.custom_compile_options = {} - self.custom_all_compile_options = {} - - @staticmethod - def _parse_digit(conf: str) -> int: - return int(conf.split('=')[1]) - - @staticmethod - def _parse_flag(conf: str) -> bool: - if 'true' == conf.split('=')[1]: - return True - return False - - @staticmethod - def _parse_str(conf: str) -> str: - return conf.split('=')[1] - - @staticmethod - def _parse_list(conf: str) -> list: - return conf.split('=')[1].split(',') - - def parse_input(self: any, conf: str): - if conf.startswith('input{}.name'.format(int(self.input_idx) + 1)): - self.input_idx += 1 - self.input_ori_name.append(self._parse_str(conf)) - self.input_name.append(self.input_ori_name[-1] + '_in__') - elif conf.startswith('input{}.paramType'.format(int(self.input_idx))): - self.input_type.append(self._parse_str(conf)) - elif conf.startswith('input{}.dtype'.format(int(self.input_idx))): - self.input_dtype.append(self._parse_str(conf)) - elif conf.startswith('input{}.format'.format(int(self.input_idx))): - self.input_fmt.append(self._parse_str(conf)) - else: - return - - def parse_output(self: any, conf: str): - if conf.startswith('output{}.name'.format(int(self.output_idx) + 1)): - self.output_idx += 1 - self.output_ori_name.append(self._parse_str(conf)) - self.output_name.append(self.output_ori_name[-1] + '_out_') - elif conf.startswith('output{}.paramType'.format(int(self.output_idx))): - self.output_type.append(self._parse_str(conf)) - elif conf.startswith('output{}.dtype'.format(int(self.output_idx))): - self.output_dtype.append(self._parse_str(conf)) - elif conf.startswith('output{}.format'.format(int(self.output_idx))): - self.output_fmt.append(self._parse_str(conf)) - else: - return - - def parse_op_format(self: any, conf: str): - self.op_fmt_sel = self._parse_flag(conf) - - def parse_check_support(self: any, conf: str): - self.op_chk_support = self._parse_flag(conf) - - def parse_range_limit(self: any, conf: str): - self.op_range_limit = self._parse_str(conf) - - def parse_kern_name(self: any, conf: str): - self.kern_name = self._parse_str(conf) - - def parse_op_intf(self: any, conf: str): - self.op_intf = self._parse_str(conf) - - def parse_op_file(self: any, conf: str): - self.op_file = self._parse_str(conf) - - def parse_dynamic_shape(self: any, conf: str): - self.dynamic_shape = self._parse_flag(conf) - - def parse_attr_list(self: any, conf: str): - self.attr_list = self._parse_list(conf) - intersection_element = set(self.attr_list) & CONFLICT_KEYWORDS - if intersection_element: - raise Exception(f'The attribute name: {intersection_element} in op: {{\'{self.op_type}\'}} conflicts with\ - the built-in variable name. Use a complex name or prefix the operator name.') - - @staticmethod - def _camel_to_snake(camel_case_str: str): - snake_case_str = '' - for i, c in enumerate(camel_case_str): - if i == 0: - snake_case_str += c.lower() - elif c.isupper(): - snake_case_str += '_' + c.lower() - else: - snake_case_str += c - return snake_case_str - - def parse_attr_val(self: any, conf: str): - for attr in self.attr_list: - if self.attr_val.get(attr) is None: - self.attr_val[attr] = {} - if conf.startswith('attr_{}.type'.format(attr)): - self.attr_val.get(attr)['type'] = self._camel_to_snake(self._parse_str(conf)) - elif conf.startswith('attr_{}.paramType'.format(attr)): - self.attr_val.get(attr)['paramType'] = self._parse_str(conf) - elif conf.startswith('attr_{}.defaultValue'.format(attr)): - self.attr_val.get(attr)['defaultValue'] = self._parse_str(conf) - - def parse_replay_val(self: any, batch_list: list, iterator_list: list): - if self.op_type in batch_list: - self.op_replay_flag = True - self.op_replay_batch = True - elif self.op_type in iterator_list: - self.op_replay_flag = True - self.op_replay_batch = False - - -def _is_op_type_in_opdesc(op_descs: list, op_type: str): - for op in op_descs: - if op_type == op.op_type: - return True - return False - - -def _set_all_options_to_opdescs(op_descs, soc_ver_compile_options): - for op in op_descs: - op.custom_all_compile_options = soc_ver_compile_options - - -def _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options): - for op in op_descs: - if op.op_type != op_type: - continue - op.custom_compile_options = soc_ver_compile_options - - -def _trans_soc_ver_to_short(soc_ver: str): - low_soc_ver = soc_ver.lower() - if low_soc_ver not in SOC_TO_SHORT_SOC_MAP: - print(f'WARNING: caution: {soc_ver} will trans into ascend910, if not your intention,' - f'use ascend910b1~4 instead') - return SOC_TO_SHORT_SOC_MAP[low_soc_ver] - - -def _get_op_custom_options(op_descs: list, auto_gen_dir: str): - if auto_gen_dir is None: - return - file = os.path.join(auto_gen_dir, "custom_compile_options.ini") - if not os.path.exists(file): - print(f'WARNING: cannot find {auto_gen_dir}/custom_compile_options.ini') - return - with open(file, 'r') as fd: - lines = fd.readlines() - for line in lines: - param_list = str.split(line.rstrip('\n'), ',') - if len(param_list) != 3: - raise Exception(f'ERROR: custom compile option {param_list} len is not 3') - op_type = param_list[0] - if op_type.upper() == 'ALL': - op_type = OP_ALL - if op_type != OP_ALL and _is_op_type_in_opdesc(op_descs, op_type) == False: - print(f'WARNING: op: {op_type} are not exists in this project') - continue - soc_ver_compile_options = {} - soc_ver = param_list[1] - options_str = param_list[2] - options = str.split(options_str, ';') - if soc_ver == '': - soc_ver_compile_options[SOC_ALL] = options - else: - soc_ver_list = str.split(soc_ver, ';') - for ver in soc_ver_list: - short_ver = _trans_soc_ver_to_short(ver) - soc_ver_compile_options[short_ver] = options - if op_type == OP_ALL: - _set_all_options_to_opdescs(op_descs, soc_ver_compile_options) - else: - _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options) - - -def get_op_desc(file: str, batch_list: list, iterator_list: list, builder: any, - op_type: list, auto_gen_dir: str = None) -> list: - op_descs = [] - op_match = False - with open (file, 'r') as fd: - lines = fd.readlines() - for line in lines: - line = line.strip() - if line.startswith('['): - name = line[1:-1] - if op_type is None or name in op_type: - op_match = True - op_desc = builder(name) - op_desc.parse_replay_val(batch_list, iterator_list) - op_descs.append(op_desc) - else: - op_match = False - if op_type is not None and len(op_descs) == len(op_type): - return op_descs - continue - if not op_match: - continue - if line.startswith('input'): - op_desc.parse_input(line) - elif line.startswith('output'): - op_desc.parse_output(line) - elif line.startswith('dynamicFormat.flag'): - op_desc.parse_op_format(line) - elif line.startswith('needCheckSupport.flag'): - op_desc.parse_check_support(line) - elif line.startswith('rangeLimit.value'): - op_desc.parse_range_limit(line) - elif line.startswith('opInterface.value'): - op_desc.parse_op_intf(line) - elif line.startswith('kernel.name'): - op_desc.parse_kern_name(line) - elif line.startswith('opFile.value'): - op_desc.parse_op_file(line) - elif line.startswith('dynamicShapeSupport.flag'): - op_desc.parse_dynamic_shape(line) - elif line.startswith('attr.list'): - op_desc.parse_attr_list(line) - elif line.startswith('attr_'): - op_desc.parse_attr_val(line) - _get_op_custom_options(op_descs, auto_gen_dir) - return op_descs diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/parse_ini_to_json.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/parse_ini_to_json.py deleted file mode 100755 index ac931e5d3..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/parse_ini_to_json.py +++ /dev/null @@ -1,338 +0,0 @@ -# Copyright 2020-2021 Huawei Technologies Co., Ltd -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -""" -parser ini to json -""" - -import json -import os -import stat -import sys - - -ATTR_TYPE_LIST = ["int", "float", "bool", "str", "listInt", "listFloat", "listBool", "listStr", "listListInt", - "type", "listType", "tensor", "listTensor"] -ATTR_PARAMTYPE_LIST = ["optional", "required"] -BOOL_FLAG_KEY = ["dynamicFormat", "dynamicShapeSupport", "dynamicRankSupport", "precision_reduce", "heavyOp", - "needCheckSupport", "enableVectorCore"] -BOOL_LIST = ["true", "false"] -DTYPE_LIST = ["float16", "float", "float32", "int8", "int16", "int32", "uint8", "uint16", "uint32", "bool", - "int64", "uint64", "qint8", "qint16", "qint32", "quint8", "quint16", "double", "complex64", - "complex128", "string", "resource", "dual", "dual_sub_int8", "dual_sub_uint8", "string_ref", - "int4", "bfloat16", "uint1"] -FORMAT_LIST = ["NCHW", "NHWC", "ND", "NC1HWC0", "FRACTAL_Z", "NC1C0HWPAD", "NHWC1C0", "FSR_NCHW", "FRACTAL_DECONV", - "C1HWNC0", "FRACTAL_DECONV_TRANSPOSE", "FRACTAL_DECONV_SP_STRIDE_TRANS", "NC1HWC0_C04", - "FRACTAL_Z_C04", "CHWN", "FRACTAL_DECONV_SP_STRIDE8_TRANS", "HWCN", "NC1KHKWHWC0", "BN_WEIGHT", - "FILTER_HWCK", "HASHTABLE_LOOKUP_LOOKUPS", "HASHTABLE_LOOKUP_KEYS", "HASHTABLE_LOOKUP_VALUE", - "HASHTABLE_LOOKUP_OUTPUT", "HASHTABLE_LOOKUP_HITS", "C1HWNCoC0", "MD", "NDHWC", "FRACTAL_ZZ", - "FRACTAL_NZ", "NCDHW", "DHWCN", "NDC1HWC0", "FRACTAL_Z_3D", "CN", "NC", "DHWNC", - "FRACTAL_Z_3D_TRANSPOSE", "FRACTAL_ZN_LSTM", "FRACTAL_ZN_RNN", "FRACTAL_Z_G", "NULL"] - - -def parse_ini_files(ini_files): - """ - parse ini files to json - Parameters: - ---------------- - ini_files:input file list - return:ops_info - ---------------- - """ - tbe_ops_info = {} - for ini_file in ini_files: - check_file_size(ini_file) - parse_ini_to_obj(ini_file, tbe_ops_info) - return tbe_ops_info - - -def check_file_size(input_file): - try: - file_size = os.path.getsize(input_file) - except OSError as os_error: - print('[ERROR] Failed to open "%s". %s' % (input_file, str(os_error))) - raise OSError from os_error - if file_size > 10*1024*1024: - print('[WARN] The size of %s exceeds 10MB, it may take more time to run, please wait.' % input_file) - - -def parse_ini_to_obj(ini_file, tbe_ops_info): - """ - parse ini file to json obj - Parameters: - ---------------- - ini_file:ini file path - tbe_ops_info:ops_info - ---------------- - """ - with open(ini_file) as ini_file: - lines = ini_file.readlines() - op_dict = {} - op_name = "" - find_op_type = False - for line in lines: - line = line.rstrip() - if line == "": - continue - if line.startswith("["): - if line.endswith("]"): - op_name = line[1:-1] - op_dict = {} - tbe_ops_info[op_name] = op_dict - find_op_type = True - elif "=" in line: - key1 = line[:line.index("=")] - key2 = line[line.index("=")+1:] - key1_0, key1_1 = key1.split(".") - if key1_0 not in op_dict: - op_dict[key1_0] = {} - if key1_1 in op_dict.get(key1_0): - raise RuntimeError("Op:" + op_name + " " + key1_0 + " " + - key1_1 + " is repeated!") - dic_key = op_dict.get(key1_0) - dic_key[key1_1] = key2 - else: - continue - if not find_op_type: - raise RuntimeError("Not find OpType in .ini file.") - - -def check_output_exist(op_dict, is_valid): - """ - Function Description: - Check output is exist - Parameter: op_dict - Parameter: is_valid - """ - if "output0" in op_dict: - output0_dict = op_dict.get("output0") - if output0_dict.get("name", None) is None: - is_valid = False - print("output0.name is required in .ini file!") - else: - is_valid = False - print("output0 is required in .ini file!") - return is_valid - - -def check_attr_dict(attr_dict, is_valid, attr): - """ - Function Description: - Check attr_dict - Parameter: attr_dict - Parameter: is_valid - Parameter: attr - """ - attr_type = attr_dict.get("type") - value = attr_dict.get("value") - param_type = attr_dict.get("paramType") - if attr_type is None or value is None: - is_valid = False - print("If attr.list is exist, {0}.type and {0}.value is required".format(attr)) - if param_type and param_type not in ATTR_PARAMTYPE_LIST: - is_valid = False - print("{0}.paramType only support {1}.".format(attr, ATTR_PARAMTYPE_LIST)) - if attr_type and attr_type not in ATTR_TYPE_LIST: - is_valid = False - print("{0}.type only support {1}.".format(attr, ATTR_TYPE_LIST)) - return is_valid - - -def check_attr(op_dict, is_valid): - """ - Function Description: - Check attr - Parameter: op_dict - Parameter: is_valid - """ - if "attr" in op_dict: - attr_dict = op_dict.get("attr") - attr_list_str = attr_dict.get("list", None) - if attr_list_str is None: - is_valid = False - print("attr.list is required in .ini file!") - else: - attr_list = attr_list_str.split(",") - for attr_name in attr_list: - attr = "attr_" + attr_name.strip() - attr_dict = op_dict.get(attr) - if attr_dict: - is_valid = check_attr_dict(attr_dict, is_valid, attr) - else: - is_valid = False - print("%s is required in .ini file, when attr.list is %s!" % (attr, attr_list_str)) - return is_valid - - -def check_bool_flag(op_dict, is_valid): - """ - Function Description: - check_bool_flag - Parameter: op_dict - Parameter: is_valid - """ - for key in BOOL_FLAG_KEY: - if key in op_dict: - op_bool_key = op_dict.get(key) - if op_bool_key.get("flag").strip() not in BOOL_LIST: - is_valid = False - print("{0}.flag only support {1}.".format(key, BOOL_LIST)) - return is_valid - - -def check_type_format(op_info, is_valid, op_info_key): - """ - Function Description: - Check type and format - Parameter: op_info - Parameter: is_valid - Parameter: op_info_key - """ - op_info_dtype_str = op_info.get("dtype") - op_info_dtype_num = 0 - op_info_format_num = 0 - if op_info_dtype_str: - op_info_dtype = op_info_dtype_str.split(",") - op_info_dtype_num = len(op_info_dtype) - for dtype in op_info_dtype: - if dtype.strip() not in DTYPE_LIST: - is_valid = False - print("{0}.dtype not support {1}.".format(op_info_key, dtype)) - op_info_format_str = op_info.get("format") - if op_info_format_str: - op_info_format = op_info_format_str.split(",") - op_info_format_num = len(op_info_format) - for op_format in op_info_format: - if op_format.strip() not in FORMAT_LIST: - is_valid = False - print("{0}.format not support {1}.".format(op_info_key, op_format)) - if op_info_dtype_num > 0 and op_info_format_num > 0: - if op_info_dtype_num != op_info_format_num: - is_valid = False - print("The number of {0}.dtype not match the number of {0}.format.".format(op_info_key)) - return is_valid - - -def check_op_info(tbe_ops): - """ - Function Description: - Check info. - Parameter: tbe_ops - Return Value: is_valid - """ - print("\n\n==============check valid for ops info start==============") - required_op_input_info_keys = ["paramType", "name"] - required_op_output_info_keys = ["paramType", "name"] - param_type_valid_value = ["dynamic", "optional", "required"] - is_valid = True - for op_key in tbe_ops: - op_dict = tbe_ops[op_key] - for op_info_key in op_dict: - if op_info_key.startswith("input"): - op_input_info = op_dict[op_info_key] - missing_keys = [] - for required_op_input_info_key in required_op_input_info_keys: - if required_op_input_info_key not in op_input_info: - missing_keys.append(required_op_input_info_key) - if len(missing_keys) > 0: - print("op: " + op_key + " " + op_info_key + " missing: " + - ",".join(missing_keys)) - is_valid = False - else: - if not op_input_info["paramType"] in param_type_valid_value: - print("op: " + op_key + " " + op_info_key + \ - " paramType not valid, valid key:[dynamic, " - "optional, required]") - is_valid = False - is_valid = check_type_format(op_input_info, is_valid, op_info_key) - if op_info_key.startswith("output"): - op_input_info = op_dict[op_info_key] - missing_keys = [] - for required_op_input_info_key in required_op_output_info_keys: - if required_op_input_info_key not in op_input_info: - missing_keys.append(required_op_input_info_key) - if len(missing_keys) > 0: - print("op: " + op_key + " " + op_info_key + " missing: " + - ",".join(missing_keys)) - is_valid = False - else: - if not op_input_info["paramType"] in param_type_valid_value: - print("op: " + op_key + " " + op_info_key + - " paramType not valid, valid key:[dynamic, " - "optional, required]") - is_valid = False - is_valid = check_type_format(op_input_info, is_valid, op_info_key) - is_valid = check_attr(op_dict, is_valid) - is_valid = check_bool_flag(op_dict, is_valid) - print("==============check valid for ops info end================\n\n") - return is_valid - - -def write_json_file(tbe_ops_info, json_file_path): - """ - Save info to json file - Parameters: - ---------------- - tbe_ops_info: ops_info - json_file_path: json file path - ---------------- - """ - json_file_real_path = os.path.realpath(json_file_path) - wr_flag = os.O_WRONLY | os.O_CREAT - wr_mode = stat.S_IWUSR | stat.S_IRUSR - with os.fdopen(os.open(json_file_real_path, wr_flag, wr_mode), 'w') as file_path: - # The owner have all rights,group only have read rights - os.chmod(json_file_real_path, stat.S_IWUSR + stat.S_IRGRP - + stat.S_IRUSR) - json.dump(tbe_ops_info, file_path, sort_keys=True, indent=4, - separators=(',', ':')) - print("Compile op info cfg successfully.") - - -def parse_ini_to_json(ini_file_paths, outfile_path): - """ - parse ini files to json file - Parameters: - ---------------- - ini_file_paths: list of ini file path - outfile_path: output file path - ---------------- - """ - tbe_ops_info = parse_ini_files(ini_file_paths) - if not check_op_info(tbe_ops_info): - print("Compile op info cfg failed.") - return False - write_json_file(tbe_ops_info, outfile_path) - return True - - -if __name__ == '__main__': - args = sys.argv - - OUTPUT_FILE_PATH = "tbe_ops_info.json" - ini_file_path_list = [] - - for arg in args: - if arg.endswith("ini"): - ini_file_path_list.append(arg) - OUTPUT_FILE_PATH = arg.replace(".ini", ".json") - if arg.endswith("json"): - OUTPUT_FILE_PATH = arg - - if len(ini_file_path_list) == 0: - ini_file_path_list.append("tbe_ops_info.ini") - - if not parse_ini_to_json(ini_file_path_list, OUTPUT_FILE_PATH): - sys.exit(1) - sys.exit(0) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/preset_parse.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/preset_parse.py deleted file mode 100755 index 8f1124b1d..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/preset_parse.py +++ /dev/null @@ -1,23 +0,0 @@ -import json -import sys -import os - - -def get_config_opts(file): - src_dir = os.path.abspath(os.path.dirname(file)) - opts = '' - with open(file, 'r') as fd: - config = json.load(fd) - for conf in config: - if conf == 'configurePresets': - for node in config[conf]: - macros = node.get('cacheVariables') - if macros is not None: - for key in macros: - opts += '-D{}={} '.format(key, macros[key]['value']) - opts = opts.replace('${sourceDir}', src_dir) - print(opts) - - -if __name__ == "__main__": - get_config_opts(sys.argv[1]) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_codegen.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_codegen.py deleted file mode 100755 index 1baa364ef..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_codegen.py +++ /dev/null @@ -1,105 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import os -import stat -import collections -import kernel_entry as keb -from tiling_data_def_build import gen_tiling -import code_channel_infer -import const_var - -PYF_PATH = os.path.dirname(__file__) - -ReplayCodeGenParams = collections.namedtuple('ReplayCodeGenParams',\ -['op_type', 'impl', 'tiling_file', 'kernel', 'entry', 'argn', 'op_replay_batch', 'max_block_dim', 'max_shape_size']) - - -class ReplayCodeGen: - def __init__(self, replayCodeGenParams): - self.op_type = replayCodeGenParams.op_type - self.impl = replayCodeGenParams.impl - self.tiling_file = replayCodeGenParams.tiling_file - self.tiling_data_file = '' - self.kernel = replayCodeGenParams.kernel - self.entry = replayCodeGenParams.entry - self.argn = replayCodeGenParams.argn - self.batch = False - self.outdir = '' - self.data_type = 'uint8_t' - self.blknum = 32 - self.op_replay_batch = replayCodeGenParams.op_replay_batch - self.max_block_dim = replayCodeGenParams.max_block_dim - self.max_shape_size = replayCodeGenParams.max_shape_size - - def set_batch(self, is_batch): - self.batch = is_batch - - def set_outdir(self, outdir): - self.outdir = outdir - - def gen_replay(self, ops_product: str): - kerentry = os.path.join(self.outdir, self.kernel + '_entry.cce') - kerimpl = os.path.join(self.outdir, self.kernel + '_impl.cpp') - replayimpl = os.path.join(self.outdir, self.kernel + '_replay.cpp') - if self.batch: - reptmp = os.path.join(PYF_PATH, 'batch_replay_impl.temp') - else: - reptmp = os.path.join(PYF_PATH, 'replay_impl.temp') - kertmp = os.path.join(PYF_PATH, 'kernel_impl.temp') - self._gen_kentry(kerentry) - self._gen_kimpl_code(kerimpl, kertmp) - self._gen_tiling_data_header() - self._gen_replay_code(replayimpl, reptmp, ops_product) - - def _gen_tiling_data_header(self): - self.tiling_data_file = os.path.join(self.outdir, self.kernel + '_tiling_data.h') - gen_tiling(self.tiling_file, self.tiling_data_file) - - def _gen_kimpl_code(self, src, tmpfile): - with open(tmpfile, 'r') as fd: - temp = fd.read() - temp = temp.replace('__CCE_FILE__', self.impl) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(temp) - - def _gen_replay_code(self, src, tmpfile, ops_product: str): - with open(tmpfile, 'r') as fd: - temp = fd.read() - temp = temp.replace('__ARG_NUM__', str(self.argn)) - argdef = [] - kargs = [] - for i in range(0, self.argn): - argdef.append('{} *'.format(self.data_type)) - kargs.append('({} *)GetArg({})'.format(self.data_type, i)) - temp = temp.replace('__ARGS_DEF__', ', '.join(argdef)) - temp = temp.replace('__KERNEL_ARGS__', ', '.join(kargs)) - temp = temp.replace('__KERNEL_FUN__', self.entry) - core_type_infer = 'core_type' - code_channel = code_channel_infer.infer_code_channel(code_channel_infer.InfoCodeChanelParams(self.impl,\ - self.tiling_data_file, self.kernel, self.outdir, ops_product, None)) - if code_channel == code_channel_infer.CODE_VEC: - core_type_infer = '0' - elif code_channel == code_channel_infer.CODE_CUBE: - core_type_infer = '1' - temp = temp.replace('__CORE_TYPE__', core_type_infer) - # regist function - temp = temp.replace('__OPS_PRODUCT__', ops_product) - temp = temp.replace('__OPTYPE__', self.op_type) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(temp) - - def _gen_kentry(self, src): - kf = '' - pre_alloc_str = 'A' * 256 - if self.batch: - kf += keb.batch_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str), self.argn, self.data_type) - else: - kf += keb.mc_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str),\ - self.argn, self.data_type, self.blknum) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(kf) diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_impl.temp b/atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_impl.temp deleted file mode 100755 index 1d30dd865..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/replay_impl.temp +++ /dev/null @@ -1,120 +0,0 @@ -#include -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#include "register/op_check.h" -#define __ASCENDC_REPLAY_CODE__ -using namespace std; -using namespace optiling; -using namespace AscendCReplay; - -extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); -extern "C" int elf_append(char *elf, uint32_t elfSize, char *jit, int kernum, int blknum[], char *atext[], - int alen[], int atlen, const char* kernelname[]); - -#define KERNEL_N 1 -#define ARG_N (__ARG_NUM__) -#define MAX_L (1024 * 1024 * 100) -#define MAX_E (1024 * 1024) - -int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) -{ - // gen type 1 : direct call codes 0: load .o file - if (param.gentype < 0 || param.gentype > 1) { - printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); - return 0; - } else if (param.gentype == 1 && param.objptr == nullptr) { - printf("Error: call replay with direct call mode, but code obj addr is null\n"); - return 0; - } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { - printf("Error: call replay with object file mode, but object file path is null\n"); - return 0; - } - // core_type 0:MIX 1:CUBE 2:VEC - if (core_type < 0 || core_type > 2) { - printf("Error: call replay core type is %d !\n", core_type); - return 0; - } - g_coreType = __CORE_TYPE__; - g_taskRation = param.task_ration; - g_tilingKey = param.tiling_key; - - unsigned char *buf, *jit; - char *kernel[KERNEL_N * 32]; - int len[KERNEL_N * 32]; - int blknum[KERNEL_N]; - int max; - block_num = param.block_dim; - g_ubBase = block_num; - uint8_t *code = (uint8_t *)malloc(MAX_L); - uint8_t *pos = code; - struct timespec tp1, tp2; - - clock_gettime(CLOCK_MONOTONIC, &tp1); - if (block_num > 32) { - printf("Error: block_num > 32\n"); - return 0; - } - //__OP_FOPEN__ - for (int i = 0; i < KERNEL_N; i++) { - for (int j = 0; j < ARG_N; j++) - AddArg(j, ARG_STEP * (j + 1)); - for (block_idx = 0; block_idx < block_num; block_idx++) { - //__OP_SET_KERNEL__ - int code_idx = i * block_num + block_idx; -#ifdef FP_CEILING - SetCtrlFloatEnable(); -#else - SetCtrlFloatDisable(); -#endif - CodeInit(pos, false); - __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); - CodeEnd(); - kernel[code_idx] = (char *)pos; - len[code_idx] = CodeLen(); - pos += len[code_idx]; - printf("kernel %d core %ld code generated len %d\n", i, block_idx, len[code_idx]); - } - blknum[i] = block_num; - } - //__OP_FCLOSE__ - clock_gettime(CLOCK_MONOTONIC, &tp2); - buf = (unsigned char *)malloc(MAX_E); - int fd = open(param.entry_file, O_RDONLY); - if (fd < 0) { - printf("[error]: cannot find entry.o : %s\n", param.entry_file); - return 0; - } - uint32_t bufSize = read(fd, buf, MAX_E); - if (bufSize <= 0) { - printf("[error]: entry.o : %s is too small ! \n", param.entry_file); - } - close(fd); - jit = (unsigned char *)malloc(MAX_L); - printf("total code generated %ld\n", pos - code); - int sz = elf_append((char *)buf, bufSize, (char *)jit, KERNEL_N, blknum, kernel, len, pos - code, ¶m.kernel_name); - if (tp1.tv_sec != tp2.tv_sec) { - printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); - } else { - printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); - } - printf("new elf size %d\n", sz); - if (param.gentype == 0) { - fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); - (void)write(fd, jit, sz); - close(fd); - free(jit); - } else if (param.gentype == 1) { - *param.objptr = (char*)jit; - } - free(buf); - free(code); - return sz; -} - -REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbEyeSample/EyeOp/cmake/util/tiling_data_def_build.py b/atb_operator/AtbEyeSample/EyeOp/cmake/util/tiling_data_def_build.py deleted file mode 100755 index 4c5b099b7..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/cmake/util/tiling_data_def_build.py +++ /dev/null @@ -1,87 +0,0 @@ -#!/usr/bin/env python -# coding=utf-8 -""" -Function: -The replay funtion entry -Copyright Information: -Huawei Technologies Co., Ltd. All Rights Reserved © 2020 -""" - -import sys -import os -import stat -import re -import const_var - - -def gen_tiling(tiling_header_file: str, tiling_file_out: str): - if not os.path.exists(tiling_header_file): - print("warning: no userdef tiling header file: ", tiling_header_file) - return - print("generate tiling def header file: ", tiling_file_out) - tmp_name = os.path.splitext(os.path.basename(tiling_header_file))[0].upper() - tiling_source = '#ifndef __{}_H__\n'.format(tmp_name) - tiling_source += '#define __{}_H__\n\n'.format(tmp_name) - tiling_source += '#include \n' - tiling_source += '#include \n\n' - tiling_source += '#include "kernel_tiling/kernel_tiling.h"\n\n' - end_source = "" - pattern = re.compile(r'[(](.*)[)]', re.S) - with open(tiling_header_file, 'r') as fd: - lines = fd.readlines() - for line in lines: - line = line.strip() - if (line.startswith('BEGIN_TILING_DATA_DEF')): - tiling_source += '#pragma pack(1)\n' - tiling_source += 'struct ' - struct_def = re.findall(pattern, line)[0] - tiling_source += struct_def + ' {\n' - elif (line.startswith('TILING_DATA_FIELD_DEF_ARR')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {}[{}] = {{}};\n'.format(fds[0].strip(), fds[2].strip(), fds[1].strip()) - elif (line.startswith('TILING_DATA_FIELD_DEF_STRUCT')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {};\n'.format(fds[0].strip(), fds[1].strip()) - elif (line.startswith('TILING_DATA_FIELD_DEF')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {} = 0;\n'.format(fds[0].strip(), fds[1].strip()) - elif (line.startswith('END_TILING_DATA_DEF')): - tiling_source += '};\n' - tiling_source += '#pragma pack()\n\n' - tiling_source += '#ifdef __NPU_TILING__\n' - tiling_source += \ - 'inline [aicore] void Init{stru}(const __gm__ uint8_t* tiling, {stru}* const_data)\n'\ - .format(stru=struct_def) - tiling_source += '{\n' - tiling_source += ' const __gm__ uint32_t *src = (const __gm__ uint32_t *)tiling;\n' - tiling_source += ' uint32_t *dst = (uint32_t *)const_data;\n' - tiling_source += ' for (auto i = 0; i < sizeof({}) / 4; i++) *(dst + i) = *(src + i);\n'\ - .format(struct_def) - tiling_source += '}\n' - tiling_source += '#else\n' - tiling_source += 'inline void Init{stru}(uint8_t* tiling, {stru}* const_data)\n'.format(stru=struct_def) - tiling_source += '{\n' - tiling_source += ' uint64_t *src = (uint64_t *)tiling;\n' - tiling_source += ' uint64_t *dst = (uint64_t *)const_data;\n' - tiling_source += ' for (auto i = 0; i < sizeof({}) / 8; i++) *(dst + i) = *(src + i);\n'\ - .format(struct_def) - tiling_source += '}\n' - tiling_source += '#endif\n\n' - end_source = ''' -#define GET_TILING_DATA(tiling_data, tiling_arg) \\ -{stru} tiling_data; \\ -Init{stru}(tiling_arg, &tiling_data)\n -'''.format(stru=struct_def) - tiling_source += end_source - tiling_source += '#endif' - with os.fdopen(os.open(tiling_file_out, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(tiling_source) - - -if __name__ == '__main__': - if len(sys.argv) <= 2: - raise RuntimeError('arguments must greater than 2') - gen_tiling(sys.argv[1], sys.argv[2]) diff --git a/atb_operator/AtbEyeSample/EyeOp/framework/CMakeLists.txt b/atb_operator/AtbEyeSample/EyeOp/framework/CMakeLists.txt deleted file mode 100755 index b6be9b492..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/framework/CMakeLists.txt +++ /dev/null @@ -1,11 +0,0 @@ -if(NOT EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/mindspore") - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/caffe_plugin") - add_subdirectory(caffe_plugin) - endif() - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/tf_plugin") - add_subdirectory(tf_plugin) - endif() - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/onnx_plugin") - add_subdirectory(onnx_plugin) - endif() -endif() diff --git a/atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/CMakeLists.txt b/atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/CMakeLists.txt deleted file mode 100755 index a6aba5c20..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/CMakeLists.txt +++ /dev/null @@ -1,14 +0,0 @@ - -aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} plugin_srcs) -add_library(cust_tf_parsers SHARED ${plugin_srcs}) -target_compile_definitions(cust_tf_parsers PRIVATE google=ascend_private) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_tf_parsers PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_tf_parsers PRIVATE intf_pub graph) -install(TARGETS cust_tf_parsers - LIBRARY DESTINATION packages/vendors/${vendor_name}/framework/tensorflow -) diff --git a/atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/tensorflow_eye_plugin.cc b/atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/tensorflow_eye_plugin.cc deleted file mode 100755 index c48076386..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/framework/tf_plugin/tensorflow_eye_plugin.cc +++ /dev/null @@ -1,12 +0,0 @@ -/* 版权所有 (c) 华为技术有限公司 2020-2021 - * 注册 Eye 操作信息到 GE - */ -#include "register/register.h" - -namespace domi { -// register op info to GE -REGISTER_CUSTOM_OP("Eye") - .FrameworkType(TENSORFLOW) // type: CAFFE, TENSORFLOW - .OriginOpType("Eye") // name in tf module - .ParseParamsByOperatorFn(AutoMappingByOpFn); -} // namespace domi diff --git a/atb_operator/AtbEyeSample/EyeOp/op_host/CMakeLists.txt b/atb_operator/AtbEyeSample/EyeOp/op_host/CMakeLists.txt deleted file mode 100755 index 40dd51cfa..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/op_host/CMakeLists.txt +++ /dev/null @@ -1,82 +0,0 @@ - -aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} ops_srcs) - -opbuild(OPS_SRC ${ops_srcs} - OUT_DIR ${ASCEND_AUTOGEN_PATH} -) - -add_library(cust_op_proto SHARED ${ops_srcs} ${ASCEND_AUTOGEN_PATH}/op_proto.cc) -target_compile_definitions(cust_op_proto PRIVATE OP_PROTO_LIB) -target_compile_options(cust_op_proto PRIVATE - -fvisibility=hidden -) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_op_proto PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_op_proto PRIVATE - intf_pub - exe_graph - register - tiling_api - -Wl,--whole-archive - rt2_registry - -Wl,--no-whole-archive -) -set_target_properties(cust_op_proto PROPERTIES OUTPUT_NAME - cust_opsproto_rt2.0 -) -add_library(cust_optiling SHARED ${ops_srcs}) -target_compile_definitions(cust_optiling PRIVATE OP_TILING_LIB) -target_compile_options(cust_optiling PRIVATE - -fvisibility=hidden -) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_optiling PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_optiling PRIVATE - intf_pub - exe_graph - register - tiling_api - -Wl,--whole-archive - rt2_registry - -Wl,--no-whole-archive -) -set_target_properties(cust_optiling PROPERTIES OUTPUT_NAME - cust_opmaster_rt2.0 -) - -file(GLOB aclnn_src ${ASCEND_AUTOGEN_PATH}/aclnn_*.cpp) -file(GLOB aclnn_inc ${ASCEND_AUTOGEN_PATH}/aclnn_*.h) -add_library(cust_opapi SHARED ${aclnn_src}) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_opapi PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_opapi PRIVATE intf_pub ascendcl nnopbase) - -add_custom_target(optiling_compat ALL - COMMAND ln -sf lib/linux/${CMAKE_SYSTEM_PROCESSOR}/$ - ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so -) - -install(TARGETS cust_op_proto - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_proto/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) -install(FILES ${ASCEND_AUTOGEN_PATH}/op_proto.h - DESTINATION packages/vendors/${vendor_name}/op_proto/inc) -install(TARGETS cust_optiling - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) -install(FILES ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so - DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling) -install(TARGETS cust_opapi - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_api/lib) -install(FILES ${aclnn_inc} - DESTINATION packages/vendors/${vendor_name}/op_api/include) diff --git a/atb_operator/AtbEyeSample/EyeOp/op_host/eye.cpp b/atb_operator/AtbEyeSample/EyeOp/op_host/eye.cpp deleted file mode 100755 index cf72d4d17..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/op_host/eye.cpp +++ /dev/null @@ -1,173 +0,0 @@ -/** -* @file eye.cpp -* -* Copyright (C) 2023. Huawei Technologies Co., Ltd. All rights reserved. -* -* This program is distributed in the hope that it will be useful, -* but WITHOUT ANY WARRANTY; without even the implied warranty of -* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. -*/ - -#include "eye_tiling.h" -#include "register/op_def_registry.h" -#include "tiling/platform/platform_ascendc.h" - -namespace optiling { - const uint32_t BLOCK_SIZE = 32; - constexpr int BATCH_SIZE_DIM_THRESHOLD = 2; - constexpr uint32_t DATATYPE_SIZE_2 = 2; - constexpr uint32_t DATATYPE_SIZE_4 = 4; - constexpr uint32_t TYPE_KEY_0 = 0; - constexpr uint32_t TYPE_KEY_1 = 1; - constexpr uint32_t DATA_NUM = 4; - static ge::graphStatus TilingFunc(gert::TilingContext* context) { - EyeTilingData tiling; - uint32_t sizeofdatatype; - uint32_t totalLengthAligned; - - // 1. 获取平台信息 - uint64_t ub_size; - auto ascendcPlatform = platform_ascendc::PlatformAscendC(context->GetPlatformInfo()); - ascendcPlatform.GetCoreMemSize(platform_ascendc::CoreMemType::UB, ub_size); - - // 2. 获取数据信息 - uint32_t dataNum; - uint32_t totalLength = context->GetInputTensor(0)->GetShapeSize(); - uint32_t tanhxLength = context->GetOptionalInputTensor(0)->GetShapeSize(); - const int64_t* pnum_rows = context->GetAttrs()->GetInt(0); - const int64_t* pnum_columns = context->GetAttrs()->GetInt(1); - const int64_t* pdtype = context->GetAttrs()->GetInt(2); - auto *pbatch_shapecv = context->GetAttrs()->GetListInt(0); - int32_t num_rows = *pnum_rows; - int32_t num_columns = *pnum_columns; - int32_t dtype = *pdtype; - const int64_t *pbatch_shape = pbatch_shapecv->GetData(); - int32_t batchShapeSize = pbatch_shapecv->GetSize(); - int32_t batch_shape_list[10]; - if(num_columns == 0){ - num_columns = num_rows; - } - - int32_t mark = 0; - int32_t batchNum = 1; - int32_t batchSize = 0; - auto shape = context->GetInputTensor(0)->GetOriginShape(); - const uint32_t dimNum = shape.GetDimNum(); - if(dimNum > BATCH_SIZE_DIM_THRESHOLD){ - mark = 1; - batchSize = num_rows * num_columns; - batchNum = totalLength / batchSize; - } - - tiling.set_num_columns(num_columns); - tiling.set_num_rows(num_rows); - tiling.set_batch_shape(batch_shape_list); - tiling.set_dtype(dtype); - tiling.set_mark(mark); - tiling.set_batchNum(batchNum); - tiling.set_batchSize(batchSize); - - auto dt = context->GetInputTensor(0)->GetDataType(); - uint32_t typeKey; - if (dt == 1) { - sizeofdatatype = DATATYPE_SIZE_2; - dataNum = DATA_NUM; - typeKey = TYPE_KEY_0; - }else{ - sizeofdatatype = DATATYPE_SIZE_4; - dataNum = DATA_NUM; - typeKey = TYPE_KEY_1; - } - - tiling.set_typeKey(typeKey); - - // 3. 填满UB大小 - uint32_t ub_block_num = ub_size / BLOCK_SIZE / dataNum - 256; - if (ub_block_num % 2 != 0) { - ub_block_num = ub_block_num - 1; - } - - // 4. 输入向量满足32字节对齐 - uint32_t ALIGN_NUM = BLOCK_SIZE / sizeofdatatype; - if (totalLength % ALIGN_NUM != 0) { //不对齐,先32位对齐 - totalLengthAligned = ((totalLength + ALIGN_NUM - 1) / ALIGN_NUM) * ALIGN_NUM; - } else { - totalLengthAligned = totalLength; - } - - // 5. Tiling参数计算 - uint32_t tile_num, block_dim = 1; - context->SetBlockDim(block_dim); - uint32_t blockLength = 0; - uint32_t tileLength = 0; - uint32_t lasttileLength = 0; - blockLength = totalLengthAligned / block_dim; - tile_num = blockLength / ALIGN_NUM / ub_block_num; - - if (tile_num == 0) { // 不足一个ub的情况 - tile_num = 1; - tileLength = ((blockLength / ALIGN_NUM) + 1) / 2 * 2 * ALIGN_NUM; - lasttileLength = tileLength; - } else if((blockLength / ALIGN_NUM) % ub_block_num == 0){ // 核内能均分 - tileLength = ub_block_num * ALIGN_NUM; - lasttileLength = tileLength; - }else{ // 核内不能均分 - tile_num = tile_num + 1; // 加一个小包的数量 - tileLength = ub_block_num * ALIGN_NUM; - lasttileLength = blockLength - (tile_num - 1) * tileLength; - lasttileLength = ((lasttileLength / ALIGN_NUM) + 1) / 2 * 2 * ALIGN_NUM; - } - - tiling.set_blockLength(blockLength); - tiling.set_tileNum(tile_num); - tiling.set_tileLength(tileLength); - tiling.set_lasttileLength(lasttileLength); - tiling.SaveToBuffer(context->GetRawTilingData()->GetData(), - context->GetRawTilingData()->GetCapacity()); - context->GetRawTilingData()->SetDataSize(tiling.GetDataSize()); - size_t* currentWorkspace = context->GetWorkspaceSizes(1); - currentWorkspace[0] = 0; - return ge::GRAPH_SUCCESS; - } -} - -namespace ge { - static ge::graphStatus InferShape(gert::InferShapeContext* context) { - const gert::Shape* x1_shape = context->GetInputShape(0); - gert::Shape* y_shape = context->GetOutputShape(0); - *y_shape = *x1_shape; - return GRAPH_SUCCESS; - } -} - - -namespace ops { -class Eye : public OpDef { -public: - explicit Eye(const char* name) : OpDef(name) { - this->Input("y") - .ParamType(REQUIRED) - .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32}) - .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) - .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); - this->Output("y") - .ParamType(REQUIRED) - .DataType({ge::DT_FLOAT, ge::DT_FLOAT16, ge::DT_INT32}) - .Format({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}) - .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND, ge::FORMAT_ND}); - this->Attr("num_rows").Int(); - this->Attr("num_columns").AttrType(OPTIONAL).Int(0); - this->Attr("batch_shape").AttrType(OPTIONAL).ListInt({}); - this->Attr("dtype").AttrType(OPTIONAL).Int(0); - - this->SetInferShape(ge::InferShape); - - this->AICore() - .SetTiling(optiling::TilingFunc); - this->AICore().AddConfig("ascend310b") - .AddConfig("ascend910b"); - } -}; - -OP_ADD(Eye); -} diff --git a/atb_operator/AtbEyeSample/EyeOp/op_host/eye_tiling.h b/atb_operator/AtbEyeSample/EyeOp/op_host/eye_tiling.h deleted file mode 100755 index 6aadcaa35..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/op_host/eye_tiling.h +++ /dev/null @@ -1,31 +0,0 @@ -/** -* @file eye_tiling.h -* -* Copyright (C) 2023. Huawei Technologies Co., Ltd. All rights reserved. -* -* This program is distributed in the hope that it will be useful, -* but WITHOUT ANY WARRANTY; without even the implied warranty of -* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. -*/ - -#include "register/tilingdata_base.h" - -namespace optiling { -BEGIN_TILING_DATA_DEF(EyeTilingData) - TILING_DATA_FIELD_DEF(uint32_t, typeKey); - TILING_DATA_FIELD_DEF(uint32_t, blockLength); - TILING_DATA_FIELD_DEF(uint32_t, tileNum); - TILING_DATA_FIELD_DEF(uint32_t, tileLength); - TILING_DATA_FIELD_DEF(uint32_t, lasttileLength); - TILING_DATA_FIELD_DEF(int32_t, num_rows); - TILING_DATA_FIELD_DEF(int32_t, num_columns); - TILING_DATA_FIELD_DEF(int32_t, dtype); - TILING_DATA_FIELD_DEF(int32_t, mark); - TILING_DATA_FIELD_DEF(int32_t, batchNum); - TILING_DATA_FIELD_DEF(int32_t, batchSize); - TILING_DATA_FIELD_DEF_ARR(int32_t, 10, batch_shape); - -END_TILING_DATA_DEF; - -REGISTER_TILING_DATA_CLASS(Eye, EyeTilingData) -} diff --git a/atb_operator/AtbEyeSample/EyeOp/op_kernel/CMakeLists.txt b/atb_operator/AtbEyeSample/EyeOp/op_kernel/CMakeLists.txt deleted file mode 100755 index 8c94a952d..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/op_kernel/CMakeLists.txt +++ /dev/null @@ -1,68 +0,0 @@ -# set custom compile options -if ("${CMAKE_BUILD_TYPE}x" STREQUAL "Debugx") - add_ops_compile_options(ALL OPTIONS -g -O0) -endif() - -foreach(compute_unit ${ASCEND_COMPUTE_UNIT}) - - # generate aic-${compute_unit}-ops-info.json - add_ops_info_target(TARGET ops_info_gen_${compute_unit} - OUTPUT ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core/${compute_unit}/aic-${compute_unit}-ops-info.json - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/config/${compute_unit} - ) - - # generate ascendc impl py once - if (NOT TARGET ascendc_impl_gen) - add_ops_impl_target(TARGET ascendc_impl_gen - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl - ) - endif() - - # dynamic shape binary compile - if (${ENABLE_BINARY_PACKAGE} AND NOT ${ENABLE_CROSS_COMPILE}) - add_bin_compile_target(TARGET ascendc_bin_${compute_unit} - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} - ADP_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/dynamic - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/binary/${compute_unit} - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/kernel - COMPUTE_UNIT ${compute_unit} - ) - add_dependencies(ascendc_bin_${compute_unit} ascendc_impl_gen) - endif() - - if (${ENABLE_CROSS_COMPILE} AND ${ENABLE_BINARY_PACKAGE}) - add_cross_compile_target( - TARGET bin_${compute_unit} - OUT_DIR ${CMAKE_CURRENT_SOURCE_DIR}/../kernel - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/ - ) - endif() -endforeach() - -# generate npu_supported_ops.json -add_npu_support_target(TARGET npu_supported_ops - OPS_INFO_DIR ${ASCEND_AUTOGEN_PATH} - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core - INSTALL_DIR packages/vendors/${vendor_name}/framework/${ASCEND_FRAMEWORK_TYPE} -) - -if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) - add_subdirectory(testcases) -endif() - -# install kernel file -if (${ENABLE_SOURCE_PACKAGE}) - file(GLOB KERNEL_FILES - ${CMAKE_CURRENT_SOURCE_DIR}/*.cpp - ${CMAKE_CURRENT_SOURCE_DIR}/*.h - ${CMAKE_CURRENT_SOURCE_DIR}/*.py - ) - install(FILES ${KERNEL_FILES} - DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl/dynamic - ) -endif() diff --git a/atb_operator/AtbEyeSample/EyeOp/op_kernel/eye.cpp b/atb_operator/AtbEyeSample/EyeOp/op_kernel/eye.cpp deleted file mode 100755 index 9aef97978..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/op_kernel/eye.cpp +++ /dev/null @@ -1,91 +0,0 @@ -/** -* @file eye.cpp -* -* Copyright (C) 2023. Huawei Technologies Co., Ltd. All rights reserved. -* -* This program is distributed in the hope that it will be useful, -* but WITHOUT ANY WARRANTY; without even the implied warranty of -* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. -*/ - -#include "kernel_operator.h" - -using namespace AscendC; -constexpr int32_t BUFFER_NUM = 2; - -class KernelEye { -public: - __aicore__ inline KernelEye() {} - __aicore__ inline void Init(GM_ADDR y, GM_ADDR y_ref, - uint32_t typeKey, uint32_t blockLength, uint32_t tileNum, - uint32_t tileLength, uint32_t lasttileLength, int32_t num_columns, - int32_t num_rows, int32_t *batch_shape, int32_t dtype, - int32_t mark, int32_t batchSize, int32_t batchNum) { - - this->batchSize = batchSize; - this->batchNum = batchNum; - this->mark = mark; - this->num_rows = num_rows; - this->num_columns = num_columns; - this->batch_shape = batch_shape; - this->dtype = dtype; - this->blockLength = blockLength; - this->tileNum = tileNum; - this->tileLength = tileLength / BUFFER_NUM; - this->lasttileLength = lasttileLength / BUFFER_NUM; - this->typeKey = typeKey; - - yGm.SetGlobalBuffer((__gm__ DTYPE_Y*)y_ref + this->blockLength * GetBlockIdx(), this->blockLength); - } - __aicore__ inline void Process() { - int32_t index, t; - if(mark == 0){ - for(int32_t i = 0; i < num_rows; i++){ - index = i * num_columns + i; - if(i < num_columns){ - yGm.SetValue(index, 1); - } - } - }else{ - for(int32_t i = 0; i < batchNum; i++){ - for(int32_t j = 0; j < num_rows; j++){ - if(j < num_columns){ - t = j * num_columns + j; - index = i * batchSize + t; - yGm.SetValue(index, 1); - } - } - } - } - } - -private: - TPipe pipe; - TBuf tmpBuf1, tmpBuf2, tmpBuf3; - TQue inQueueIN; - TQue outQueueOUT; - GlobalTensor yGm; - uint32_t blockLength; - uint32_t tileNum; - uint32_t tileLength; - uint32_t lasttileLength; - uint32_t typeKey; - int32_t *batch_shape; - int32_t dtype; - int32_t num_columns; - int32_t num_rows; - int32_t mark; - int32_t batchSize, batchNum; -}; - -extern "C" __global__ __aicore__ void eye(GM_ADDR y, GM_ADDR y_ref, GM_ADDR workspace, GM_ADDR tiling) { - GET_TILING_DATA(tiling_data, tiling); - - KernelEye op; - - op.Init(y, y_ref, tiling_data.typeKey, tiling_data.blockLength, - tiling_data.tileNum, tiling_data.tileLength, tiling_data.lasttileLength, - tiling_data.num_columns, tiling_data.num_rows, tiling_data.batch_shape, - tiling_data.dtype, tiling_data.mark, tiling_data.batchSize, tiling_data.batchNum); - op.Process(); -} \ No newline at end of file diff --git a/atb_operator/AtbEyeSample/EyeOp/scripts/install.sh b/atb_operator/AtbEyeSample/EyeOp/scripts/install.sh deleted file mode 100755 index 8468c5a25..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/scripts/install.sh +++ /dev/null @@ -1,318 +0,0 @@ -#!/bin/bash -vendor_name=customize -targetdir=/usr/local/Ascend/opp -target_custom=0 - -sourcedir=$PWD/packages -vendordir=vendors/$vendor_name - -QUIET="y" - -while true -do - case $1 in - --quiet) - QUIET="y" - shift - ;; - --install-path=*) - INSTALL_PATH=$(echo $1 | cut -d"=" -f2-) - INSTALL_PATH=${INSTALL_PATH%*/} - shift - ;; - --*) - shift - ;; - *) - break - ;; - esac -done - -log() { - cur_date=`date +"%Y-%m-%d %H:%M:%S"` - echo "[runtime] [$cur_date] "$1 -} - -if [ -n "${INSTALL_PATH}" ]; then - if [[ ! "${INSTALL_PATH}" = /* ]]; then - log "[ERROR] use absolute path for --install-path argument" - exit 1 - fi - if [ ! -d ${INSTALL_PATH} ]; then - mkdir ${INSTALL_PATH} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${INSTALL_PATH} failed" - exit 1 - fi - fi - targetdir=${INSTALL_PATH} -elif [ -n "${ASCEND_CUSTOM_OPP_PATH}" ]; then - if [ ! -d ${ASCEND_CUSTOM_OPP_PATH} ]; then - mkdir -p ${ASCEND_CUSTOM_OPP_PATH} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${ASCEND_CUSTOM_OPP_PATH} failed" - fi - fi - targetdir=${ASCEND_CUSTOM_OPP_PATH} -else - if [ "x${ASCEND_OPP_PATH}" == "x" ]; then - log "[ERROR] env ASCEND_OPP_PATH no exist" - exit 1 - fi - targetdir="${ASCEND_OPP_PATH}" -fi - -if [ ! -d $targetdir ];then - log "[ERROR] $targetdir no exist" - exit 1 -fi - -upgrade() -{ - if [ ! -d ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 files" - return 0 - fi - - if [ ! -d ${targetdir}/$vendordir/$1 ];then - log "[INFO] create ${targetdir}/$vendordir/$1." - mkdir -p ${targetdir}/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/$1 failed" - return 1 - fi - else - has_same_file=-1 - for file_a in ${sourcedir}/$vendordir/$1/*; do - file_b=${file_a##*/}; - if [ "ls ${targetdir}/$vendordir/$1" = "" ]; then - log "[INFO] ${targetdir}/$vendordir/$1 is empty !!" - return 1 - fi - grep -q $file_b <<<`ls ${targetdir}/$vendordir/$1`; - if [[ $? -eq 0 ]]; then - echo -n "${file_b} " - has_same_file=0 - fi - done - if [ 0 -eq $has_same_file ]; then - if test $QUIET = "n"; then - echo "[INFO]: has old version in ${targetdir}/$vendordir/$1, \ - you want to Overlay Installation , please enter:[o]; \ - or replace directory installation , please enter: [r]; \ - or not install , please enter:[n]." - - while true - do - read orn - if [ "$orn" = n ]; then - return 0 - elif [ "$orn" = m ]; then - break; - elif [ "$0rn" = r ]; then - [ -n "${targetdir}/$vendordir/$1/" ] && rm -rf "${targetdir}/$vendordir/$1"/* - break; - else - echo "[ERROR] input error, please input again!" - fi - done - fi - fi - log "[INFO] replace or merge old ops $1 files .g....." - fi - - log "copy new ops $1 files ......" - if [ -d ${targetdir}/$vendordir/$1/ ]; then - chmod -R +w "$targetdir/$vendordir/$1/" >/dev/null 2>&1 - fi - cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 files failed" - return 1 - fi - - return 0 -} -upgrade_proto() -{ - if [ ! -f ${sourcedir}/$vendordir/custom.proto ]; then - log "[INFO] no need to upgrade custom.proto files" - return 0 - fi - if [ ! -d ${targetdir}/$vendordir/framework/caffe ];then - log "[INFO] create ${targetdir}/$vendordir/framework/caffe." - mkdir -p ${targetdir}/$vendordir/framework/caffe - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/framework/caffe failed" - return 1 - fi - else - if [ -f ${targetdir}/$vendordir/framework/caffe/custom.proto ]; then - # 有老版本,判断是否要覆盖式安装 - if test $QUIET = "n"; then - echo "[INFO] ${targetdir}/$vendordir/framework/caffe has old version"\ - "custom.proto file. Do you want to replace? [y/n] " - - while true - do - read yn - if [ "$yn" = n ]; then - return 0 - elif [ "$yn" = y ]; then - break; - else - echo "[ERROR] input error, please input again!" - fi - done - fi - fi - log "[INFO] replace old caffe.proto files ......" - fi - chmod -R +w "$targetdir/$vendordir/framework/caffe/" >/dev/null 2>&1 - cp -rf ${sourcedir}/$vendordir/custom.proto ${targetdir}/$vendordir/framework/caffe/ - if [ $? -ne 0 ];then - log "[ERROR] copy new custom.proto failed" - return 1 - fi - log "[INFO] copy custom.proto success" - - return 0 -} - -upgrade_file() -{ - if [ ! -e ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 file" - return 0 - fi - - log "copy new $1 files ......" - cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 file failed" - return 1 - fi - - return 0 -} - -delete_optiling_file() -{ - if [ ! -d ${targetdir}/vendors ];then - log "[INFO] $1 not exist, no need to uninstall" - return 0 - fi - sys_info=$(uname -m) - if [ ! -d ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/lib/linux/${sys_info} ];then - rm -rf ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/liboptiling.so - fi - return 0 -} - -log "[INFO] copy uninstall sh success" - -if [ ! -d ${targetdir}/vendors ];then - log "[INFO] create ${targetdir}/vendors." - mkdir -p ${targetdir}/vendors - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/vendors failed" - return 1 - fi -fi -chmod u+w ${targetdir}/vendors - -echo "[ops_custom]upgrade framework" -upgrade framework -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op proto" -upgrade op_proto -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade version.info" -upgrade_file version.info -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op impl" -delete_optiling_file op_impl -upgrade op_impl -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op api" -upgrade op_api -if [ $? -ne 0 ];then - exit 1 -fi - -upgrade_proto -if [ $? -ne 0 ];then - exit 1 -fi - -# set the set_env.bash -if [ -n "${INSTALL_PATH}" ] && [ -d ${INSTALL_PATH} ]; then - _ASCEND_CUSTOM_OPP_PATH=${targetdir}/${vendordir} - bin_path="${_ASCEND_CUSTOM_OPP_PATH}/bin" - set_env_variable="#!/bin/bash\nexport ASCEND_CUSTOM_OPP_PATH=${_ASCEND_CUSTOM_OPP_PATH}:\${ASCEND_CUSTOM_OPP_PATH}" - if [ ! -d ${bin_path} ]; then - mkdir -p ${bin_path} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${bin_path} failed" - exit 1 - fi - fi - echo -e ${set_env_variable} > ${bin_path}/set_env.bash - if [ $? -ne 0 ]; then - log "[ERROR] write ASCEND_CUSTOM_OPP_PATH to set_env.bash failed" - exit 1 - else - log "[INFO] using requirements: when custom module install finished or before you run the custom module, \ - execute the command [ source ${bin_path}/set_env.bash ] to set the environment path" - fi -else - config_file=${targetdir}/vendors/config.ini - if [ ! -f ${config_file} ]; then - touch ${config_file} - chmod 640 ${config_file} - echo "load_priority=$vendor_name" > ${config_file} - if [ $? -ne 0 ];then - echo "echo load_priority failed" - exit 1 - fi - else - found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" - found_vendor=$(echo $found_vendors | sed "s/$vendor_name//g" | tr ',' ' ') - vendor=$(echo $found_vendor | tr -s ' ' ',') - if [ "$vendor" != "" ]; then - sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" - fi - fi -fi - -chmod u-w ${targetdir}/vendors - -if [ -d ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/ ]; then - chmod -R 440 ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/* >/dev/null 2>&1 -fi -if [ -f ${targetdir}/ascend_install.info ]; then - chmod -R 440 ${targetdir}/ascend_install.info -fi -if [ -f ${targetdir}/scene.info ]; then - chmod -R 440 ${targetdir}/scene.info -fi -if [ -f ${targetdir}/version.info ]; then - chmod -R 440 ${targetdir}/version.info -fi - -echo "SUCCESS" -exit 0 - diff --git a/atb_operator/AtbEyeSample/EyeOp/scripts/upgrade.sh b/atb_operator/AtbEyeSample/EyeOp/scripts/upgrade.sh deleted file mode 100755 index e09173485..000000000 --- a/atb_operator/AtbEyeSample/EyeOp/scripts/upgrade.sh +++ /dev/null @@ -1,151 +0,0 @@ -#!/bin/bash -vendor_name=customize -targetdir=/usr/local/Ascend/opp -target_custom=0 - -sourcedir=$PWD/packages -vendordir=vendors/$vendor_name - -log() { - cur_date=`date +"%Y-%m-%d %H:%M:%S"` - echo "[runtime] [$cur_date] "$1 -} - -if [[ "x${ASCEND_OPP_PATH}" == "x" ]];then - log "[ERROR] env ASCEND_OPP_PATH no exist" - exit 1 -fi - -targetdir=${ASCEND_OPP_PATH} - -if [ ! -d $targetdir ];then - log "[ERROR] $targetdir no exist" - exit 1 -fi - -upgrade() -{ - if [ ! -d ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 files" - return 0 - fi - - if [ ! -d ${targetdir}/$vendordir/$1 ];then - log "[INFO] create ${targetdir}/$vendordir/$1." - mkdir -p ${targetdir}/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/$1 failed" - return 1 - fi - else - vendor_installed_dir=$(ls "$targetdir/vendors" 2> /dev/null) - for i in $vendor_installed_dir;do - vendor_installed_file=$(ls "$vendor_installed_dir/$vendor_name/$i" 2> /dev/null) - if [ "$i" = "$vendor_name" ] && [ "$vendor_installed_file" != "" ]; then - echo "[INFO]: $vendor_name custom opp package has been installed on the path $vendor_installed_dir, \ - you want to Overlay Installation , please enter:[o]; \ - or replace directory installation , please enter: [r]; \ - or not install , please enter:[n]." - fi - while true - do - read mrn - if [ "$mrn" = m ]; then - break - elif [ "$mrn" = r ]; then - [ -n "$vendor_installed_file"] && rm -rf "$vendor_installed_file" - break - elif [ "$mrn" = n ]; then - return 0 - else - echo "[WARNING]: Input error, please input m or r or n to choose!" - fi - done - done - log "[INFO] replace old ops $1 files ......" - fi - - log "copy new ops $1 files ......" - cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 files failed" - return 1 - fi - - return 0 -} - -upgrade_file() -{ - if [ ! -e ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 file" - return 0 - fi - - log "copy new $1 files ......" - cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 file failed" - return 1 - fi - - return 0 -} - -log "[INFO] copy uninstall sh success" - -echo "[ops_custom]upgrade framework" -upgrade framework -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op proto" -upgrade op_proto -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op impl" -upgrade op_impl -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op api" -upgrade op_api -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade version.info" -upgrade_file version.info -if [ $? -ne 0 ];then - exit 1 -fi - -config_file=${targetdir}/vendors/config.ini -found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" -found_vendor=$(echo $found_vendors | sed "s/$vendor_name//g" | tr ',' ' ') -vendor=$(echo $found_vendor | tr -s ' ' ',') -if [ "$vendor" != "" ]; then - sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" -fi - -changemode() -{ - if [ -d ${targetdir} ];then - chmod -R 550 ${targetdir}>/dev/null 2>&1 - fi - - return 0 -} -echo "[ops_custom]changemode..." -#changemode -if [ $? -ne 0 ];then - exit 1 -fi - -echo "SUCCESS" -exit 0 - diff --git a/atb_operator/AtbTrilSample/TrilOp/CMakeLists.txt b/atb_operator/AtbTrilSample/TrilOp/CMakeLists.txt deleted file mode 100755 index 584132d80..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/CMakeLists.txt +++ /dev/null @@ -1,69 +0,0 @@ -cmake_minimum_required(VERSION 3.16.0) -project(opp) -if(ENABLE_CROSS_COMPILE) - if(${CMAKE_SYSTEM_PROCESSOR} STREQUAL x86_64) - set(CROSS_COMPILE_PLATFORM aarch64) - else() - set(CROSS_COMPILE_PLATFORM x86_64) - endif() - set(PLATFORM ${CMAKE_SYSTEM_PROCESSOR}) - set(CMAKE_COMPILE_COMPILER_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/linux/${CROSS_COMPILE_PLATFORM}/) - set(CMAKE_COMPILE_RUNTIME_LIBRARY ${ASCEND_CANN_PACKAGE_PATH}/${PLATFORM}-linux/devlib/${CROSS_COMPILE_PLATFORM}/) - set(CMAKE_SYSTEM_PROCESSOR ${CROSS_COMPILE_PLATFORM}) - set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) - set(CMAKE_CXX_COMPILER ${CMAKE_CROSS_PLATFORM_COMPILER}) -else() - set(CMAKE_COMPILE ${CMAKE_CXX_COMPILER}) -endif() - -include(cmake/config.cmake) -include(cmake/func.cmake) -include(cmake/intf.cmake) - -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/framework) - add_subdirectory(framework) -endif() -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_host) - add_subdirectory(op_host) -endif() -if(EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/op_kernel) - add_subdirectory(op_kernel) -endif() -if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) - add_subdirectory(testcases) -endif() - -# modify vendor_name in install.sh and upgrade.sh -add_custom_command(OUTPUT ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh - COMMAND mkdir -p ${CMAKE_BINARY_DIR}/scripts - COMMAND cp -r ${CMAKE_SOURCE_DIR}/scripts/* ${CMAKE_BINARY_DIR}/scripts/ - COMMAND sed -i "s/vendor_name=customize/vendor_name=${vendor_name}/g" ${CMAKE_BINARY_DIR}/scripts/* -) -add_custom_target(modify_vendor ALL DEPENDS ${CMAKE_BINARY_DIR}/scripts/install.sh ${CMAKE_BINARY_DIR}/scripts/upgrade.sh) -install(DIRECTORY ${CMAKE_BINARY_DIR}/scripts/ DESTINATION . FILE_PERMISSIONS OWNER_EXECUTE OWNER_READ GROUP_READ) - -install(FILES ${CMAKE_SOURCE_DIR}/custom.proto DESTINATION packages OPTIONAL) - -get_system_info(SYSTEM_INFO) - -# gen version.info -add_custom_target(gen_version_info ALL - COMMAND bash ${CMAKE_CURRENT_SOURCE_DIR}/cmake/util/gen_version_info.sh ${ASCEND_CANN_PACKAGE_PATH} ${CMAKE_CURRENT_BINARY_DIR} -) - -install(FILES ${CMAKE_CURRENT_BINARY_DIR}/version.info - DESTINATION packages/vendors/${vendor_name}/) - -# CPack config -set(CPACK_PACKAGE_NAME ${CMAKE_PROJECT_NAME}) -set(CPACK_PACKAGE_VERSION ${CMAKE_PROJECT_VERSION}) -set(CPACK_PACKAGE_DESCRIPTION "CPack opp project") -set(CPACK_PACKAGE_DESCRIPTION_SUMMARY "CPack opp project") -set(CPACK_PACKAGE_DIRECTORY ${CMAKE_INSTALL_PREFIX}) -set(CPACK_PACKAGE_FILE_NAME "custom_opp_${SYSTEM_INFO}.run") -set(CPACK_GENERATOR External) -set(CPACK_CMAKE_GENERATOR "Unix Makefiles") -set(CPACK_EXTERNAL_ENABLE_STAGING TRUE) -set(CPACK_EXTERNAL_PACKAGE_SCRIPT ${CMAKE_SOURCE_DIR}/cmake/makeself.cmake) -set(CPACK_EXTERNAL_BUILT_PACKAGES ${CPACK_PACKAGE_DIRECTORY}/_CPack_Packages/Linux/External/${CPACK_PACKAGE_FILE_NAME}/${CPACK_PACKAGE_FILE_NAME}) -include(CPack) diff --git a/atb_operator/AtbTrilSample/TrilOp/CMakePresets.json b/atb_operator/AtbTrilSample/TrilOp/CMakePresets.json deleted file mode 100755 index d32bf6d9a..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/CMakePresets.json +++ /dev/null @@ -1,63 +0,0 @@ -{ - "version": 1, - "cmakeMinimumRequired": { - "major": 3, - "minor": 19, - "patch": 0 - }, - "configurePresets": [ - { - "name": "default", - "displayName": "Default Config", - "description": "Default build using Unix Makefiles generator", - "generator": "Unix Makefiles", - "binaryDir": "${sourceDir}/build_out", - "cacheVariables": { - "CMAKE_BUILD_TYPE": { - "type": "STRING", - "value": "Release" - }, - "ENABLE_SOURCE_PACKAGE": { - "type": "BOOL", - "value": "True" - }, - "ENABLE_BINARY_PACKAGE": { - "type": "BOOL", - "value": "True" - }, - "ASCEND_COMPUTE_UNIT": { - "type": "STRING", - "value": "ascend310b;ascend910b" - }, - "ENABLE_TEST": { - "type": "BOOL", - "value": "True" - }, - "vendor_name": { - "type": "STRING", - "value": "customize" - }, - "ASCEND_CANN_PACKAGE_PATH": { - "type": "PATH", - "value": "/usr/local/Ascend/ascend-toolkit/latest" - }, - "ASCEND_PYTHON_EXECUTABLE": { - "type": "STRING", - "value": "python3" - }, - "CMAKE_INSTALL_PREFIX": { - "type": "PATH", - "value": "${sourceDir}/build_out" - }, - "ENABLE_CROSS_COMPILE": { - "type": "BOOL", - "value": "False" - }, - "CMAKE_CROSS_PLATFORM_COMPILER": { - "type": "PATH", - "value": "/usr/bin/aarch64-linux-gnu-g++" - } - } - } - ] -} \ No newline at end of file diff --git a/atb_operator/AtbTrilSample/TrilOp/README.md b/atb_operator/AtbTrilSample/TrilOp/README.md deleted file mode 100644 index a8addc946..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/README.md +++ /dev/null @@ -1,193 +0,0 @@ -# 概述 -本样例基于TrilOp算子工程,介绍了单算子工程及aclnn接口说明。 - -## 算子工程介绍 -### 算子描述 -Tril算子是PyTorch中的一种常见矩阵构造函数。Tril函数默认情况下返回一个矩阵主对角线以下的下三角矩阵,其它元素全部为0。主对角线的偏移由可选参数diagonal决定,其缺省值为0。diagonal为正值时,主对角线向上偏移。当输入是一个多维张量时,其最后两个维度构成矩阵,Tril以迭代的方式处理多维张量中的每个矩阵,最终返回对应的下三角矩阵构成的多维张量。返回的多维张量与输入张量维度保持一致。 - -### 算子规格描述 - - - - - - - -
算子类型(OpType)Tril
算子输入nameshapedata typeformat默认值
x-float32, float16ND\
算子输出y-float32, float16ND\
attr属性diagonal\int\0
核函数名tril
- - - -### 算子工程文件介绍 -其中,算子工程目录TrilOp包含算子实现的模板文件、编译脚本等,如下所示: -``` -├── TrilOp //Add自定义算子工程 -│ ├── cmake -│ ├── framework // 算子插件实现文件目录,单算子模型文件的生成不依赖算子适配插件,无需关注 -│ ├── op_host // host侧实现文件 -│ ├── op_kernel // kernel侧实现文件 -│ ├── scripts // 自定义算子工程打包相关脚本所在目录 -│ ├── build.sh // 编译入口脚本 -│ ├── CMakeLists.txt // 算子工程的CMakeLists.txt -│ └── CMakePresets.json // 编译配置项 -``` - -### 编译运行样例算子 -针对自定义算子工程,编译运行包含如下步骤: -- 编译自定义算子工程生成算子安装包; -- 安装自定义算子到算子库中; -- 调用执行自定义算子; - -详细操作如下所示。 -#### 1. 获取源码包 -编译运行此样例前,请参考[准备:获取样例代码](../README.md#codeready)获取源码包。 -#### 2. 编译算子工程 - 编译自定义算子工程,构建生成自定义算子包。 - - - 执行如下命令,切换到算子工程TrilOp目录。 - - ```bash - cd ${git_clone_path}/samples/atb_operator/AtbTrilSample/TrilOp - ``` - - - 修改CMakePresets.json中ASCEND_CANN_PACKAGE_PATH为CANN软件包安装后的实际路径。 - - - ```json - { - …… - "configurePresets": [ - { - …… - "ASCEND_CANN_PACKAGE_PATH": { - "type": "PATH", - "value": "/usr/local/Ascend/ascend-toolkit/latest" //请替换为CANN软件包安装后的实际路径。eg:/home/HwHiAiUser/Ascend/ascend-toolkit/latest - }, - …… - } - ] - } - ``` - - 在算子工程TrilOp目录下执行如下命令,进行算子工程编译。 - - ```bash - ./build.sh - ``` -编译成功后,会在当前目录下创建build_out目录,并在build_out目录下生成自定义算子安装包custom_opp_\_\.run,例如“custom_opp_ubuntu_x86_64.run”。 - -备注:如果要使用dump调试功能,需要移除op_host内和CMakeLists.txt内的Atlas 200/500 A2 推理产品的配置。 - -#### 3. 部署算子包 - -执行如下命令,在自定义算子安装包所在路径下,安装自定义算子包。 - ```bash - cd build_out - ./custom_opp__.run - ``` -命令执行成功后,自定义算子包中的相关文件将部署至当前环境的OPP算子库的vendors/customize目录中。 - -#### 4. 配置环境变量 - - 请根据当前环境上CANN开发套件包的[安装方式](https://hiascend.com/document/redirect/CannCommunityInstSoftware),选择对应配置环境变量的命令。 - - 默认路径,root用户安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=/usr/local/Ascend/ascend-toolkit/latest - ``` - - 默认路径,非root用户安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=$HOME/Ascend/ascend-toolkit/latest - ``` - - 指定路径install_path,安装CANN软件包 - ```bash - export ASCEND_INSTALL_PATH=${install_path}/ascend-toolkit/latest - ``` - -## aclnn接口说明 - -自定义算子编译部署后,会自动生成单算子API,可以直接在应用程序中调用。算子API的形式一般定义为“两段式接口”,形如: - ```cpp - aclnnStatus aclnnTrilGetWorkspaceSize(const aclTensor *x, int64_t diagonal, const aclTensor *out, uint64_t *workspaceSize,aclOpExecutor **executor); - aclnnStatus aclnnTril(void *workspace, int64_t workspaceSize, aclOpExecutor **executor, aclrtStream stream); - ``` -其中aclnnTrilGetWorkspaceSize为第一段接口,主要用于计算本次API调用计算过程中需要多少的workspace内存。获取到本次API计算需要的workspace大小之后,按照workspaceSize大小申请Device侧内存,然后调用第二段接口aclnnTriu执行计算。具体参考[AscendCL单算子调用](https://hiascend.com/document/redirect/CannCommunityAscendCInVorkSingleOp)>单算子API执行 章节。 - -### 功能描述 -* 算子功能:完成加法计算。 -* 计算公式: -``` -z = x + y -``` - -#### 参数说明 -##### aclnnTrilGetWorkspaceSize: - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
接口参数描述
aclnnTrilGetWorkspaceSizex输入x的Tensor,支持flaot/half类型,ND排布格式
diagonal可选属性,对角线的位置,支持int64类型数值,默认值为0
out输出z的Tensor,支持flaot/half类型,ND排布格式
workspaceSize返回需要在Device侧申请的workspace大小
executor返回op执行器,包含了算子计算流程
- - -| 返回值 | 描述 | -| ---- | ---- | -| ACL_SUCCESS | 成功 | -| ACL_ERROR | 失败 | - -##### aclnnTril - - - - - - - - - - - - - - - - - - - - - - - -
接口参数描述
aclnnTrilworkspace输入Device侧申请的workspace内存
workspaceSizeworkspace大小
executor输入op执行器,包含了算子计算流程
stream输入执行流
- -| 返回值 | 描述 | -| ---- | ---- | -| ACL_SUCCESS | 成功 | -| ACL_ERROR | 失败 | - - -## 更新说明 -| 时间 | 更新事项 | -|----|------| -| 2025/03/27 | 新增README | \ No newline at end of file diff --git a/atb_operator/AtbTrilSample/TrilOp/build.sh b/atb_operator/AtbTrilSample/TrilOp/build.sh deleted file mode 100755 index decae7c43..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/build.sh +++ /dev/null @@ -1,74 +0,0 @@ -#!/bin/bash -script_path=$(realpath $(dirname $0)) - - -mkdir -p build_out -rm -rf build_out/* -cd build_out - -opts=$(python3 $script_path/cmake/util/preset_parse.py $script_path/CMakePresets.json) -ENABLE_CROSS="-DENABLE_CROSS_COMPILE=True" -ENABLE_BINARY="-DENABLE_BINARY_PACKAGE=True" -cmake_version=$(cmake --version | grep "cmake version" | awk '{print $3}') - -cmake_run_package() -{ - target=$1 - cmake --build . --target $target -j16 - if [ $? -ne 0 ]; then exit 1; fi - - if [ $target = "package" ]; then - if test -d ./op_kernel/binary ; then - ./cust*.run - if [ $? -ne 0 ]; then exit 1; fi - cmake --build . --target binary -j16 - if [ $? -ne 0 ]; then exit 1; fi - cmake --build . --target $target -j16 - fi - fi -} - -if [[ $opts =~ $ENABLE_CROSS ]] && [[ $opts =~ $ENABLE_BINARY ]] -then - target=package - if [ "$1"x != ""x ]; then target=$1; fi - if [ "$cmake_version" \< "3.19.0" ] ; then - cmake .. $opts -DENABLE_CROSS_COMPILE=0 - else - cmake .. --preset=default -DENABLE_CROSS_COMPILE=0 - fi - cmake_run_package $target - cp -r kernel ../ - rm -rf * - if [ "$cmake_version" \< "3.19.0" ] ; then - cmake .. $opts - else - cmake .. --preset=default - fi - - cmake --build . --target $target -j16 - if [ $? -ne 0 ]; then exit 1; fi - if [ $target = "package" ]; then - if test -d ./op_kernel/binary ; then - ./cust*.run - fi - fi - rm -rf ../kernel - -else - target=package - if [ "$1"x != ""x ]; then target=$1; fi - if [ "$cmake_version" \< "3.19.0" ] ; then - cmake .. $opts - else - cmake .. --preset=default - fi - cmake_run_package $target -fi - - -# for debug -# cd build_out -# make -# cpack -# verbose append -v \ No newline at end of file diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/config.cmake b/atb_operator/AtbTrilSample/TrilOp/cmake/config.cmake deleted file mode 100755 index 886119daa..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/config.cmake +++ /dev/null @@ -1,25 +0,0 @@ - -set(CMAKE_CXX_FLAGS_DEBUG "") -set(CMAKE_CXX_FLAGS_RELEASE "") - -if (NOT DEFINED vendor_name) - set(vendor_name customize CACHE STRING "") -endif() -if (NOT DEFINED ASCEND_CANN_PACKAGE_PATH) - set(ASCEND_CANN_PACKAGE_PATH /usr/local/Ascend/latest CACHE PATH "") -endif() -if (NOT DEFINED ASCEND_PYTHON_EXECUTABLE) - set(ASCEND_PYTHON_EXECUTABLE python3 CACHE STRING "") -endif() -if (NOT DEFINED ASCEND_COMPUTE_UNIT) - message(FATAL_ERROR "ASCEND_COMPUTE_UNIT not set in CMakePreset.json ! -") -endif() -set(ASCEND_TENSOR_COMPILER_PATH ${ASCEND_CANN_PACKAGE_PATH}/compiler) -set(ASCEND_CCEC_COMPILER_PATH ${ASCEND_TENSOR_COMPILER_PATH}/ccec_compiler/bin) -set(ASCEND_AUTOGEN_PATH ${CMAKE_BINARY_DIR}/autogen) -set(ASCEND_FRAMEWORK_TYPE tensorflow) -file(MAKE_DIRECTORY ${ASCEND_AUTOGEN_PATH}) -set(CUSTOM_COMPILE_OPTIONS "custom_compile_options.ini") -execute_process(COMMAND rm -rf ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} - COMMAND touch ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS}) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/func.cmake b/atb_operator/AtbTrilSample/TrilOp/cmake/func.cmake deleted file mode 100755 index 4179dfd25..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/func.cmake +++ /dev/null @@ -1,192 +0,0 @@ - -function(get_system_info SYSTEM_INFO) - if (UNIX) - execute_process(COMMAND grep -i ^id= /etc/os-release OUTPUT_VARIABLE TEMP) - string(REGEX REPLACE "\n|id=|ID=|\"" "" SYSTEM_NAME ${TEMP}) - set(${SYSTEM_INFO} ${SYSTEM_NAME}_${CMAKE_SYSTEM_PROCESSOR} PARENT_SCOPE) - elseif (WIN32) - message(STATUS "System is Windows. Only for pre-build.") - else () - message(FATAL_ERROR "${CMAKE_SYSTEM_NAME} not support.") - endif () -endfunction() - -function(opbuild) - message(STATUS "Opbuild generating sources") - cmake_parse_arguments(OPBUILD "" "OUT_DIR;PROJECT_NAME;ACCESS_PREFIX" "OPS_SRC" ${ARGN}) - execute_process(COMMAND ${CMAKE_COMPILE} -g -fPIC -shared -std=c++11 ${OPBUILD_OPS_SRC} -D_GLIBCXX_USE_CXX11_ABI=0 - -I ${ASCEND_CANN_PACKAGE_PATH}/include -L ${ASCEND_CANN_PACKAGE_PATH}/lib64 -lexe_graph -lregister -ltiling_api - -o ${OPBUILD_OUT_DIR}/libascend_all_ops.so - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("build ops lib info: ${EXEC_INFO}") - message("build ops lib error: ${EXEC_ERROR}") - message(FATAL_ERROR "opbuild run failed!") - endif() - set(proj_env "") - set(prefix_env "") - if (NOT "${OPBUILD_PROJECT_NAME}x" STREQUAL "x") - set(proj_env "OPS_PROJECT_NAME=${OPBUILD_PROJECT_NAME}") - endif() - if (NOT "${OPBUILD_ACCESS_PREFIX}x" STREQUAL "x") - set(prefix_env "OPS_DIRECT_ACCESS_PREFIX=${OPBUILD_ACCESS_PREFIX}") - endif() - execute_process(COMMAND ${proj_env} ${prefix_env} ${ASCEND_CANN_PACKAGE_PATH}/toolkit/tools/opbuild/op_build - ${OPBUILD_OUT_DIR}/libascend_all_ops.so ${OPBUILD_OUT_DIR} - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("opbuild ops info: ${EXEC_INFO}") - message("opbuild ops error: ${EXEC_ERROR}") - endif() - message(STATUS "Opbuild generating sources - done") -endfunction() - -function(add_ops_info_target) - cmake_parse_arguments(OPINFO "" "TARGET;OPS_INFO;OUTPUT;INSTALL_DIR" "" ${ARGN}) - get_filename_component(opinfo_file_path "${OPINFO_OUTPUT}" DIRECTORY) - add_custom_command(OUTPUT ${OPINFO_OUTPUT} - COMMAND mkdir -p ${opinfo_file_path} - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/parse_ini_to_json.py - ${OPINFO_OPS_INFO} ${OPINFO_OUTPUT} - ) - add_custom_target(${OPINFO_TARGET} ALL - DEPENDS ${OPINFO_OUTPUT} - ) - install(FILES ${OPINFO_OUTPUT} - DESTINATION ${OPINFO_INSTALL_DIR} - ) -endfunction() - -function(add_ops_compile_options OP_TYPE) - cmake_parse_arguments(OP_COMPILE "" "OP_TYPE" "COMPUTE_UNIT;OPTIONS" ${ARGN}) - file(APPEND ${ASCEND_AUTOGEN_PATH}/${CUSTOM_COMPILE_OPTIONS} - "${OP_TYPE},${OP_COMPILE_COMPUTE_UNIT},${OP_COMPILE_OPTIONS}\n") -endfunction() - -function(add_ops_impl_target) - cmake_parse_arguments(OPIMPL "" "TARGET;OPS_INFO;IMPL_DIR;OUT_DIR;INSTALL_DIR" "OPS_BATCH;OPS_ITERATE" ${ARGN}) - add_custom_command(OUTPUT ${OPIMPL_OUT_DIR}/.impl_timestamp - COMMAND mkdir -m 700 -p ${OPIMPL_OUT_DIR}/dynamic - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py - ${OPIMPL_OPS_INFO} - \"${OPIMPL_OPS_BATCH}\" \"${OPIMPL_OPS_ITERATE}\" - ${OPIMPL_IMPL_DIR} - ${OPIMPL_OUT_DIR}/dynamic - ${ASCEND_AUTOGEN_PATH} - - COMMAND rm -rf ${OPIMPL_OUT_DIR}/.impl_timestamp - COMMAND touch ${OPIMPL_OUT_DIR}/.impl_timestamp - DEPENDS ${OPIMPL_OPS_INFO} - ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_impl_build.py - ) - add_custom_target(${OPIMPL_TARGET} ALL - DEPENDS ${OPIMPL_OUT_DIR}/.impl_timestamp) - if (${ENABLE_SOURCE_PACKAGE}) - install(DIRECTORY ${OPIMPL_OUT_DIR}/dynamic - DESTINATION ${OPIMPL_INSTALL_DIR} - ) - endif() -endfunction() - -function(add_npu_support_target) - cmake_parse_arguments(NPUSUP "" "TARGET;OPS_INFO_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) - get_filename_component(npu_sup_file_path "${NPUSUP_OUT_DIR}" DIRECTORY) - add_custom_command(OUTPUT ${NPUSUP_OUT_DIR}/npu_supported_ops.json - COMMAND mkdir -p ${NPUSUP_OUT_DIR} - COMMAND ${CMAKE_SOURCE_DIR}/cmake/util/gen_ops_filter.sh - ${NPUSUP_OPS_INFO_DIR} - ${NPUSUP_OUT_DIR} - ) - add_custom_target(npu_supported_ops ALL - DEPENDS ${NPUSUP_OUT_DIR}/npu_supported_ops.json - ) - install(FILES ${NPUSUP_OUT_DIR}/npu_supported_ops.json - DESTINATION ${NPUSUP_INSTALL_DIR} - ) -endfunction() - -function(add_bin_compile_target) - cmake_parse_arguments(BINCMP "" "TARGET;OPS_INFO;COMPUTE_UNIT;IMPL_DIR;ADP_DIR;OUT_DIR;INSTALL_DIR" "" ${ARGN}) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/src) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/gen) - execute_process(COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_bin_param_build.py - ${BINCMP_OPS_INFO} ${BINCMP_OUT_DIR}/gen ${BINCMP_COMPUTE_UNIT} - RESULT_VARIABLE EXEC_RESULT - OUTPUT_VARIABLE EXEC_INFO - ERROR_VARIABLE EXEC_ERROR - ) - if (${EXEC_RESULT}) - message("ops binary compile scripts gen info: ${EXEC_INFO}") - message("ops binary compile scripts gen error: ${EXEC_ERROR}") - message(FATAL_ERROR "ops binary compile scripts gen failed!") - endif() - if (NOT TARGET binary) - add_custom_target(binary) - endif() - add_custom_target(${BINCMP_TARGET} - COMMAND cp -r ${BINCMP_IMPL_DIR}/*.* ${BINCMP_OUT_DIR}/src - ) - add_custom_target(${BINCMP_TARGET}_gen_ops_config - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/insert_simplified_keys.py -p ${BINCMP_OUT_DIR}/bin - COMMAND ${ASCEND_PYTHON_EXECUTABLE} ${CMAKE_SOURCE_DIR}/cmake/util/ascendc_ops_config.py -p ${BINCMP_OUT_DIR}/bin - -s ${BINCMP_COMPUTE_UNIT} - ) - add_dependencies(binary ${BINCMP_TARGET}_gen_ops_config) - file(GLOB bin_scripts ${BINCMP_OUT_DIR}/gen/*.sh) - foreach(bin_script ${bin_scripts}) - get_filename_component(bin_file ${bin_script} NAME_WE) - string(REPLACE "-" ";" bin_sep ${bin_file}) - list(GET bin_sep 0 op_type) - list(GET bin_sep 1 op_file) - list(GET bin_sep 2 op_index) - if (NOT TARGET ${BINCMP_TARGET}_${op_file}_copy) - file(MAKE_DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file}) - add_custom_target(${BINCMP_TARGET}_${op_file}_copy - COMMAND cp ${BINCMP_ADP_DIR}/${op_file}.py ${BINCMP_OUT_DIR}/src/${op_type}.py - ) - install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} - DESTINATION ${BINCMP_INSTALL_DIR}/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) - install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json - DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT}/ OPTIONAL - ) - endif() - add_custom_target(${BINCMP_TARGET}_${op_file}_${op_index} - COMMAND export HI_PYTHON=${ASCEND_PYTHON_EXECUTABLE} && bash ${bin_script} ${BINCMP_OUT_DIR}/src/${op_type}.py ${BINCMP_OUT_DIR}/bin/${op_file} && echo $(MAKE) - WORKING_DIRECTORY ${BINCMP_OUT_DIR} - ) - add_dependencies(${BINCMP_TARGET}_${op_file}_${op_index} ${BINCMP_TARGET} ${BINCMP_TARGET}_${op_file}_copy) - add_dependencies(${BINCMP_TARGET}_gen_ops_config ${BINCMP_TARGET}_${op_file}_${op_index}) - endforeach() - install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json - DESTINATION ${BINCMP_INSTALL_DIR}/config/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) - - install(DIRECTORY ${BINCMP_OUT_DIR}/bin/${op_file} - DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/../build_out/kernel/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) - install(FILES ${BINCMP_OUT_DIR}/bin/binary_info_config.json - DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/../build_out/kernel/config/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) - install(FILES ${BINCMP_OUT_DIR}/bin/${op_file}.json - DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/../build_out/kernel/config/${BINCMP_COMPUTE_UNIT} OPTIONAL - ) - -endfunction() - -function(add_cross_compile_target) - cmake_parse_arguments(CROSSMP "" "TARGET;OUT_DIR;INSTALL_DIR" "" ${ARGN}) - add_custom_target(${CROSSMP_TARGET} ALL - DEPENDS ${CROSSMP_OUT_DIR} - ) - install(DIRECTORY ${CROSSMP_OUT_DIR} - DESTINATION ${CROSSMP_INSTALL_DIR} - ) -endfunction() diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/intf.cmake b/atb_operator/AtbTrilSample/TrilOp/cmake/intf.cmake deleted file mode 100755 index 2f362c396..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/intf.cmake +++ /dev/null @@ -1,26 +0,0 @@ - -add_library(intf_pub INTERFACE) -target_compile_options(intf_pub INTERFACE - -fPIC - -fvisibility=hidden - -fvisibility-inlines-hidden - $<$:-O2> - $<$:-O0 -g> - $<$:-std=c++11> - $<$,$>:-ftrapv -fstack-check> - $<$:-pthread -Wfloat-equal -Wshadow -Wformat=2 -Wno-deprecated -Wextra> - $,-fstack-protector-strong,-fstack-protector-all> -) -target_compile_definitions(intf_pub INTERFACE - _GLIBCXX_USE_CXX11_ABI=0 - $<$:_FORTIFY_SOURCE=2> -) -target_include_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/include) -target_link_options(intf_pub INTERFACE - $<$,EXECUTABLE>:-pie> - $<$:-s> - -Wl,-z,relro - -Wl,-z,now - -Wl,-z,noexecstack -) -target_link_directories(intf_pub INTERFACE ${ASCEND_CANN_PACKAGE_PATH}/lib64) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/makeself.cmake b/atb_operator/AtbTrilSample/TrilOp/cmake/makeself.cmake deleted file mode 100755 index 48c565bfb..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/makeself.cmake +++ /dev/null @@ -1,17 +0,0 @@ -execute_process(COMMAND chmod +x ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh) -execute_process(COMMAND ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself.sh - --header ${CMAKE_CURRENT_LIST_DIR}/util/makeself/makeself-header.sh - --help-header ./help.info - --gzip --complevel 4 --nomd5 --sha256 - ./ ${CPACK_PACKAGE_FILE_NAME} "version:1.0" ./install.sh - WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} - RESULT_VARIABLE EXEC_RESULT - ERROR_VARIABLE EXEC_ERROR -) -if (NOT "${EXEC_RESULT}x" STREQUAL "0x") - message(FATAL_ERROR "CPack Command error: ${EXEC_RESULT}\n${EXEC_ERROR}") -endif() -execute_process(COMMAND cp ${CPACK_EXTERNAL_BUILT_PACKAGES} ${CPACK_PACKAGE_DIRECTORY}/ - COMMAND echo "Copy ${CPACK_EXTERNAL_BUILT_PACKAGES} to ${CPACK_PACKAGE_DIRECTORY}/" - WORKING_DIRECTORY ${CPACK_TEMPORARY_DIRECTORY} -) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/__init__.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/__init__.py deleted file mode 100755 index c4ddc893a..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/__init__.py +++ /dev/null @@ -1,8 +0,0 @@ -#!/usr/bin/env python -# -*- coding: utf-8 -*- - -import sys -import os - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) -sys.path.append(PYF_PATH) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_bin_param_build.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_bin_param_build.py deleted file mode 100755 index 63b15c900..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_bin_param_build.py +++ /dev/null @@ -1,127 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import json -import hashlib -import const_var -import opdesc_parser - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - - -class BinParamBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - super().__init__(op_type) - self.soc = '' - self.out_path = '' - - def set_soc_version(self: any, soc: str): - self.soc = soc - - def set_out_path(self: any, out_path: str): - self.out_path = out_path - - def gen_input_json(self: any): - key_map = {} - count = len(self.input_dtype[0].split(',')) - for i in range(0, count): - inputs = [] - outputs = [] - attrs = [] - op_node = {} - for idx in range(0, len(self.input_name)): - idtypes = self.input_dtype[idx].split(',') - ifmts = self.input_fmt[idx].split(',') - itype = self.input_type[idx] - para = {} - para['name'] = self.input_name[idx][:-5] - para['index'] = idx - para['dtype'] = idtypes[i] - para['format'] = ifmts[i] - para['paramType'] = itype - para['shape'] = [-2] - if itype == 'dynamic': - inputs.append([para]) - else: - inputs.append(para) - for idx in range(0, len(self.output_name)): - odtypes = self.output_dtype[idx].split(',') - ofmts = self.output_fmt[idx].split(',') - otype = self.output_type[idx] - para = {} - para['name'] = self.output_name[idx][:-5] - para['index'] = idx - para['dtype'] = odtypes[i] - para['format'] = ofmts[i] - para['paramType'] = otype - para['shape'] = [-2] - if otype == 'dynamic': - outputs.append([para]) - else: - outputs.append(para) - for attr in self.attr_list: - att = {} - att['name'] = attr - atype = self.attr_val.get(attr).get('type').lower() - att['dtype'] = atype - att['value'] = const_var.ATTR_DEF_VAL.get(atype) - attrs.append(att) - op_node['bin_filename'] = '' - op_node['inputs'] = inputs - op_node['outputs'] = outputs - if len(attrs) > 0: - op_node['attrs'] = attrs - param = {} - param['op_type'] = self.op_type - param['op_list'] = [op_node] - objstr = json.dumps(param, indent=' ') - md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() - while key_map.get(md5sum) is not None: - objstr += '1' - md5sum = hashlib.md5(objstr.encode('utf-8')).hexdigest() - key_map[md5sum] = md5sum - bin_file = self.op_type + '_' + md5sum - op_node['bin_filename'] = bin_file - param_file = os.path.join(self.out_path, bin_file + '_param.json') - param_file = os.path.realpath(param_file) - with os.fdopen(os.open(param_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - json.dump(param, fd, indent=' ') - self._write_buld_cmd(param_file, bin_file, i) - - def _write_buld_cmd(self: any, param_file: str, bin_file: str, index: int): - hard_soc = const_var.SOC_MAP_EXT.get(self.soc) - if not hard_soc: - hard_soc = self.soc.capitalize() - name_com = [self.op_type, self.op_file, str(index)] - compile_file = os.path.join(self.out_path, '-'.join(name_com) + '.sh') - compile_file = os.path.realpath(compile_file) - with os.fdopen(os.open(compile_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - fd.write('#!/bin/bash\n') - fd.write('echo "[{}] Generating {} ..."\n'.format(hard_soc, bin_file)) - cmd = const_var.BIN_CMD.format(fun=self.op_intf, soc=hard_soc, param=param_file, impl='""') - fd.write(cmd) - chk = const_var.CHK_CMD.format(res_file=bin_file + '.json') - fd.write(chk) - chk = const_var.CHK_CMD.format(res_file=bin_file + '.o') - fd.write(chk) - fd.write('echo "[{}] Generating {} Done"\n'.format(hard_soc, bin_file)) - - -def gen_bin_param_file(cfgfile: str, out_dir: str, soc: str): - op_descs = opdesc_parser.get_op_desc(cfgfile, [], [], BinParamBuilder, None) - for op_desc in op_descs: - op_desc.set_soc_version(soc) - op_desc.set_out_path(out_dir) - op_desc.gen_input_json() - - -if __name__ == '__main__': - if len(sys.argv) <= 3: - raise RuntimeError('arguments must greater than 3') - gen_bin_param_file(sys.argv[1], sys.argv[2], sys.argv[3]) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_impl_build.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_impl_build.py deleted file mode 100755 index e8cf2b9ab..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_impl_build.py +++ /dev/null @@ -1,496 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import re -import stat -import opdesc_parser -import const_var - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - -IMPL_HEAD = ''' -import os, sys -import ctypes -import json -import shutil -from tbe.common.platform import get_soc_spec -from tbe.common.utils import para_check -from tbe.tikcpp import compile_op, replay_op, check_op_cap, generalize_op_params, get_code_channel, OpInfo -from tbe.common.buildcfg import get_default_build_config -from impl.util.platform_adapter import tbe_register -from tbe.common.buildcfg import get_current_build_config -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - -DTYPE_MAP = {"float32": ["DT_FLOAT", "float"], - "float16": ["DT_FLOAT16", "half"], - "int8": ["DT_INT8", "int8_t"], - "int16": ["DT_INT16", "int16_t"], - "int32": ["DT_INT32", "int32_t"], - "int64": ["DT_INT64", "int64_t"], - "uint1": ["DT_UINT1", "uint8_t"], - "uint8": ["DT_UINT8", "uint8_t"], - "uint16": ["DT_UINT16", "uint16_t"], - "uint32": ["DT_UINT32", "uint32_t"], - "uint64": ["DT_UINT64", "uint64_t"], - "bool": ["DT_BOOL", "bool"], - "double": ["DT_DOUBLE", "double"], - "dual": ["DT_DUAL", "unknown"], - "dual_sub_int8": ["DT_DUAL_SUB_INT8", "unknown"], - "dual_sub_uint8": ["DT_DUAL_SUB_UINT8", "unknown"], - "string": ["DT_STRING", "unknown"], - "complex64": ["DT_COMPLEX64", "unknown"], - "complex128": ["DT_COMPLEX128", "unknown"], - "qint8": ["DT_QINT8", "unknown"], - "qint16": ["DT_QINT16", "unknown"], - "qint32": ["DT_QINT32", "unknown"], - "quint8": ["DT_QUINT8", "unknown"], - "quint16": ["DT_QUINT16", "unknown"], - "resource": ["DT_RESOURCE", "unknown"], - "string_ref": ["DT_STRING_REF", "unknown"], - "int4": ["DT_INT4", "int4b_t"], - "bfloat16": ["DT_BF16", "bfloat16_t"]} - -def add_dtype_fmt_option_single(x, x_n, is_ref: bool = False): - options = [] - x_fmt = x.get("format") - x_dtype = x.get("dtype") - x_n_in_kernel = x_n + '_REF' if is_ref else x_n - options.append("-DDTYPE_{n}={t}".format(n=x_n_in_kernel, t=DTYPE_MAP.get(x_dtype)[1])) - options.append("-DORIG_DTYPE_{n}={ot}".format(n=x_n_in_kernel, ot=DTYPE_MAP.get(x_dtype)[0])) - options.append("-DFORMAT_{n}=FORMAT_{f}".format(n=x_n_in_kernel, f=x_fmt)) - return options - -def get_dtype_fmt_options(__inputs__, __outputs__): - options = [] - unique_param_name_set = set() - for x in __inputs__: - if x is None: - continue - x_n = x.get("param_name")[:-5].upper() - unique_param_name_set.add(x_n) - options += add_dtype_fmt_option_single(x, x_n) - - for x in __outputs__: - if x is None: - continue - x_n = x.get("param_name")[:-5].upper() - if x_n in unique_param_name_set: - options += add_dtype_fmt_option_single(x, x_n, True) - else: - options += add_dtype_fmt_option_single(x, x_n) - return options - -def load_dso(so_path): - try: - ctypes.CDLL(so_path) - except OSError as error : - print(error) - raise RuntimeError("cannot open %s" %(so_path)) - else: - print("load so succ ", so_path) - -def get_shortsoc_compile_option(compile_option_list: list, shortsoc:str): - compile_options = [] - if shortsoc in compile_option_list: - compile_options = compile_option_list[shortsoc] - elif '__ALLSOC__' in compile_option_list: - compile_options = compile_option_list['__ALLSOC__'] - return compile_options - -def get_kernel_source(src_file, dir_snake, dir_ex): - src_ex = os.path.join(PYF_PATH, "..", "ascendc", dir_ex, src_file) - if os.path.exists(src_ex): - return src_ex - src = os.path.join(PYF_PATH, "..", "ascendc", dir_snake, src_file) - if os.path.exists(src): - return src - src = os.path.join(PYF_PATH, src_file) - if os.path.exists(src): - return src - return src_ex - -''' - -IMPL_API = ''' -@tbe_register.register_operator("{}", trans_bool_to_s8=False) -@para_check.check_op_params({}) -def {}({}, kernel_name="{}", impl_mode=""): - if get_current_build_config("enable_op_prebuild"): - return - __inputs__, __outputs__, __attrs__ = _build_args({}) - options = get_dtype_fmt_options(__inputs__, __outputs__) - options += ["-x", "cce"] - bisheng = shutil.which("bisheng") - if bisheng != None: - bisheng_path = os.path.dirname(bisheng) - tikcpp_path = os.path.realpath(os.path.join(bisheng_path, "..", "..", "tikcpp")) - else: - tikcpp_path = os.path.realpath("/usr/local/Ascend/latest/compiler/tikcpp") - options.append("-I" + tikcpp_path) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw")) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "impl")) - options.append("-I" + os.path.join(tikcpp_path, "tikcfw", "interface")) - options.append("-I" + os.path.join(PYF_PATH, "..", "ascendc", "common")) - if impl_mode == "high_performance": - options.append("-DHIGH_PERFORMANCE=1") - elif impl_mode == "high_precision": - options.append("-DHIGH_PRECISION=1") - if get_default_build_config("enable_deterministic_mode") == 1: - options.append("-DDETEMINISTIC_MODE=1") - - custom_compile_options = {}, - custom_all_compile_options = {}, - soc_version = get_soc_spec("SOC_VERSION") - soc_short = get_soc_spec("SHORT_SOC_VERSION").lower() - custom_compile_options_soc = get_shortsoc_compile_option(custom_compile_options[0], soc_short) - custom_all_compile_options_soc = get_shortsoc_compile_option(custom_all_compile_options[0], soc_short) - options += custom_all_compile_options_soc - options += custom_compile_options_soc - - origin_func_name = "{}" - ascendc_src_dir_ex = "{}" - ascendc_src_dir = "{}" - ascendc_src_file = "{}" - src = get_kernel_source(ascendc_src_file, ascendc_src_dir, ascendc_src_dir_ex) -''' - -REPLAY_OP_API = ''' - print("start replay Acend C Operator {}, kernel name is {}") - tikreplay_codegen_path = tikcpp_path + "/tikreplaylib/lib" - tikreplay_stub_path = tikcpp_path + "/tikreplaylib/lib/" + soc_version - print("start load libtikreplaylib_codegen.so and libtikreplaylib_stub.so") - codegen_so_path = tikreplay_codegen_path + "/libtikreplaylib_codegen.so" - replaystub_so_path = tikreplay_stub_path + "/libtikreplaylib_stub.so" - if PYF_PATH.endswith("dynamic"): - op_replay_path = os.path.join(PYF_PATH, "..", "..", "op_replay") - else: - op_replay_path = os.path.join(PYF_PATH, "..", "op_replay") - replayapi_so_path = os.path.join(op_replay_path, "libreplay_{}_" + soc_short + ".so") - load_dso(codegen_so_path) - load_dso(replaystub_so_path) - load_dso(replayapi_so_path) - op_type = "{}" - entry_obj = os.path.join(op_replay_path, "{}_entry_" + soc_short + ".o") - code_channel = get_code_channel(src, kernel_name, op_type, options) - op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ - attrs = __attrs__, impl_mode = impl_mode) - res, msg = replay_op(op_info, entry_obj, code_channel, src, options) - if not res: - print("call replay op failed for %s and get into call compile op" %(msg)) - compile_op(src, origin_func_name, op_info, options, code_channel, '{}') -''' - -COMPILE_OP_API = ''' - print("start compile Ascend C operator {}. kernel name is " + kernel_name) - op_type = "{}" - code_channel = get_code_channel(src, kernel_name, op_type, options) - op_info = OpInfo(kernel_name = kernel_name, op_type = op_type, inputs = __inputs__, outputs = __outputs__,\\ - attrs = __attrs__, impl_mode = impl_mode, origin_inputs=[{}], origin_outputs = [{}]) - compile_op(src, origin_func_name, op_info, options, code_channel, '{}') -''' - -SUP_API = ''' -def {}({}, impl_mode=""): - __inputs__, __outputs__, __attrs__ = _build_args({}) - ret_str = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) - ret_dict = json.loads(ret_str) - err_code = ret_dict.get("ret_code") - sup = "Unknown" - reason = "Unknown reason" - if err_code is not None: - if err_code is 0: - sup = "True" - reason = "" - elif err_code is 1: - sup = "False" - reason = ret_dict.get("reason") - else: - sup = "Unknown" - reason = ret_dict.get("reason") - return sup, reason -''' -CAP_API = ''' -def {}({}, impl_mode=""): - __inputs__, __outputs__, __attrs__ = _build_args({}) - result = check_op_cap("{}", "{}", __inputs__, __outputs__, __attrs__) - return result.decode("utf-8") -''' -GLZ_API = ''' -@tbe_register.register_param_generalization("{}") -def {}_generalization({}, generalize_config=None): - __inputs__, __outputs__, __attrs__ = _build_args({}) - ret_str = generalize_op_params("{}", __inputs__, __outputs__, __attrs__, generalize_config) - return [json.loads(ret_str)] -''' - -ATTR_DEFAULT = {'bool': 'False', 'int': '0', 'float': '0.0', 'list_int': '[]', - 'list_float': '[]', 'list_bool': '[]', 'list_list_int': '[[]]', 'str': ''} - - -def _get_snake_str(s, i): - if s[i - 1] != '_': - if not s[i - 1].isupper(): - return "_" - elif s[i - 1].isupper() and (i + 1) < len(s) and s[i + 1].islower(): - return "_" - return "" - return "" - - -def optype_snake(origin_str): - temp_str = origin_str[0].lower() + origin_str[1:] - new_str = re.sub(r'([A-Z])', r'_\1', temp_str).lower() - return new_str - - -def optype_snake_ex(s): - snake_case = "" - for i, c in enumerate(s): - if i == 0: - snake_case += c.lower() - elif c.isupper(): - snake_case += _get_snake_str(s, i) - snake_case += c.lower() - else: - snake_case += c - return snake_case - - -class AdpBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - self.argsname = [] - self.argsdefv = [] - self.op_compile_option:str = '{}' - super().__init__(op_type) - - - def write_adapt(self: any, impl_path, path: str, op_compile_option_all:list = None): - self._build_paradefault() - if impl_path != "": - src_file = os.path.join(impl_path, self.op_file + '.cpp') - if not os.path.exists(src_file): - return - out_path = os.path.abspath(path) - if self.dynamic_shape and not out_path.endswith('dynamic'): - out_path = os.path.join(path, 'dynamic') - os.makedirs(out_path, mode=0o700, exist_ok=True) - adpfile = os.path.join(out_path, self.op_file + '.py') - self._gen_op_compile_option(op_compile_option_all) - with os.fdopen(os.open(adpfile, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - self._write_head(fd) - self._write_argparse(fd) - self._write_impl(fd) - if self.op_chk_support: - self._write_cap('check_supported', fd) - self._write_cap('get_op_support_info', fd) - if self.op_fmt_sel: - self._write_cap('op_select_format', fd) - self._write_cap('get_op_specific_info', fd) - if self.op_range_limit == 'limited' or self.op_range_limit == 'dynamic': - self._write_glz(fd) - - - def _gen_op_compile_option(self:any, op_compile_option_all:list =None): - if op_compile_option_all is not None: - if self.op_type in op_compile_option_all: - self.op_compile_option = op_compile_option_all[self.op_type] - elif "__all__" in op_compile_option_all: - self.op_compile_option = op_compile_option_all["__all__"] - - - def _ip_argpack(self: any, default: bool = True) -> list: - args = [] - for i in range(len(self.input_name)): - arg = self.input_name[i] - if default and self.argsdefv[i] is not None: - arg += '=' + self.argsdefv[i] - args.append(arg) - return args - - def _op_argpack(self: any, default: bool = True) -> list: - args = [] - argidx = len(self.input_name) - for i in range(len(self.output_name)): - arg = self.output_name[i] - if default and self.argsdefv[i + argidx] is not None: - arg += '=' + self.argsdefv[i + argidx] - args.append(arg) - return args - - def _attr_argpack(self: any, default: bool = True) -> list: - args = [] - argidx = len(self.input_name) + len(self.output_name) - for i in range(len(self.attr_list)): - att = self.attr_list[i] - arg = att - if default and self.argsdefv[i + argidx] is not None: - if self.attr_val.get(att).get('type') == 'str': - arg += '="' + self.argsdefv[i + argidx] + '"' - elif self.attr_val.get(att).get('type') == 'bool': - arg += '=' + self.argsdefv[i + argidx].capitalize() - else: - arg += '=' + self.argsdefv[i + argidx] - args.append(arg) - return args - - def _build_paralist(self: any, default: bool = True) -> str: - args = [] - args.extend(self._ip_argpack(default)) - args.extend(self._op_argpack(default)) - args.extend(self._attr_argpack(default)) - return ', '.join(args) - - def _io_parachk(self: any, types: list, type_name: str) -> list: - chk = [] - for iot in types: - if iot == 'optional': - ptype = 'OPTION' - else: - ptype = iot.upper() - chk.append('para_check.{}_{}'.format(ptype, type_name)) - return chk - - def _attr_parachk(self: any) -> list: - chk = [] - for att in self.attr_list: - if self.attr_val.get(att).get('paramType') == 'optional': - pt = 'OPTION' - else: - pt = self.attr_val.get(att).get('paramType').upper() - att_type = self.attr_val.get(att).get('type').upper() - chk.append('para_check.{}_ATTR_{}'.format(pt, att_type)) - return chk - - def _build_parachk(self: any) -> str: - chk = [] - chk.extend(self._io_parachk(self.input_type, 'INPUT')) - chk.extend(self._io_parachk(self.output_type, 'OUTPUT')) - chk.extend(self._attr_parachk()) - chk.append('para_check.KERNEL_NAME') - return ', '.join(chk) - - def _build_paradefault(self: any): - optional = False - argtypes = [] - argtypes.extend(self.input_type) - argtypes.extend(self.output_type) - for atype in argtypes: - if atype == 'optional': - optional = True - if optional: - self.argsdefv.append('None') - else: - self.argsdefv.append(None) - for attr in self.attr_list: - atype = self.attr_val.get(attr).get('paramType') - if atype == 'optional': - optional = True - attrval = self.attr_val.get(attr).get('defaultValue') - if attrval is not None: - optional = True - if type == "bool": - attrval = attrval.capitalize() - elif type == "str": - attrval = "\"" + attrval + "\"" - self.argsdefv.append(attrval) - continue - if optional: - self.argsdefv.append(ATTR_DEFAULT.get(self.attr_val.get(attr).get('type'))) - else: - self.argsdefv.append(None) - - def _write_head(self: any, fd: object): - fd.write(IMPL_HEAD) - - def _write_argparse(self: any, fd: object): - args = self._build_paralist(False) - fd.write('def _build_args({}):\n'.format(args)) - fd.write(' __inputs__ = []\n') - fd.write(' for arg in [{}]:\n'.format(', '.join(self.input_name))) - fd.write(' if arg != None:\n') - fd.write(' if isinstance(arg, (list, tuple)):\n') - fd.write(' if len(arg) == 0:\n') - fd.write(' continue\n') - fd.write(' __inputs__.append(arg[0])\n') - fd.write(' else:\n') - fd.write(' __inputs__.append(arg)\n') - fd.write(' __outputs__ = []\n') - fd.write(' for arg in [{}]:\n'.format(', '.join(self.output_name))) - fd.write(' if arg != None:\n') - fd.write(' if isinstance(arg, (list, tuple)):\n') - fd.write(' if len(arg) == 0:\n') - fd.write(' continue\n') - fd.write(' __outputs__.append(arg[0])\n') - fd.write(' else:\n') - fd.write(' __outputs__.append(arg)\n') - fd.write(' __attrs__ = []\n') - for attr in self.attr_list: - fd.write(' if {} != None:\n'.format(attr)) - fd.write(' attr = {}\n') - fd.write(' attr["name"] = "{}"\n'.format(attr)) - fd.write(' attr["dtype"] = "{}"\n'.format(self.attr_val.get(attr).get('type'))) - fd.write(' attr["value"] = {}\n'.format(attr)) - fd.write(' __attrs__.append(attr)\n') - fd.write(' return __inputs__, __outputs__, __attrs__\n') - - def _write_impl(self: any, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - pchk = self._build_parachk() - if len(self.kern_name) > 0: - kern_name = self.kern_name - else: - kern_name = self.op_intf - src = self.op_file + '.cpp' - fd.write(IMPL_API.format(self.op_type, pchk, self.op_intf, argsdef, kern_name, argsval,\ - self.custom_compile_options, self.custom_all_compile_options, self.op_intf,\ - optype_snake_ex(self.op_type), optype_snake(self.op_type), src)) - if self.op_replay_flag: - fd.write(REPLAY_OP_API.format(self.op_type, kern_name, self.op_file, self.op_type, self.op_file,\ - self.op_compile_option)) - else: - fd.write(COMPILE_OP_API.format(self.op_type, self.op_type, ', '.join(self.input_name),\ - ', '.join(self.output_name), self.op_compile_option)) - - def _write_cap(self: any, cap_name: str, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - if cap_name == 'check_supported': - fd.write(SUP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) - else: - fd.write(CAP_API.format(cap_name, argsdef, argsval, cap_name, self.op_type)) - - def _write_glz(self: any, fd: object): - argsdef = self._build_paralist() - argsval = self._build_paralist(False) - fd.write(GLZ_API.format(self.op_type, self.op_intf, argsdef, argsval, self.op_type)) - - -def write_scripts(cfgfile: str, cfgs: dict, dirs: dict, ops: list = None, op_compile_option:list = None): - batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') - iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') - file_map = {} - op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, AdpBuilder,\ - ops, dirs.get(const_var.AUTO_GEN_DIR)) - for op_desc in op_descs: - op_desc.write_adapt(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), op_compile_option) - file_map[op_desc.op_type] = op_desc.op_file - return file_map - -if __name__ == '__main__': - if len(sys.argv) <= 6: - raise RuntimeError('arguments must greater equal than 6') - rep_cfg = {} - rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] - rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] - cfg_dir = {} - cfg_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] - cfg_dir[const_var.CFG_OUT_DIR] = sys.argv[5] - cfg_dir[const_var.AUTO_GEN_DIR] = sys.argv[6] - write_scripts(cfgfile=sys.argv[1], cfgs=rep_cfg, dirs=cfg_dir) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_ops_config.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_ops_config.py deleted file mode 100755 index 7a97180be..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_ops_config.py +++ /dev/null @@ -1,114 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import glob -import json -import argparse -import const_var - - -def load_json(json_file: str): - with open(json_file, encoding='utf-8') as file: - json_content = json.load(file) - return json_content - - -def get_specified_suffix_file(root_dir, suffix): - specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) - all_suffix_files = glob.glob(specified_suffix, recursive=True) - return all_suffix_files - - -def add_simplified_config(op_type, key, core_type, objfile, config): - simple_cfg = config.get('binary_info_config.json') - op_cfg = simple_cfg.get(op_type) - if not op_cfg: - op_cfg = {} - op_cfg['dynamicRankSupport'] = True - op_cfg['simplifiedKeyMode'] = 0 - op_cfg['binaryList'] = [] - simple_cfg[op_type] = op_cfg - bin_list = op_cfg.get('binaryList') - bin_list.append({'coreType': core_type, 'simplifiedKey': key, 'binPath': objfile}) - - -def add_op_config(op_file, bin_info, config): - op_cfg = config.get(op_file) - if not op_cfg: - op_cfg = {} - op_cfg['binList'] = [] - config[op_file] = op_cfg - op_cfg.get('binList').append(bin_info) - - -def gen_ops_config(json_file, soc, config): - core_type_map = {"MIX": 0, "AiCore": 1, "VectorCore": 2} - contents = load_json(json_file) - if ('binFileName' not in contents) or ('supportInfo' not in contents): - return - json_base_name = os.path.basename(json_file) - op_dir = os.path.basename(os.path.dirname(json_file)) - support_info = contents.get('supportInfo') - bin_name = contents.get('binFileName') - bin_suffix = contents.get('binFileSuffix') - core_type = core_type_map.get(contents.get("coreType")) - bin_file_name = bin_name + bin_suffix - op_type = bin_name.split('_')[0] - op_file = op_dir + '.json' - bin_info = {} - keys = support_info.get('simplifiedKey') - if keys: - bin_info['simplifiedKey'] = keys - for key in keys: - add_simplified_config(op_type, key, core_type, os.path.join(soc, op_dir, bin_file_name), config) - bin_info['staticKey'] = support_info.get('staticKey') - bin_info['int64Mode'] = support_info.get('int64Mode') - bin_info['inputs'] = support_info.get('inputs') - bin_info['outputs'] = support_info.get('outputs') - if support_info.get('attrs'): - bin_info['attrs'] = support_info.get('attrs') - bin_info['binInfo'] = {'jsonFilePath': os.path.join(soc, op_dir, json_base_name)} - add_op_config(op_file, bin_info, config) - - -def gen_all_config(root_dir, soc): - suffix = 'json' - config = {} - config['binary_info_config.json'] = {} - all_json_files = get_specified_suffix_file(root_dir, suffix) - for _json in all_json_files: - gen_ops_config(_json, soc, config) - for cfg_key in config.keys(): - cfg_file = os.path.join(root_dir, cfg_key) - with os.fdopen(os.open(cfg_file, const_var.WFLAGS, const_var.WMODES), 'w') as fd: - json.dump(config.get(cfg_key), fd, indent=' ') - - -def args_prase(): - parser = argparse.ArgumentParser() - parser.add_argument('-p', - '--path', - nargs='?', - required=True, - help='Parse the path of the json file.') - parser.add_argument('-s', - '--soc', - nargs='?', - required=True, - help='Parse the soc_version of ops.') - return parser.parse_args() - - -def main(): - args = args_prase() - gen_all_config(args.path, args.soc) - - -if __name__ == '__main__': - main() diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_replay_build.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_replay_build.py deleted file mode 100755 index 1cac7d911..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/ascendc_replay_build.py +++ /dev/null @@ -1,65 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os -import opdesc_parser -import replay_codegen -import const_var -from replay_codegen import ReplayCodeGenParams - -PYF_PATH = os.path.dirname(os.path.realpath(__file__)) - - -class ReplayBuilder(opdesc_parser.OpDesc): - def __init__(self: any, op_type: str): - super().__init__(op_type) - - def gen_replay_source(self: any, impl_path: str, out_path: str, ops_product: str): - if not self.op_replay_flag: - print('{} replay not enabled'.format(self.op_type)) - return - argn = len(self.input_name) + len(self.output_name) + 1 - if self.op_replay_batch: - print('{} replay in batch mode'.format(self.op_type)) - else: - print('{} replay in normal mode'.format(self.op_type)) - if impl_path.endswith('op_kernel'): - implf = os.path.join(impl_path, self.op_file + '.cpp') - tiling_file = os.path.join(impl_path, "../op_host", self.op_file + '_tiling.h') - else: - if self.dynamic_shape: - dyn_path = 'dynamic' - else: - dyn_path = '' - implf = os.path.join(impl_path, dyn_path, self.op_file + '.cpp') - tiling_file = os.path.join(impl_path, "../../op_tiling", self.op_file + '_tiling.h') - rep_conf = replay_codegen.ReplayCodeGen(ReplayCodeGenParams(self.op_type, implf, tiling_file, self.op_file, \ - self.op_intf, argn, self.op_replay_batch, self.max_block_dim, self.max_shape_size)) - rep_conf.set_batch(self.op_replay_batch) - rep_conf.set_outdir(out_path) - rep_conf.gen_replay(ops_product) - - -def gen_replay(cfgfile: str, cfgs: dict, dirs: dict, ops_product: str, ops: list = None): - batch_lists = cfgs.get(const_var.REPLAY_BATCH).split(';') - iterator_lists = cfgs.get(const_var.REPLAY_ITERATE).split(';') - op_descs = opdesc_parser.get_op_desc(cfgfile, batch_lists, iterator_lists, ReplayBuilder, ops) - for op_desc in op_descs: - op_desc.gen_replay_source(dirs.get(const_var.CFG_IMPL_DIR), dirs.get(const_var.CFG_OUT_DIR), ops_product) - - -if __name__ == '__main__': - if len(sys.argv) <= 6: - raise RuntimeError('arguments must greater than 6') - rep_cfg = {} - rep_cfg[const_var.REPLAY_BATCH] = sys.argv[2] - rep_cfg[const_var.REPLAY_ITERATE] = sys.argv[3] - rep_dir = {} - rep_dir[const_var.CFG_IMPL_DIR] = sys.argv[4] - rep_dir[const_var.CFG_OUT_DIR] = sys.argv[5] - gen_replay(sys.argv[1], rep_cfg, rep_dir, sys.argv[6]) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/batch_replay_impl.temp b/atb_operator/AtbTrilSample/TrilOp/cmake/util/batch_replay_impl.temp deleted file mode 100755 index 0e8834664..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/batch_replay_impl.temp +++ /dev/null @@ -1,117 +0,0 @@ -#include -#include -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#include "register/op_check.h" -#define __ASCENDC_REPLAY_CODE__ -#include - -using namespace std; -using namespace optiling; -using namespace AscendCReplay; - -extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); -extern "C" int elf_batch_append(char *elf, uint32_t elfSize, char *jit, int kernum, char *atext[], int alen[], - int atlen, const char* kernelname[]); - -#define KERNEL_N 1 -#define ARG_N (__ARG_NUM__) -#define MAX_L (1024 * 1024 * 100) -#define MAX_E (1024 * 1024) - -int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) -{ - // gen type 1 : direct call codes 0: load .o file - if (param.gentype < 0 || param.gentype > 1) { - printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); - return 0; - } else if (param.gentype == 1 && param.objptr == nullptr) { - printf("Error: call replay with direct call mode, but code obj addr is null\n"); - return 0; - } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { - printf("Error: call replay with object file mode, but object file path is null\n"); - return 0; - } - // core_type 0:MIX 1:CUBE 2:VEC - if (core_type < 0 || core_type > 2) { - printf("Error: call replay core type is %d !\n", core_type); - return 0; - } - g_coreType = __CORE_TYPE__; - g_taskRation = param.task_ration; - g_tilingKey = param.tiling_key; - - unsigned char *buf, *jit; - char *kernel[KERNEL_N]; - int len[KERNEL_N]; - block_idx = 0; - block_num = param.block_dim; - g_ubBase = block_num; - uint8_t *code = (uint8_t *)malloc(MAX_L); - uint8_t *pos = code; - struct timespec tp1, tp2; - - clock_gettime(CLOCK_MONOTONIC, &tp1); - if (block_num > 32) { - printf("Error: block_num > 32\n"); - return 0; - } - //__OP_FOPEN__ - for (int i = 0; i < KERNEL_N; i++) { - //__OP_SET_KERNEL__ - for (int j = 0; j < ARG_N; j++) - AddArg(j, ARG_STEP * (j + 1)); -#ifdef FP_CEILING - SetCtrlFloatEnable(); -#else - SetCtrlFloatDisable(); -#endif - CodeInit(pos, true); - __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); - CodeEnd(); - kernel[i] = (char *)pos; - len[i] = CodeLen(); - pos += len[i]; - } - //__OP_FCLOSE__ - clock_gettime(CLOCK_MONOTONIC, &tp2); - buf = (unsigned char *)malloc(MAX_E); - int fd = open(param.entry_file, O_RDONLY); - if (fd < 0) { - printf("[error]: cannot find entry.o : %s\n", param.entry_file); - return 0; - } - uint32_t bufSize = read(fd, buf, MAX_E); - if (bufSize <= 0) { - printf("[error]: entry.o : %s is too small ! \n", param.entry_file); - } - close(fd); - jit = (unsigned char *)malloc(MAX_L); - printf("total code generated %ld\n", pos - code); - int sz = elf_batch_append((char *)buf, bufSize, (char *)jit, KERNEL_N, kernel, len, pos - code, ¶m.kernel_name); - if (tp1.tv_sec != tp2.tv_sec) { - printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); - } else { - printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); - } - printf("new elf size %d\n", sz); - if (param.gentype == 0) { - fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); - (void)write(fd, jit, sz); - close(fd); - free(jit); - } else if (param.gentype == 1) { - *param.objptr = (char*)jit; - } - free(buf); - free(code); - return sz; -} - -REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/code_channel_infer.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/code_channel_infer.py deleted file mode 100755 index d4e6a540c..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/code_channel_infer.py +++ /dev/null @@ -1,58 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" -import os -import stat -import ctypes -import collections -import shutil -import subprocess -import copy - -"""CODE_* is used to cube/vector api is called in operator code -CODE_MIX means both cube and vector api is called -CODE_CUBE means only cube api is called -CODE_VEC means only vector api is called -""" -CODE_MIX = 0 -CODE_CUBE = 1 -CODE_VEC = 2 - - -def _is_v220(op_product: str): - """return if current soc version is V220 - - Returns: - res: True means V220 - """ - if op_product in ["ascend910b", "ascend910c"]: - return True - return False - - -InfoCodeChanelParams = collections.namedtuple('InfoCodeChanelParams',\ -['src_file', 'tiling_header', 'kernel_name', 'outdir', 'op_product', 'compile_options']) - - -def infer_code_channel(params: InfoCodeChanelParams): - """get code channel for v220, return CODE_MIX if soc version is not V220 - - Args: - src_file (str): AscendC operator code file - src_file (str): AscendC operator tiling header file - kernel_name (str): kernel function name - optype (str): operator type - compile_options (list): compile options for bisheng cmd - - Raises: - Exception: if not exist L1/L0/UB if code, it's not a aicore code - - Returns: - res (int): CODE_MIX/CODE_CUBE/CODE_VEC - """ - if not _is_v220(params.op_product): - return CODE_MIX - return CODE_VEC diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/const_var.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/const_var.py deleted file mode 100755 index 3e71f5ab8..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/const_var.py +++ /dev/null @@ -1,33 +0,0 @@ - -#!/usr/bin/env python -# coding=utf-8 -""" -Function: -The replay funtion entry -Copyright Information: -Huawei Technologies Co., Ltd. All Rights Reserved © 2020 -""" - -import os -import stat - - -REPLAY_BATCH = 'batch' -REPLAY_ITERATE = 'iterate' -CFG_IMPL_DIR = 'impl_dir' -CFG_OUT_DIR = 'out_dir' -AUTO_GEN_DIR = 'auto_gen_dir' -WFLAGS = os.O_WRONLY | os.O_CREAT | os.O_TRUNC -WMODES = stat.S_IWUSR | stat.S_IRUSR -SOC_MAP_EXT = {'ascend310p': 'Ascend310P3', 'ascend310b': 'Ascend310B1', - 'ascend910': 'Ascend910A', 'ascend910b': 'Ascend910B1', 'ascend910c': 'Ascend910C1'} -BIN_CMD = 'opc $1 --main_func={fun} --input_param={param} --soc_version={soc} \ ---output=$2 --impl_mode={impl} --simplified_key_mode=0 --op_mode=dynamic\n' -CHK_CMD = ''' -if ! test -f $2/{res_file} ; then - echo "$2/{res_file} not generated!" - exit 1 -fi -''' -ATTR_DEF_VAL = {'str' : '', 'int': 0, 'float': 0.0, 'bool': False, 'list_bool': [], - 'list_int': [], 'list_float': [], 'list_list_int': [[]]} diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_impl_and_mrege_json.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_impl_and_mrege_json.sh deleted file mode 100755 index 492034e49..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_impl_and_mrege_json.sh +++ /dev/null @@ -1,20 +0,0 @@ -#!/usr/bin/bash - -project_path=$1 -build_path=$2 -vendor_name=customize -if [[ ! -d "$project_path" ]]; then - echo "[ERROR] No projcet path is provided" - exit 1 -fi - -if [[ ! -d "$build_path" ]]; then - echo "[ERROR] No build path is provided" - exit 1 -fi - -# copy aicpu kernel so operators -if [[ -d "${project_path}/cpukernel/aicpu_kernel_lib" ]]; then - cp -f ${project_path}/cpukernel/aicpu_kernel_lib/* ${build_path}/makepkg/packages/vendors/$vendor_name/op_impl/cpu/aicpu_kernel/impl - rm -rf ${project_path}/cpukernel/aicpu_kernel_lib -fi diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_ops_filter.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_ops_filter.sh deleted file mode 100755 index d4c27d17f..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_ops_filter.sh +++ /dev/null @@ -1,62 +0,0 @@ -#!/bin/bash -# Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -# Description: Generate npu_supported_ops.json -# ============================================================================== - -if [[ -z "$1" ]]; then - echo -e "[ERROR] No source dir provided" - exit 1 -fi - -if [[ -z "$2" ]]; then - echo -e "[ERROR] No destination dir provided" - exit 1 -fi - -src=$1 -dest_file=$2/npu_supported_ops.json - -if [ -f "$dest_file" ];then - chmod u+w $dest_file -fi - -echo $* - -add_ops() { - name=$1 - isHeavy=$2 - file=$3 - grep -w "\"$name\"" ${file} >/dev/null - if [ $? == 0 ];then - return - fi - echo " \"${name}\": {" >> ${file} - echo " \"isGray\": false," >> ${file} - echo " \"isHeavy\": ${isHeavy}" >> ${file} - echo " }," >> ${file} -} - -echo "{" > ${dest_file} -ini_files=$(find ${src} -name "*.ini") -for file in ${ini_files} ; do - name=$(grep '^\[' ${file} | sed 's/\[//g' | sed 's/]//g' | sed 's/\r//g') - grep 'heavyOp.flag' ${file} >/dev/null - if [ $? == 0 ];then - isHeavy=$(grep 'heavyOp.flag' ${file} | awk -F= '{print $2}') - else - isHeavy="false" - fi - for op in ${name} ; do - add_ops ${op} "false" ${dest_file} - done -done -echo "}" >> ${dest_file} -file_count=$(cat ${dest_file} | wc -l) -line=$(($file_count-1)) -sed -i "${line}{s/,//g}" ${dest_file} - -chmod 640 "${dest_file}" -echo -e "[INFO] Succed generated ${dest_file}" - -exit 0 - diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_version_info.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_version_info.sh deleted file mode 100755 index a06cfc78d..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/gen_version_info.sh +++ /dev/null @@ -1,6 +0,0 @@ -ascend_install_dir=$1 -gen_file_dir=$2 - -# create version.info -compiler_version=$(grep "Version" -w ${ascend_install_dir}/compiler/version.info | awk -F = '{print $2}') -echo "custom_opp_compiler_version=${compiler_version}" > ${gen_file_dir}/version.info \ No newline at end of file diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_op_info.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_op_info.py deleted file mode 100755 index 28ba08757..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_op_info.py +++ /dev/null @@ -1,36 +0,0 @@ -# -*- coding: utf-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" -import json -import os -import sys -import stat -import const_var - - -if __name__ == '__main__': - if len(sys.argv) != 3: - print(sys.argv) - print('argv error, inert_op_info.py your_op_file lib_op_file') - sys.exit(2) - - with open(sys.argv[1], 'r') as load_f: - insert_operator = json.load(load_f) - - all_operators = {} - if os.path.exists(sys.argv[2]): - if os.path.getsize(sys.argv[2]) != 0: - with open(sys.argv[2], 'r') as load_f: - all_operators = json.load(load_f) - - for k in insert_operator.keys(): - if k in all_operators.keys(): - print('replace op:[', k, '] success') - else: - print('insert op:[', k, '] success') - all_operators[k] = insert_operator[k] - - with os.fdopen(os.open(sys.argv[2], const_var.WFLAGS, const_var.WMODES), 'w') as json_file: - json_file.write(json.dumps(all_operators, indent=4)) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_simplified_keys.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_simplified_keys.py deleted file mode 100755 index 7e8dc2c55..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/insert_simplified_keys.py +++ /dev/null @@ -1,242 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import os -import re -import glob -import json -import argparse - - -DATA_TPYE_DICT = { - 'float32': 0, - 'float16': 1, - 'int8': 2, - 'int16': 6, - 'uint16': 7, - 'uint8': 4, - 'int32': 3, - 'int64': 9, - 'uint32': 8, - 'uint64': 10, - 'bool': 12, - 'double': 11, - 'complex64': 16, - 'complex128': 17, - 'qint8': 18, - 'qint16': 19, - 'qint32': 20, - 'quint8': 21, - 'quint16': 22, - 'resource': 23, - 'string': 24, - 'dual': 25, - 'variant': 26, - 'bf16': 27, - 'bfloat16': 27, - 'undefined': 28, - 'int4': 29, - 'uint1': 30, - 'int2': 31 -} - -FORMAT_DICT = { - 'NCHW': 0, - 'NHWC': 1, - 'ND': 2, - 'NC1HWC0': 3, - 'FRACTAL_Z': 4, - 'NC1C0HWPAD': 5, - 'NHWC1C0': 6, - 'FSR_NCHW': 7, - 'FRACTAL_DECONV': 8, - 'C1HWNC0': 9, - 'FRACTAL_DECONV_TRANSPOSE': 10, - 'FRACTAL_DECONV_SP_STRIDE_TRANS': 11, - 'NC1HWC0_C04': 12, - 'FRACTAL_Z_C04': 13, - 'CHWN': 14, - 'FRACTAL_DECONV_SP_STRIDE8_TRANS': 15, - 'HWCN': 16, - 'NC1KHKWHWC0': 17, - 'BN_WEIGHT': 18, - 'FILTER_HWCK': 19, - 'HASHTABLE_LOOKUP_LOOKUPS': 20, - 'HASHTABLE_LOOKUP_KEYS': 21, - 'HASHTABLE_LOOKUP_VALUE': 22, - 'HASHTABLE_LOOKUP_OUTPUT': 23, - 'HASHTABLE_LOOKUP_HITS': 24, - 'C1HWNCoC0': 25, - 'MD': 26, - 'NDHWC': 27, - 'FRACTAL_ZZ': 28, - 'FRACTAL_NZ': 29, - 'NCDHW': 30, - 'DHWCN': 31, - 'NDC1HWC0': 32, - 'FRACTAL_Z_3D': 33, - 'CN': 34, - 'NC': 35, - 'DHWNC': 36, - 'FRACTAL_Z_3D_TRANSPOSE': 37, - 'FRACTAL_ZN_LSTM': 38, - 'FRACTAL_Z_G': 39, - 'RESERVED': 40, - 'ALL': 41, - 'NULL': 42, - 'ND_RNN_BIAS': 43, - 'FRACTAL_ZN_RNN': 44, - 'NYUV': 45, - 'NYUV_A': 46 -} - - -def load_json(json_file: str): - with open(json_file, encoding='utf-8') as file: - json_content = json.load(file) - return json_content - - -def get_specified_suffix_file(root_dir, suffix): - specified_suffix = os.path.join(root_dir, '**/*.{}'.format(suffix)) - all_suffix_files = glob.glob(specified_suffix, recursive=True) - return all_suffix_files - - -def get_deterministic_value(support_info): - deterministic_key = 'deterministic' - if deterministic_key not in support_info: - return 0 - deterministic_value = support_info.get(deterministic_key) - if deterministic_value == 'true': - return 1 - else: - return 0 - - -def get_precision_value(support_info): - precision_key = 'implMode' - precision_value = support_info.get(precision_key) - if precision_value == 'high_performance': - _value = 1 - elif precision_value == 'high_precision': - _value = 2 - else: - _value = 0 - return _value - - -def get_overflow_value(support_info): - return 0 - - -def get_parameters(info): - if info: - if 'dtype' in info: - data_type = info['dtype'] - data_type_value = DATA_TPYE_DICT.get(data_type) - else: - data_type_value = 0 - if 'format' in info: - _format = info['format'] - _format_value = FORMAT_DICT.get(_format) - else: - _format_value = 0 - else: - data_type_value = 0 - _format_value = 0 - return str(data_type_value), str(_format_value) - - -def get_dynamic_parameters(info): - # 动态输入时只需获取第一个参数 - return get_parameters(info[0]) - - -def get_all_parameters(support_info, _type): - result_list = list() - info_lists = support_info.get(_type) - if info_lists: - for _info in info_lists: - # 输入为列表时是动态输入 - if isinstance(_info, (list, tuple)): - data_type_value, _format_value = get_dynamic_parameters(_info) - else: - data_type_value, _format_value = get_parameters(_info) - result_list.append("{},{}".format(data_type_value, _format_value)) - return result_list - - -def get_all_input_parameters(support_info): - result = get_all_parameters(support_info, 'inputs') - return '/'.join(result) - - -def insert_content_into_file(input_file, content): - with open(input_file, 'r+') as file: - lines = file.readlines() - for index, line in enumerate(lines): - match_result = re.search(r'"staticKey":', line) - if match_result: - count = len(line) - len(line.lstrip()) - new_content = "{}{}".format(' ' * count, content) - # 插入到前一行,防止插入最后时还需要考虑是否添加逗号 - lines.insert(index, new_content) - break - file.seek(0) - file.write(''.join(lines)) - - -def insert_simplified_keys(json_file): - contents = load_json(json_file) - # 不存在'binFileName'或者'supportInfo'字段时,非需要替换的解析json文件 - if ('binFileName' not in contents) or ('supportInfo' not in contents): - return - support_info = contents.get('supportInfo') - bin_file_name = contents.get('binFileName') - # 'simplifiedKey'字段已经存在时,直接返回,不重复生成 - if 'simplifiedKey' in support_info: - return - op_type = bin_file_name.split('_')[0] - deterministic = str(get_deterministic_value(support_info)) - precision = str(get_precision_value(support_info)) - overflow = str(get_overflow_value(support_info)) - input_parameters = get_all_input_parameters(support_info) - key = '{}/d={},p={},o={}/{}/'.format( - op_type, - deterministic, - precision, - overflow, - input_parameters) - result = '"simplifiedKey": "' + key + '",\n' - insert_content_into_file(json_file, result) - - -def insert_all_simplified_keys(root_dir): - suffix = 'json' - all_json_files = get_specified_suffix_file(root_dir, suffix) - for _json in all_json_files: - insert_simplified_keys(_json) - - -def args_prase(): - parser = argparse.ArgumentParser() - parser.add_argument('-p', - '--path', - nargs='?', - required=True, - help='Parse the path of the json file.') - return parser.parse_args() - - -def main(): - args = args_prase() - insert_all_simplified_keys(args.path) - - -if __name__ == '__main__': - main() diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_entry.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_entry.py deleted file mode 100755 index 2b77c970d..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_entry.py +++ /dev/null @@ -1,115 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - - -def gen_fun_def(title, kernel, argn, arg_type, arg_name): - entry = [] - entry.append(title) - entry.append(kernel) - entry.append('(') - args = [] - for i in range(0, argn): - args.append(arg_type + ' ' + arg_name + str(i)) - entry.append(', '.join(args)) - entry.append(')') - return ' '.join(entry) - - -def gen_batch_kernel_body(fname, argn, arg_name): - body = [] - body.append('{') - fun = [] - fun.append(fname) - fun.append('(') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - fun.append(', '.join(args)) - fun.append(');') - body.append(' '.join(fun)) - body.append('}') - return '\n'.join(body) - - -def gen_mc_kernel_body(kn, argn, arg_name, blknum): - body = [] - body.append('{') - body.append(' switch(block_idx) {') - for blk in range(0, blknum): - fun = [] - fun.append('{}_blk{:02d}'.format(kn, blk)) - fun.append('(') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - fun.append(', '.join(args)) - fun.append(')') - body.append(' case {}: {}; break;'.format(blk, ' '.join(fun))) - body.append(' default: break;') - body.append(' }') - body.append('}') - return '\n'.join(body) - - -def gen_proc_body(argn, arg_name): - body = [] - body.append('{') - args = [] - for i in range(0, argn): - args.append(arg_name + str(i)) - body.append('uint64_t __x = (uint64_t)' + ' + (uint64_t)'.join(args) + ';') - body.append('__asm__ ("NOP");') - body.append('__asm__ ("NOP");') - body.append('__asm__ ("NOP");') - body.append('}') - return '\n'.join(body) - - -def batch_code_gen(kn, argn, argt): - codes = [] - kernel_name = kn - proc_name = kernel_name + '_percore' - arg_num = int(argn) - data_type = argt - arg_type = '__gm__ ' + data_type + '* __restrict__' - arg_name = 'arg' - kernel_title = 'extern \"C\" __global__ __aicore__ void' - proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' - codes.append('#ifndef __aicore__') - codes.append('#define __aicore__ [aicore]') - codes.append('#endif') - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') - codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) - codes.append(gen_batch_kernel_body(proc_name, arg_num, arg_name)) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) - codes.append(gen_proc_body(arg_num, arg_name)) - return '\n'.join(codes) + '\n' - - -def mc_code_gen(kn, argn, argt, blknum): - codes = [] - kernel_name = kn - core_num = int(blknum) - arg_num = int(argn) - data_type = argt - arg_type = '__gm__ ' + data_type + '* __restrict__' - arg_name = 'arg' - kernel_title = 'extern \"C\" __global__ __aicore__ void' - proc_title = 'extern \"C\" __attribute__((noinline)) __aicore__ void' - codes.append('#ifndef __aicore__') - codes.append('#define __aicore__ [aicore]') - codes.append('#endif') - for i in range(0, core_num): - proc_name = '{}_blk{:02d}'.format(kernel_name, i) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name) + ';') - codes.append(gen_fun_def(kernel_title, kernel_name, arg_num, arg_type, arg_name)) - codes.append(gen_mc_kernel_body(kernel_name, arg_num, arg_name, core_num)) - for i in range(0, core_num): - proc_name = '{}_blk{:02d}'.format(kernel_name, i) - codes.append(gen_fun_def(proc_title, proc_name, arg_num, arg_type, arg_name)) - codes.append(gen_proc_body(arg_num, arg_name)) - return '\n'.join(codes) + '\n' diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_impl.temp b/atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_impl.temp deleted file mode 100755 index 5079a1043..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/kernel_impl.temp +++ /dev/null @@ -1,10 +0,0 @@ -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#define __ASCENDC_REPLAY_CODE__ -#include "__CCE_FILE__" diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/COPYING b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/COPYING deleted file mode 100755 index d159169d1..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/COPYING +++ /dev/null @@ -1,339 +0,0 @@ - GNU GENERAL PUBLIC LICENSE - Version 2, June 1991 - - Copyright (C) 1989, 1991 Free Software Foundation, Inc., - 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA - Everyone is permitted to copy and distribute verbatim copies - of this license document, but changing it is not allowed. - - Preamble - - The licenses for most software are designed to take away your -freedom to share and change it. By contrast, the GNU General Public -License is intended to guarantee your freedom to share and change free -software--to make sure the software is free for all its users. This -General Public License applies to most of the Free Software -Foundation's software and to any other program whose authors commit to -using it. (Some other Free Software Foundation software is covered by -the GNU Lesser General Public License instead.) You can apply it to -your programs, too. - - When we speak of free software, we are referring to freedom, not -price. Our General Public Licenses are designed to make sure that you -have the freedom to distribute copies of free software (and charge for -this service if you wish), that you receive source code or can get it -if you want it, that you can change the software or use pieces of it -in new free programs; and that you know you can do these things. - - To protect your rights, we need to make restrictions that forbid -anyone to deny you these rights or to ask you to surrender the rights. -These restrictions translate to certain responsibilities for you if you -distribute copies of the software, or if you modify it. - - For example, if you distribute copies of such a program, whether -gratis or for a fee, you must give the recipients all the rights that -you have. You must make sure that they, too, receive or can get the -source code. And you must show them these terms so they know their -rights. - - We protect your rights with two steps: (1) copyright the software, and -(2) offer you this license which gives you legal permission to copy, -distribute and/or modify the software. - - Also, for each author's protection and ours, we want to make certain -that everyone understands that there is no warranty for this free -software. If the software is modified by someone else and passed on, we -want its recipients to know that what they have is not the original, so -that any problems introduced by others will not reflect on the original -authors' reputations. - - Finally, any free program is threatened constantly by software -patents. We wish to avoid the danger that redistributors of a free -program will individually obtain patent licenses, in effect making the -program proprietary. To prevent this, we have made it clear that any -patent must be licensed for everyone's free use or not licensed at all. - - The precise terms and conditions for copying, distribution and -modification follow. - - GNU GENERAL PUBLIC LICENSE - TERMS AND CONDITIONS FOR COPYING, DISTRIBUTION AND MODIFICATION - - 0. This License applies to any program or other work which contains -a notice placed by the copyright holder saying it may be distributed -under the terms of this General Public License. The "Program", below, -refers to any such program or work, and a "work based on the Program" -means either the Program or any derivative work under copyright law: -that is to say, a work containing the Program or a portion of it, -either verbatim or with modifications and/or translated into another -language. (Hereinafter, translation is included without limitation in -the term "modification".) Each licensee is addressed as "you". - -Activities other than copying, distribution and modification are not -covered by this License; they are outside its scope. The act of -running the Program is not restricted, and the output from the Program -is covered only if its contents constitute a work based on the -Program (independent of having been made by running the Program). -Whether that is true depends on what the Program does. - - 1. You may copy and distribute verbatim copies of the Program's -source code as you receive it, in any medium, provided that you -conspicuously and appropriately publish on each copy an appropriate -copyright notice and disclaimer of warranty; keep intact all the -notices that refer to this License and to the absence of any warranty; -and give any other recipients of the Program a copy of this License -along with the Program. - -You may charge a fee for the physical act of transferring a copy, and -you may at your option offer warranty protection in exchange for a fee. - - 2. You may modify your copy or copies of the Program or any portion -of it, thus forming a work based on the Program, and copy and -distribute such modifications or work under the terms of Section 1 -above, provided that you also meet all of these conditions: - - a) You must cause the modified files to carry prominent notices - stating that you changed the files and the date of any change. - - b) You must cause any work that you distribute or publish, that in - whole or in part contains or is derived from the Program or any - part thereof, to be licensed as a whole at no charge to all third - parties under the terms of this License. - - c) If the modified program normally reads commands interactively - when run, you must cause it, when started running for such - interactive use in the most ordinary way, to print or display an - announcement including an appropriate copyright notice and a - notice that there is no warranty (or else, saying that you provide - a warranty) and that users may redistribute the program under - these conditions, and telling the user how to view a copy of this - License. (Exception: if the Program itself is interactive but - does not normally print such an announcement, your work based on - the Program is not required to print an announcement.) - -These requirements apply to the modified work as a whole. If -identifiable sections of that work are not derived from the Program, -and can be reasonably considered independent and separate works in -themselves, then this License, and its terms, do not apply to those -sections when you distribute them as separate works. But when you -distribute the same sections as part of a whole which is a work based -on the Program, the distribution of the whole must be on the terms of -this License, whose permissions for other licensees extend to the -entire whole, and thus to each and every part regardless of who wrote it. - -Thus, it is not the intent of this section to claim rights or contest -your rights to work written entirely by you; rather, the intent is to -exercise the right to control the distribution of derivative or -collective works based on the Program. - -In addition, mere aggregation of another work not based on the Program -with the Program (or with a work based on the Program) on a volume of -a storage or distribution medium does not bring the other work under -the scope of this License. - - 3. You may copy and distribute the Program (or a work based on it, -under Section 2) in object code or executable form under the terms of -Sections 1 and 2 above provided that you also do one of the following: - - a) Accompany it with the complete corresponding machine-readable - source code, which must be distributed under the terms of Sections - 1 and 2 above on a medium customarily used for software interchange; or, - - b) Accompany it with a written offer, valid for at least three - years, to give any third party, for a charge no more than your - cost of physically performing source distribution, a complete - machine-readable copy of the corresponding source code, to be - distributed under the terms of Sections 1 and 2 above on a medium - customarily used for software interchange; or, - - c) Accompany it with the information you received as to the offer - to distribute corresponding source code. (This alternative is - allowed only for noncommercial distribution and only if you - received the program in object code or executable form with such - an offer, in accord with Subsection b above.) - -The source code for a work means the preferred form of the work for -making modifications to it. For an executable work, complete source -code means all the source code for all modules it contains, plus any -associated interface definition files, plus the scripts used to -control compilation and installation of the executable. However, as a -special exception, the source code distributed need not include -anything that is normally distributed (in either source or binary -form) with the major components (compiler, kernel, and so on) of the -operating system on which the executable runs, unless that component -itself accompanies the executable. - -If distribution of executable or object code is made by offering -access to copy from a designated place, then offering equivalent -access to copy the source code from the same place counts as -distribution of the source code, even though third parties are not -compelled to copy the source along with the object code. - - 4. You may not copy, modify, sublicense, or distribute the Program -except as expressly provided under this License. Any attempt -otherwise to copy, modify, sublicense or distribute the Program is -void, and will automatically terminate your rights under this License. -However, parties who have received copies, or rights, from you under -this License will not have their licenses terminated so long as such -parties remain in full compliance. - - 5. You are not required to accept this License, since you have not -signed it. However, nothing else grants you permission to modify or -distribute the Program or its derivative works. These actions are -prohibited by law if you do not accept this License. Therefore, by -modifying or distributing the Program (or any work based on the -Program), you indicate your acceptance of this License to do so, and -all its terms and conditions for copying, distributing or modifying -the Program or works based on it. - - 6. Each time you redistribute the Program (or any work based on the -Program), the recipient automatically receives a license from the -original licensor to copy, distribute or modify the Program subject to -these terms and conditions. You may not impose any further -restrictions on the recipients' exercise of the rights granted herein. -You are not responsible for enforcing compliance by third parties to -this License. - - 7. If, as a consequence of a court judgment or allegation of patent -infringement or for any other reason (not limited to patent issues), -conditions are imposed on you (whether by court order, agreement or -otherwise) that contradict the conditions of this License, they do not -excuse you from the conditions of this License. If you cannot -distribute so as to satisfy simultaneously your obligations under this -License and any other pertinent obligations, then as a consequence you -may not distribute the Program at all. For example, if a patent -license would not permit royalty-free redistribution of the Program by -all those who receive copies directly or indirectly through you, then -the only way you could satisfy both it and this License would be to -refrain entirely from distribution of the Program. - -If any portion of this section is held invalid or unenforceable under -any particular circumstance, the balance of the section is intended to -apply and the section as a whole is intended to apply in other -circumstances. - -It is not the purpose of this section to induce you to infringe any -patents or other property right claims or to contest validity of any -such claims; this section has the sole purpose of protecting the -integrity of the free software distribution system, which is -implemented by public license practices. Many people have made -generous contributions to the wide range of software distributed -through that system in reliance on consistent application of that -system; it is up to the author/donor to decide if he or she is willing -to distribute software through any other system and a licensee cannot -impose that choice. - -This section is intended to make thoroughly clear what is believed to -be a consequence of the rest of this License. - - 8. If the distribution and/or use of the Program is restricted in -certain countries either by patents or by copyrighted interfaces, the -original copyright holder who places the Program under this License -may add an explicit geographical distribution limitation excluding -those countries, so that distribution is permitted only in or among -countries not thus excluded. In such case, this License incorporates -the limitation as if written in the body of this License. - - 9. The Free Software Foundation may publish revised and/or new versions -of the General Public License from time to time. Such new versions will -be similar in spirit to the present version, but may differ in detail to -address new problems or concerns. - -Each version is given a distinguishing version number. If the Program -specifies a version number of this License which applies to it and "any -later version", you have the option of following the terms and conditions -either of that version or of any later version published by the Free -Software Foundation. If the Program does not specify a version number of -this License, you may choose any version ever published by the Free Software -Foundation. - - 10. If you wish to incorporate parts of the Program into other free -programs whose distribution conditions are different, write to the author -to ask for permission. For software which is copyrighted by the Free -Software Foundation, write to the Free Software Foundation; we sometimes -make exceptions for this. Our decision will be guided by the two goals -of preserving the free status of all derivatives of our free software and -of promoting the sharing and reuse of software generally. - - NO WARRANTY - - 11. BECAUSE THE PROGRAM IS LICENSED FREE OF CHARGE, THERE IS NO WARRANTY -FOR THE PROGRAM, TO THE EXTENT PERMITTED BY APPLICABLE LAW. EXCEPT WHEN -OTHERWISE STATED IN WRITING THE COPYRIGHT HOLDERS AND/OR OTHER PARTIES -PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED -OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF -MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE. THE ENTIRE RISK AS -TO THE QUALITY AND PERFORMANCE OF THE PROGRAM IS WITH YOU. SHOULD THE -PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF ALL NECESSARY SERVICING, -REPAIR OR CORRECTION. - - 12. IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING -WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MAY MODIFY AND/OR -REDISTRIBUTE THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, -INCLUDING ANY GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING -OUT OF THE USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED -TO LOSS OF DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY -YOU OR THIRD PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER -PROGRAMS), EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE -POSSIBILITY OF SUCH DAMAGES. - - END OF TERMS AND CONDITIONS - - How to Apply These Terms to Your New Programs - - If you develop a new program, and you want it to be of the greatest -possible use to the public, the best way to achieve this is to make it -free software which everyone can redistribute and change under these terms. - - To do so, attach the following notices to the program. It is safest -to attach them to the start of each source file to most effectively -convey the exclusion of warranty; and each file should have at least -the "copyright" line and a pointer to where the full notice is found. - - - Copyright (C) - - This program is free software; you can redistribute it and/or modify - it under the terms of the GNU General Public License as published by - the Free Software Foundation; either version 2 of the License, or - (at your option) any later version. - - This program is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the - GNU General Public License for more details. - - You should have received a copy of the GNU General Public License along - with this program; if not, write to the Free Software Foundation, Inc., - 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. - -Also add information on how to contact you by electronic and paper mail. - -If the program is interactive, make it output a short notice like this -when it starts in an interactive mode: - - Gnomovision version 69, Copyright (C) year name of author - Gnomovision comes with ABSOLUTELY NO WARRANTY; for details type `show w'. - This is free software, and you are welcome to redistribute it - under certain conditions; type `show c' for details. - -The hypothetical commands `show w' and `show c' should show the appropriate -parts of the General Public License. Of course, the commands you use may -be called something other than `show w' and `show c'; they could even be -mouse-clicks or menu items--whatever suits your program. - -You should also get your employer (if you work as a programmer) or your -school, if any, to sign a "copyright disclaimer" for the program, if -necessary. Here is a sample; alter the names: - - Yoyodyne, Inc., hereby disclaims all copyright interest in the program - `Gnomovision' (which makes passes at compilers) written by James Hacker. - - , 1 April 1989 - Ty Coon, President of Vice - -This General Public License does not permit incorporating your program into -proprietary programs. If your program is a subroutine library, you may -consider it more useful to permit linking proprietary applications with the -library. If this is what you want to do, use the GNU Lesser General -Public License instead of this License. diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/README.md b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/README.md deleted file mode 100755 index b41f01682..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/README.md +++ /dev/null @@ -1,246 +0,0 @@ -[![License: GPL v2](https://img.shields.io/badge/License-GPL%20v2-blue.svg)](https://www.gnu.org/licenses/old-licenses/gpl-2.0.en.html) -![Build Status](https://github.com/megastep/makeself/workflows/CI/badge.svg) - -# makeself - Make self-extractable archives on Unix - -[makeself.sh][1] is a small shell script that generates a self-extractable -compressed tar archive from a directory. The resulting file appears as a shell script -(many of those have a **.run** suffix), and can be launched as is. The archive -will then uncompress itself to a temporary directory and an optional arbitrary -command will be executed (for example an installation script). This is pretty -similar to archives generated with WinZip Self-Extractor in the Windows world. -Makeself archives also include checksums for integrity self-validation (CRC -and/or MD5/SHA256 checksums). - -The makeself.sh script itself is used only to create the archives from a -directory of files. The resultant archive is actually a compressed (using -gzip, bzip2, or compress) TAR archive, with a small shell script stub at the -beginning. This small stub performs all the steps of extracting the files, -running the embedded command, and removing the temporary files when done. -All the user has to do to install the software contained in such an -archive is to "run" the archive, i.e **sh nice-software.run**. I recommend -using the ".run" (which was introduced by some Makeself archives released by -Loki Software) or ".sh" suffix for such archives not to confuse the users, -so that they will know they are actually shell scripts (with quite a lot of binary data -attached to them though!). - -I am trying to keep the code of this script as portable as possible, i.e it is -not relying on any bash-specific features and only calls commands that are -installed on any functioning UNIX-compatible system. This script as well as -the archives it generates should run on any Unix flavor, with any compatible -Bourne shell, provided of course that the compression programs are available. - -As of version 2.1, Makeself has been rewritten and tested on the following -platforms : - - * Linux (all distributions) - * Sun Solaris (8 and above) - * HP-UX (tested on 11.0 and 11i on HPPA RISC) - * SCO OpenUnix and OpenServer - * IBM AIX 5.1L - * macOS (Darwin) - * SGI IRIX 6.5 - * FreeBSD - * UnicOS / Cray - * Cygwin (Windows) - -If you successfully run Makeself and/or archives created with it on another -system, then please [let me know][2]! - -Examples of publicly available archives made using makeself are : - - * Game patches and installers for [Id Software][3] games like Quake 3 for Linux or Return To Castle Wolfenstein ; - * All game patches released by [Loki Software][4] for the Linux version of popular games ; - * The [nVidia drivers][5] for Linux - * The installer for the Linux version of [Google Earth][6] - * The [VirtualBox][7] installers for Linux - * The [Makeself][1] distribution itself ;-) - * and countless others... - -**Important note for Apache users:** By default, most Web servers will think that Makeself archives are regular text files and thus they may show up as text in a Web browser. The correct way to prevent this is to add a MIME type for this file format, like so (in httpd.conf) : - -`AddType application/x-makeself .run` - -**Important note for certain GNU/Linux distributions:** Archives created with Makeself prior to v2.1.2 were using an old syntax for the _head_ and _tail_ Unix commands that is being progressively obsoleted in their GNU forms. Therefore you may have problems uncompressing some of these archives. A workaround for this is to set the environment variable $_POSIX2_VERSION to enable the old syntax, i.e. : - -`export _POSIX2_VERSION=199209` - -## Usage - -The syntax of makeself is the following: - -``` -makeself.sh [args] archive_dir file_name label startup_script [script_args] -``` - - * _args_ are optional options for Makeself. The available ones are : - - * **`--version`** : Prints the version number on stdout, then exits immediately - * **`--gzip`** : Use gzip for compression (the default on platforms on which gzip is commonly available, like Linux) - * **`--bzip2`** : Use bzip2 instead of gzip for better compression. The bzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. - * **`--pbzip2`** : Use pbzip2 instead of gzip for better and faster compression on machines having multiple CPUs. The pbzip2 command must be available in the command path. It is recommended that the archive prefix be set to something like '.bz2.run', so that potential users know that they'll need bzip2 to extract it. - * **`--xz`** : Use xz instead of gzip for better compression. The xz command must be available in the command path. It is recommended that the archive prefix be set to something like '.xz.run' for the archive, so that potential users know that they'll need xz to extract it. - * **`--lzo`** : Use lzop instead of gzip for better compression. The lzop command must be available in the command path. It is recommended that the archive prefix be set to something like `.lzo.run` for the archive, so that potential users know that they'll need lzop to extract it. - * **`--lz4`** : Use lz4 instead of gzip for better compression. The lz4 command must be available in the command path. It is recommended that the archive prefix be set to something like '.lz4.run' for the archive, so that potential users know that they'll need lz4 to extract it. - * **`--zstd`** : Use zstd instead of gzip for better compression. The zstd command must be available in the command path. It is recommended that the archive prefix be set to something like '.zstd.run' for the archive, so that potential users know that they'll need zstd to extract it. - * **`--pigz`** : Use pigz for compression. - * **`--base64`** : Encode the archive to ASCII in Base64 format instead of compressing (base64 command required). - * **`--gpg-encrypt`** : Encrypt the archive using `gpg -ac -z $COMPRESS_LEVEL`. This will prompt for a password to encrypt with. Assumes that potential users have `gpg` installed. - * **`--ssl-encrypt`** : Encrypt the archive using `openssl aes-256-cbc -a -salt`. This will prompt for a password to encrypt with. Assumes that the potential users have the OpenSSL tools installed. - * **`--compress`** : Use the UNIX `compress` command to compress the data. This should be the default on all platforms that don't have gzip available. - * **`--nocomp`** : Do not use any compression for the archive, which will then be an uncompressed TAR. - * **`--complevel`** : Specify the compression level for gzip, bzip2, pbzip2, zstd, xz, lzo or lz4. (defaults to 9) - * **`--threads`** : Specify the number of threads to be used by compressors that support parallelization. Omit to use compressor's default. Most useful (and required) for opting into xz's threading, usually with `--threads=0` for all available cores. pbzip2 and pigz are parallel by default, and setting this value allows limiting the number of threads they use. - * **`--notemp`** : The generated archive will not extract the files to a temporary directory, but in a new directory created in the current directory. This is better to distribute software packages that may extract and compile by themselves (i.e. launch the compilation through the embedded script). - * **`--current`** : Files will be extracted to the current directory, instead of in a subdirectory. This option implies `--notemp` above. - * **`--follow`** : Follow the symbolic links inside of the archive directory, i.e. store the files that are being pointed to instead of the links themselves. - * **`--append`** _(new in 2.1.x)_: Append data to an existing archive, instead of creating a new one. In this mode, the settings from the original archive are reused (compression type, label, embedded script), and thus don't need to be specified again on the command line. - * **`--header`** : Makeself uses a separate file to store the header stub, called `makeself-header.sh`. By default, it is assumed that it is stored in the same location as makeself.sh. This option can be used to specify its actual location if it is stored someplace else. - * **`--cleanup`** : Specify a script that is run when execution is interrupted or finishes successfully. The script is executed with the same environment and initial `script_args` as `startup_script`. - * **`--copy`** : Upon extraction, the archive will first extract itself to a temporary directory. The main application of this is to allow self-contained installers stored in a Makeself archive on a CD, when the installer program will later need to unmount the CD and allow a new one to be inserted. This prevents "Filesystem busy" errors for installers that span multiple CDs. - * **`--nox11`** : Disable the automatic spawning of a new terminal in X11. - * **`--nowait`** : When executed from a new X11 terminal, disable the user prompt at the end of the script execution. - * **`--nomd5`** and **`--nocrc`** : Disable the creation of a MD5 / CRC checksum for the archive. This speeds up the extraction process if integrity checking is not necessary. - * **`--sha256`** : Adds a SHA256 checksum for the archive. This is in addition to the MD5 / CRC checksums unless `--nomd5` is also used. - * **`--lsm` _file_** : Provide and LSM file to makeself, that will be embedded in the generated archive. LSM files are describing a software package in a way that is easily parseable. The LSM entry can then be later retrieved using the `--lsm` argument to the archive. An example of a LSM file is provided with Makeself. - * **`--tar-format opt`** : Specify the tar archive format (default is ustar); you may use any value accepted by your tar command (such as posix, v7, etc). - * **`--tar-extra opt`** : Append more options to the tar command line. - - For instance, in order to exclude the `.git` directory from the packaged archive directory using the GNU `tar`, one can use `makeself.sh --tar-extra "--exclude=.git" ...` - - * **`--keep-umask`** : Keep the umask set to shell default, rather than overriding when executing self-extracting archive. - * **`--packaging-date date`** : Use provided string as the packaging date instead of the current date. - * **`--license`** : Append a license file. - * **`--nooverwrite`** : Do not extract the archive if the specified target directory already exists. - * **`--help-header file`** : Add a header to the archive's `--help` output. - * `archive_dir` is the name of the directory that contains the files to be archived - * `file_name` is the name of the archive to be created - * `label` is an arbitrary text string describing the package. It will be displayed while extracting the files. - * `startup_script` is the command to be executed _from within_ the directory of extracted files. Thus, if you wish to execute a program contained in this directory, you must prefix your command with `./`. For example, `./program` will be fine. The `script_args` are additional arguments for this command. - -Here is an example, assuming the user has a package image stored in a **/home/joe/mysoft**, and he wants to generate a self-extracting package named -**mysoft.sh**, which will launch the "setup" script initially stored in /home/joe/mysoft : - -`makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup -` - -Here is also how I created the [makeself.run][9] archive which contains the Makeself distribution : - -`makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" ` - -Archives generated with Makeself can be passed the following arguments: - - * **`--keep`** : Prevent the files to be extracted in a temporary directory that will be removed after the embedded script's execution. The files will then be extracted in the current working directory and will stay here until you remove them. - * **`--verbose`** : Will prompt the user before executing the embedded command - * **`--target dir`** : Allows to extract the archive in an arbitrary place. - * **`--nox11`** : Do not spawn a X11 terminal. - * **`--confirm`** : Prompt the user for confirmation before running the embedded command. - * **`--info`** : Print out general information about the archive (does not extract). - * **`--lsm`** : Print out the LSM entry, if it is present. - * **`--list`** : List the files in the archive. - * **`--check`** : Check the archive for integrity using the embedded checksums. Does not extract the archive. - * **`--nochown`** : By default, a `chown -R` command is run on the target directory after extraction, so that all files belong to the current user. This is mostly needed if you are running as root, as tar will then try to recreate the initial user ownerships. You may disable this behavior with this flag. - * **`--tar`** : Run the tar command on the contents of the archive, using the following arguments as parameter for the command. - * **`--noexec`** : Do not run the embedded script after extraction. - * **`--noexec-cleanup`** : Do not run the embedded cleanup script. - * **`--nodiskspace`** : Do not check for available disk space before attempting to extract. - * **`--cleanup-args`** : Specify arguments to be passed to the cleanup script. Wrap value in quotes to specify multiple arguments. - -Any subsequent arguments to the archive will be passed as additional arguments to the embedded command. You must explicitly use the `--` special command-line construct before any such options to make sure that Makeself will not try to interpret them. - -## Startup Script - -The startup script must be a regular Shell script. - -Within the startup script, you can use the `$USER_PWD` variable to get the path of the folder from which the self-extracting script is executed. This is especially useful to access files that are located in the same folder as the script, as shown in the example below. - -`my-self-extracting-script.sh --fooBarFileParameter foo.bar` - -## Building and Testing - -Clone the git repo and execute `git submodule update --init --recursive` to obtain all submodules. - -* To make a release: `make` -* To run all tests: `make test` - -## Maven Usage - -Makeself is now supported by the following maven plugin [makeself-maven-plugin](https://github.com/hazendaz/makeself-maven-plugin). Please refer to project for usage and report any bugs in regards to maven plugin on that project. - -## License - -Makeself itself is covered by the [GNU General Public License][8] (GPL) version 2 and above. Archives generated by Makeself don't have to be placed under this license (although I encourage it ;-)), since the archive itself is merely data for Makeself. - -## Contributing - -I will gladly consider merging your pull requests on the [GitHub][10] repository. However, please keep the following in mind: - - * One of the main purposes of Makeself is portability. Do not submit patches that will break supported platforms. The more platform-agnostic, the better. - * Please explain clearly what the purpose of the patch is, and how you achieved it. - -## Download - -Get the latest official distribution [here][9] (version 2.4.2). - -The latest development version can be grabbed from [GitHub][10]. Feel free to submit any patches there through the fork and pull request process. - -## Version history - - * **v1.0:** Initial public release - * **v1.1:** The archive can be passed parameters that will be passed on to the embedded script, thanks to John C. Quillan - * **v1.2:** Cosmetic updates, support for bzip2 compression and non-temporary archives. Many ideas thanks to Francois Petitjean. - * **v1.3:** More patches from Bjarni R. Einarsson and Francois Petitjean: Support for no compression (`--nocomp`), script is no longer mandatory, automatic launch in an xterm, optional verbose output, and -target archive option to indicate where to extract the files. - * **v1.4:** Many patches from Francois Petitjean: improved UNIX compatibility, automatic integrity checking, support of LSM files to get info on the package at run time.. - * **v1.5.x:** A lot of bugfixes, and many other patches, including automatic verification through the usage of checksums. Version 1.5.5 was the stable release for a long time, even though the Web page didn't get updated ;-). Makeself was also officially made a part of the [Loki Setup installer][11], and its source is being maintained as part of this package. - * **v2.0:** Complete internal rewrite of Makeself. The command-line parsing was vastly improved, the overall maintenance of the package was greatly improved by separating the stub from makeself.sh. Also Makeself was ported and tested to a variety of Unix platforms. - * **v2.0.1:** First public release of the new 2.0 branch. Prior versions are officially obsoleted. This release introduced the `--copy` argument that was introduced in response to a need for the [UT2K3][12] Linux installer. - * **v2.1.0:** Big change : Makeself can now support multiple embedded tarballs, each stored separately with their own checksums. An existing archive can be updated with the `--append` flag. Checksums are also better managed, and the `--nochown` option for archives appeared. - * **v2.1.1:** Fixes related to the Unix compression (compress command). Some Linux distributions made the insane choice to make it unavailable, even though gzip is capable of uncompressing these files, plus some more bugfixes in the extraction and checksum code. - * **v2.1.2:** Some bug fixes. Use head -n to avoid problems with POSIX conformance. - * **v2.1.3:** Bug fixes with the command line when spawning terminals. Added `--tar`, `--noexec` for archives. Added `--nomd5` and `--nocrc` to avoid creating checksums in archives. The embedded script is now run through "eval". The `--info` output now includes the command used to create the archive. A man page was contributed by Bartosz Fenski. - * **v2.1.4:** Fixed `--info` output. Generate random directory name when extracting files to . to avoid problems. Better handling of errors with wrong permissions for the directory containing the files. Avoid some race conditions, Unset the $CDPATH variable to avoid problems if it is set. Better handling of dot files in the archive directory. - * **v2.1.5:** Made the md5sum detection consistent with the header code. Check for the presence of the archive directory. Added `--encrypt` for symmetric encryption through gpg (Eric Windisch). Added support for the digest command on Solaris 10 for MD5 checksums. Check for available disk space before extracting to the target directory (Andreas Schweitzer). Allow extraction to run asynchronously (patch by Peter Hatch). Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo). - * **v2.1.6:** Replaced one dot per file progress with a realtime progress percentage and a spinning cursor. Added `--noprogress` to prevent showing the progress during the decompression. Added `--target` dir to allow extracting directly to a target directory. (Guy Baconniere) - * **v2.2.0:** First major new release in years! Includes many bugfixes and user contributions. Please look at the [project page on Github][10] for all the details. - * **v2.3.0:** Support for archive encryption via GPG or OpenSSL. Added LZO and LZ4 compression support. Options to set the packaging date and stop the umask from being overriden. Optionally ignore check for available disk space when extracting. New option to check for root permissions before extracting. - * **v2.3.1:** Various compatibility updates. Added unit tests for Travis CI in the GitHub repo. New `--tar-extra`, `--untar-extra`, `--gpg-extra`, `--gpg-asymmetric-encrypt-sign` options. - * **v2.4.0:** Added optional support for SHA256 archive integrity checksums. - * **v2.4.2:** New --cleanup and --cleanup-args arguments for cleanup scripts. Added threading support for supported compressors. Now supports zstd compression. - * **v2.4.3:** Make explicit POSIX tar archives for increased compatibility. - * **v2.4.4:** Fixed various compatibility issues (no longer use POSIX tar archives), Github Actions to check on Solaris and FreeBSD. - * **v2.4.5:** Added `--tar-format` option to set the tar archive format (default is ustar) - -## Links - - * Check out the ["Loki Setup"][11] installer, used to install many Linux games and other applications, and of which I am the co-author. Since the demise of Loki, I am now the official maintainer of the project, and it is now being hosted here on GitHub. - * Bjarni R. Einarsson also wrote the **setup.sh** installer script, inspired by Makeself. [Check it out !][14] - -## Contact - -This script was written by [Stéphane Peter][15] (megastep at megastep.org). Any enhancements and suggestions are welcome. - -Contributions were included from John C. Quillan, Bjarni R. Einarsson, -Francois Petitjean, Ryan C. Gordon, and many contributors on GitHub. If you think I forgot -your name, don't hesitate to contact me. - -This project is now hosted on GitHub. Feel free to submit patches and bug reports on the [project page][10]. - -* * * - -[Stephane Peter][2] - - [1]: http://makeself.io/ - [2]: mailto:megastep@megastep.org - [3]: http://www.idsoftware.com/ - [4]: http://www.lokigames.com/products/myth2/updates.php3 - [5]: http://www.nvidia.com/ - [6]: http://earth.google.com/ - [7]: http://www.virtualbox.org/ - [8]: http://www.gnu.org/copyleft/gpl.html - [9]: https://github.com/megastep/makeself/releases/download/release-2.4.5/makeself-2.4.5.run - [10]: https://github.com/megastep/makeself - [11]: https://github.com/megastep/loki_setup/ - [12]: http://www.unrealtournament2003.com/ - [13]: http://www.icculus.org/ - [14]: http://bre.klaki.net/programs/setup.sh/ - [15]: https://stephanepeter.com/ diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/VERSION b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/VERSION deleted file mode 100755 index 59aa62c1f..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/VERSION +++ /dev/null @@ -1 +0,0 @@ -2.4.5 diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/make-release.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/make-release.sh deleted file mode 100755 index b5692d490..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/make-release.sh +++ /dev/null @@ -1,9 +0,0 @@ -#!/bin/sh -# -# Create a distributable archive of the current version of Makeself - -VER=`cat VERSION` -mkdir -p /tmp/makeself-$VER release -cp -pPR makeself* test README.md COPYING VERSION .gitmodules /tmp/makeself-$VER/ -./makeself.sh --notemp /tmp/makeself-$VER release/makeself-$VER.run "Makeself v$VER" echo "Makeself has extracted itself" - diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself-header.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself-header.sh deleted file mode 100755 index 940903148..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself-header.sh +++ /dev/null @@ -1,660 +0,0 @@ -cat << EOF > "$archname" -#!/bin/bash -# This script was generated using Makeself $MS_VERSION -# The license covering this archive and its contents, if any, is wholly independent of the Makeself license (GPL) -# 2022.3.19-Modified the MS_Help function and some options -# Huawei Technologies Co., Ltd. - -ORIG_UMASK=\`umask\` - -CRCsum="$CRCsum" -MD5="$MD5sum" -SHA="$SHAsum" -SIGNATURE="$Signature" -TMPROOT=\${TMPDIR:="\$HOME"} -if ! test -d "\$TMPROOT"; then - TMPROOT="\$PWD" -fi -export TMPDIR="\$TMPROOT" -USER_PWD="\$PWD" -if ! test -d "\$USER_PWD"; then - exit 1 -fi -export USER_PWD -ARCHIVE_DIR=\`dirname "\$0"\` -export ARCHIVE_DIR - -name_of_file="\$0 " -pwd_of_file="\$PWD" -label="$LABEL" -script="$SCRIPT" -scriptargs="$SCRIPTARGS" -cleanup_script="${CLEANUP_SCRIPT}" -licensetxt="$LICENSE" -helpheader='$HELPHEADER' -targetdir="$archdirname" -filesizes="$filesizes" -totalsize="$totalsize" -keep="$KEEP" -nooverwrite="$NOOVERWRITE" -quiet="n" -accept="n" -nodiskspace="n" -export_conf="$EXPORT_CONF" -decrypt_cmd="$DECRYPT_CMD" -skip="$SKIP" - -print_cmd_arg="" -if type printf > /dev/null; then - print_cmd="printf" -elif test -x /usr/ucb/echo; then - print_cmd="/usr/ucb/echo" -else - print_cmd="echo" -fi - -if test -d /usr/xpg4/bin; then - PATH=/usr/xpg4/bin:\$PATH - export PATH -fi - -if test -d /usr/sfw/bin; then - PATH=\$PATH:/usr/sfw/bin - export PATH -fi - -unset CDPATH - -MS_Printf() -{ - \$print_cmd \$print_cmd_arg "\$1" -} - -MS_PrintLicense() -{ - PAGER=\${PAGER:=more} - if test x"\$licensetxt" != x; then - PAGER_PATH=\`exec <&- 2>&-; which \$PAGER || command -v \$PAGER || type \$PAGER\` - if test -x "\$PAGER_PATH"; then - echo "\$licensetxt" | \$PAGER - else - echo "\$licensetxt" - fi - if test x"\$accept" != xy; then - while true - do - MS_Printf "Please type y to accept, n otherwise: " - read yn - if test x"\$yn" = xn; then - keep=n - eval \$finish; exit 1 - break; - elif test x"\$yn" = xy; then - break; - fi - done - fi - fi -} - -MS_diskspace() -{ - ( - df -kP "\$1" | tail -1 | awk '{ if (\$4 ~ /%/) {print \$3} else {print \$4} }' - ) -} - -MS_dd() -{ - blocks=\`expr \$3 / 1024\` - bytes=\`expr \$3 % 1024\` - # Test for ibs, obs and conv feature - if dd if=/dev/zero of=/dev/null count=1 ibs=512 obs=512 conv=sync 2> /dev/null; then - dd if="\$1" ibs=\$2 skip=1 obs=1024 conv=sync 2> /dev/null | \\ - { test \$blocks -gt 0 && dd ibs=1024 obs=1024 count=\$blocks ; \\ - test \$bytes -gt 0 && dd ibs=1 obs=1024 count=\$bytes ; } 2> /dev/null - else - dd if="\$1" bs=\$2 skip=1 2> /dev/null - fi -} - -MS_dd_Progress() -{ - if test x"\$noprogress" = xy; then - MS_dd "\$@" - return \$? - fi - file="\$1" - offset=\$2 - length=\$3 - pos=0 - bsize=4194304 - while test \$bsize -gt \$length; do - bsize=\`expr \$bsize / 4\` - done - blocks=\`expr \$length / \$bsize\` - bytes=\`expr \$length % \$bsize\` - ( - dd ibs=\$offset skip=1 2>/dev/null - pos=\`expr \$pos \+ \$bsize\` - MS_Printf " 0%% " 1>&2 - if test \$blocks -gt 0; then - while test \$pos -le \$length; do - dd bs=\$bsize count=1 2>/dev/null - pcent=\`expr \$length / 100\` - pcent=\`expr \$pos / \$pcent\` - if test \$pcent -lt 100; then - MS_Printf "\b\b\b\b\b\b\b" 1>&2 - if test \$pcent -lt 10; then - MS_Printf " \$pcent%% " 1>&2 - else - MS_Printf " \$pcent%% " 1>&2 - fi - fi - pos=\`expr \$pos \+ \$bsize\` - done - fi - if test \$bytes -gt 0; then - dd bs=\$bytes count=1 2>/dev/null - fi - MS_Printf "\b\b\b\b\b\b\b" 1>&2 - MS_Printf " 100%% " 1>&2 - ) < "\$file" -} - -MS_Help() -{ - cat << EOH >&2 -Usage: \$0 [options] -Options: - --help | -h Print this message - --info Print embedded info : title, default target directory, embedded script ... - --list Print the list of files in the archive - --check Checks integrity and version dependency of the archive - --quiet Quiet install mode, skip human-computer interactions - --nox11 Do not spawn an xterm - --noexec Do not run embedded script - --extract= Extract directly to a target directory (absolute or relative) - Usually used with --noexec to just extract files without running - --tar arg1 [arg2 ...] Access the contents of the archive through the tar command -\${helpheader} -EOH -} - -MS_Verify_Sig() -{ - GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` - MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` - test -x "\$GPG_PATH" || GPG_PATH=\`exec <&- 2>&-; which gpg || command -v gpg || type gpg\` - test -x "\$MKTEMP_PATH" || MKTEMP_PATH=\`exec <&- 2>&-; which mktemp || command -v mktemp || type mktemp\` - offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` - temp_sig=\`mktemp -t XXXXX\` - echo \$SIGNATURE | base64 --decode > "\$temp_sig" - gpg_output=\`MS_dd "\$1" \$offset \$totalsize | LC_ALL=C "\$GPG_PATH" --verify "\$temp_sig" - 2>&1\` - gpg_res=\$? - rm -f "\$temp_sig" - if test \$gpg_res -eq 0 && test \`echo \$gpg_output | grep -c Good\` -eq 1; then - if test \`echo \$gpg_output | grep -c \$sig_key\` -eq 1; then - test x"\$quiet" = xn && echo "GPG signature is good" >&2 - else - echo "GPG Signature key does not match" >&2 - exit 2 - fi - else - test x"\$quiet" = xn && echo "GPG signature failed to verify" >&2 - exit 2 - fi -} - -MS_Check() -{ - OLD_PATH="\$PATH" - PATH=\${GUESS_MD5_PATH:-"\$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} - MD5_ARG="" - MD5_PATH=\`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum\` - test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which md5 || command -v md5 || type md5\` - test -x "\$MD5_PATH" || MD5_PATH=\`exec <&- 2>&-; which digest || command -v digest || type digest\` - PATH="\$OLD_PATH" - - SHA_PATH=\`exec <&- 2>&-; which shasum || command -v shasum || type shasum\` - test -x "\$SHA_PATH" || SHA_PATH=\`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum\` - - if test x"\$quiet" = xn; then - MS_Printf "Verifying archive integrity..." - fi - offset=\`head -n "\$skip" "\$1" | wc -c | tr -d " "\` - fsize=\`cat "\$1" | wc -c | tr -d " "\` - if test \$totalsize -ne \`expr \$fsize - \$offset\`; then - echo " Unexpected archive size." >&2 - exit 2 - fi - verb=\$2 - i=1 - for s in \$filesizes - do - crc=\`echo \$CRCsum | cut -d" " -f\$i\` - if test -x "\$SHA_PATH"; then - if test x"\`basename \$SHA_PATH\`" = xshasum; then - SHA_ARG="-a 256" - fi - sha=\`echo \$SHA | cut -d" " -f\$i\` - if test x"\$sha" = x0000000000000000000000000000000000000000000000000000000000000000; then - test x"\$verb" = xy && echo " \$1 does not contain an embedded SHA256 checksum." >&2 - else - shasum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$SHA_PATH \$SHA_ARG" | cut -b-64\`; - if test x"\$shasum" != x"\$sha"; then - echo "Error in SHA256 checksums: \$shasum is different from \$sha" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " SHA256 checksums are OK." >&2 - fi - crc="0000000000"; - fi - fi - if test -x "\$MD5_PATH"; then - if test x"\`basename \$MD5_PATH\`" = xdigest; then - MD5_ARG="-a md5" - fi - md5=\`echo \$MD5 | cut -d" " -f\$i\` - if test x"\$md5" = x00000000000000000000000000000000; then - test x"\$verb" = xy && echo " \$1 does not contain an embedded MD5 checksum." >&2 - else - md5sum=\`MS_dd_Progress "\$1" \$offset \$s | eval "\$MD5_PATH \$MD5_ARG" | cut -b-32\`; - if test x"\$md5sum" != x"\$md5"; then - echo "Error in MD5 checksums: \$md5sum is different from \$md5" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " MD5 checksums are OK." >&2 - fi - crc="0000000000"; verb=n - fi - fi - if test x"\$crc" = x0000000000; then - test x"\$verb" = xy && echo " \$1 does not contain a CRC checksum." >&2 - else - sum1=\`MS_dd_Progress "\$1" \$offset \$s | CMD_ENV=xpg4 cksum | awk '{print \$1}'\` - if test x"\$sum1" != x"\$crc"; then - echo "Error in checksums: \$sum1 is different from \$crc" >&2 - exit 2 - elif test x"\$quiet" = xn; then - MS_Printf " CRC checksums are OK." >&2 - fi - fi - i=\`expr \$i + 1\` - offset=\`expr \$offset + \$s\` - done - if test x"\$quiet" = xn; then - echo " All good." - fi -} - -MS_Decompress() -{ - if test x"\$decrypt_cmd" != x""; then - { eval "\$decrypt_cmd" || echo " ... Decryption failed." >&2; } | eval "$GUNZIP_CMD" - else - eval "$GUNZIP_CMD" - fi - - if test \$? -ne 0; then - echo " ... Decompression failed." >&2 - fi -} - -UnTAR() -{ - if test x"\$quiet" = xn; then - tar \$1vf - $UNTAR_EXTRA 2>&1 || { echo " ... Extraction failed." >&2; kill -15 \$$; } - else - tar \$1f - $UNTAR_EXTRA 2>&1 || { echo Extraction failed. >&2; kill -15 \$$; } - fi -} - -MS_exec_cleanup() { - if test x"\$cleanup" = xy && test x"\$cleanup_script" != x""; then - cleanup=n - cd "\$tmpdir" - eval "\"\$cleanup_script\" \$scriptargs \$cleanupargs" - fi -} - -MS_cleanup() -{ - echo 'Signal caught, cleaning up' >&2 - MS_exec_cleanup - cd "\$TMPROOT" - rm -rf "\$tmpdir" - eval \$finish; exit 15 -} - -Script_Args_Check() -{ - script_supported_args=\$(echo \${helpheader} | grep -o -E "\-\-[^ ]+" | awk -F"=" {'print \$1'}) - arg_to_test=\$(echo \$1|awk -F"=" {'print \$1'}) - - for arg in \${script_supported_args}; - do - if test x"\$arg_to_test" = x"\$arg" ;then - return - fi - done - - MS_Help - exit 1 -} - -finish=true -xterm_loop= -noprogress=$NOPROGRESS -nox11=$NOX11 -copy=$COPY -ownership=$OWNERSHIP -verbose=n -cleanup=y -cleanupargs= -sig_key= - -initargs="\$@" - -while [ -n "\$*" ] -do - case "\$1" in - -h | --help) - MS_Help - exit 0 - ;; - -q | --quiet) - quiet=y - noprogress=y - shift - ;; - --info) - echo Identification: "\$label" - echo Target directory: "\$targetdir" - echo Uncompressed size: $USIZE KB - echo Compression: $COMPRESS - if test x"$ENCRYPT" != x""; then - echo Encryption: $ENCRYPT - fi - echo Date of packaging: $DATE - echo Built with Makeself version $MS_VERSION - echo Build command was: "$MS_COMMAND" - if test x"\$script" != x; then - echo Script run after extraction: - echo " " \$script \$scriptargs - fi - if test x"$copy" = xcopy; then - echo "Archive will copy itself to a temporary location" - fi - if test x"$NEED_ROOT" = xy; then - echo "Root permissions required for extraction" - fi - if test x"$KEEP" = xy; then - echo "directory \$targetdir is permanent" - else - echo "\$targetdir will be removed after extraction" - fi - exit 0 - ;; - --list) - echo Target directory: \$targetdir - offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - for s in \$filesizes - do - MS_dd "\$0" \$offset \$s | MS_Decompress | UnTAR t - offset=\`expr \$offset + \$s\` - done - exit 0 - ;; - --tar) - offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - arg1="\$2" - shift 2 || { MS_Help; exit 1; } - for s in \$filesizes - do - MS_dd "\$0" \$offset \$s | MS_Decompress | tar "\$arg1" - "\$@" - offset=\`expr \$offset + \$s\` - done - exit 0 - ;; - --check) - MS_Check "\$0" y - scriptargs="\$scriptargs \$1" - shift - ;; - --noexec) - script="" - cleanup_script="" - shift - ;; - --extract=*) - keep=y - targetdir=\`echo \$1 | cut -d"=" -f2 \` - if ! shift; then MS_Help; exit 1; fi - ;; - --nox11) - nox11=y - shift - ;; - --xwin) - if test "$NOWAIT" = n; then - finish="echo Press Return to close this window...; read junk" - fi - xterm_loop=1 - shift - ;; - --phase2) - copy=phase2 - shift - ;; - --repack | --repack-path=*) - Script_Args_Check \$1 - scriptargs="\$scriptargs '\$1'" - shift - if [[ ! "\$1" =~ ^-.* ]]; then - scriptargs="\$scriptargs '\$1'" - shift - fi - ;; - *) - Script_Args_Check \$1 - scriptargs="\$scriptargs '\$1'" - shift - ;; - esac -done - -quiet_para="" -if test x"\$quiet" = xy; then - quiet_para="--quiet " -fi -scriptargs="--\$name_of_file""--\"\$pwd_of_file\""" \$quiet_para""\$scriptargs" - -if test x"\$quiet" = xy -a x"\$verbose" = xy; then - echo Cannot be verbose and quiet at the same time. >&2 - exit 1 -fi - -if test x"$NEED_ROOT" = xy -a \`id -u\` -ne 0; then - echo "Administrative privileges required for this archive (use su or sudo)" >&2 - exit 1 -fi - -if test x"\$copy" \!= xphase2; then - MS_PrintLicense -fi - -case "\$copy" in -copy) - tmpdir="\$TMPROOT"/makeself.\$RANDOM.\`date +"%y%m%d%H%M%S"\`.\$\$ - mkdir "\$tmpdir" || { - echo "Could not create temporary directory \$tmpdir" >&2 - exit 1 - } - SCRIPT_COPY="\$tmpdir/makeself" - echo "Copying to a temporary location..." >&2 - cp "\$0" "\$SCRIPT_COPY" - chmod +x "\$SCRIPT_COPY" - cd "\$TMPROOT" - exec "\$SCRIPT_COPY" --phase2 -- \$initargs - ;; -phase2) - finish="\$finish ; rm -rf \`dirname \$0\`" - ;; -esac - -if test x"\$nox11" = xn; then - if tty -s; then # Do we have a terminal? - : - else - if test x"\$DISPLAY" != x -a x"\$xterm_loop" = x; then # No, but do we have X? - if xset q > /dev/null 2>&1; then # Check for valid DISPLAY variable - GUESS_XTERMS="xterm gnome-terminal rxvt dtterm eterm Eterm xfce4-terminal lxterminal kvt konsole aterm terminology" - for a in \$GUESS_XTERMS; do - if type \$a >/dev/null 2>&1; then - XTERM=\$a - break - fi - done - chmod a+x \$0 || echo Please add execution rights on \$0 - if test \`echo "\$0" | cut -c1\` = "/"; then # Spawn a terminal! - exec \$XTERM -e "\$0 --xwin \$initargs" - else - exec \$XTERM -e "./\$0 --xwin \$initargs" - fi - fi - fi - fi -fi - -if test x"\$targetdir" = x.; then - tmpdir="." -else - if test x"\$keep" = xy; then - if test x"\$nooverwrite" = xy && test -d "\$targetdir"; then - echo "Target directory \$targetdir already exists, aborting." >&2 - exit 1 - fi - if test x"\$quiet" = xn; then - echo "Creating directory \$targetdir" >&2 - fi - tmpdir="\$targetdir" - dashp="-p" - else - tmpdir="\$TMPROOT/selfgz\$\$\$RANDOM" - dashp="" - fi - mkdir \$dashp "\$tmpdir" || { - echo 'Cannot create target directory' \$tmpdir >&2 - echo 'You should try option --extract=' >&2 - eval \$finish - exit 1 - } -fi - -location="\`pwd\`" -if test x"\$SETUP_NOCHECK" != x1; then - MS_Check "\$0" -fi -offset=\`head -n "\$skip" "\$0" | wc -c | tr -d " "\` - -if test x"\$verbose" = xy; then - MS_Printf "About to extract $USIZE KB in \$tmpdir ... Proceed ? [Y/n] " - read yn - if test x"\$yn" = xn; then - eval \$finish; exit 1 - fi -fi - -if test x"\$quiet" = xn; then - # Decrypting with openssl will ask for password, - # the prompt needs to start on new line - if test x"$ENCRYPT" = x"openssl"; then - echo "Decrypting and uncompressing \$label..." - else - MS_Printf "Uncompressing \$label" - fi -fi -res=3 -if test x"\$keep" = xn; then - trap MS_cleanup 1 2 3 15 -fi - -if test x"\$nodiskspace" = xn; then - leftspace=\`MS_diskspace "\$tmpdir"\` - if test -n "\$leftspace"; then - if test "\$leftspace" -lt $USIZE; then - echo - echo "Not enough space left in "\`dirname \$tmpdir\`" (\$leftspace KB) to decompress \$0 ($USIZE KB)" >&2 - if test x"\$keep" = xn; then - echo "Consider setting TMPDIR to a directory with more free space." - fi - eval \$finish; exit 1 - fi - fi -fi - -for s in \$filesizes -do - if MS_dd_Progress "\$0" \$offset \$s | MS_Decompress | ( cd "\$tmpdir"; umask \$ORIG_UMASK ; UnTAR xp ) 1>/dev/null; then - if test x"\$ownership" = xy; then - (cd "\$tmpdir"; chown -R \`id -u\` .; chgrp -R \`id -g\` .) - fi - else - echo >&2 - echo "Unable to decompress \$0" >&2 - eval \$finish; exit 1 - fi - offset=\`expr \$offset + \$s\` -done -if test x"\$quiet" = xn; then - echo -fi - -cd "\$tmpdir" -res=0 -if test x"\$script" != x; then - if test x"\$export_conf" = x"y"; then - MS_BUNDLE="\$0" - MS_LABEL="\$label" - MS_SCRIPT="\$script" - MS_SCRIPTARGS="\$scriptargs" - MS_ARCHDIRNAME="\$archdirname" - MS_KEEP="\$KEEP" - MS_NOOVERWRITE="\$NOOVERWRITE" - MS_COMPRESS="\$COMPRESS" - MS_CLEANUP="\$cleanup" - export MS_BUNDLE MS_LABEL MS_SCRIPT MS_SCRIPTARGS - export MS_ARCHDIRNAME MS_KEEP MS_NOOVERWRITE MS_COMPRESS - fi - - if test x"\$verbose" = x"y"; then - yn="x" - while test x"\$yn" != x -a x"\$yn" != xy -a x"\$yn" != xY -a x"\$yn" != xn -a x"\$yn" != xN - do - MS_Printf "OK to execute: \$script \$scriptargs \$* ? [Y/n] " - read yn - if test x"\$yn" = x -o x"\$yn" = xy -o x"\$yn" = xY; then - eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$?; - elif test x"\$yn" = xn -o x"\$yn" = xN; then - echo "Unable to decompress \$script ,because of aborting! ";res=\$? - else - echo "Input value is unacceptable,please try again." - fi - done - else - eval "\"\$script\" \$scriptargs \"\\\$@\""; res=\$? - fi - if test "\$res" -ne 0; then - test x"\$verbose" = xy && echo "The program '\$script' returned an error code (\$res)" >&2 - fi -fi - -MS_exec_cleanup - -if test x"\$keep" = xn; then - cd "\$TMPROOT" - rm -rf "\$tmpdir" -fi -eval \$finish; exit \$res -EOF diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.1 b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.1 deleted file mode 100755 index 81bf6e4ff..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.1 +++ /dev/null @@ -1,110 +0,0 @@ -.TH "MAKESELF" "1" "2.4.5" -.SH "NAME" -makeself \- An utility to generate self-extractable archives. -.SH "SYNTAX" -.B makeself [\fIoptions\fP] archive_dir file_name label -.B [\fIstartup_script\fP] [\fIargs\fP] -.SH "DESCRIPTION" -This program is a free (GPL) utility designed to create self-extractable -archives from a directory. -.SH "OPTIONS" -The following options are supported. -.TP 15 -.B -v, --version -Prints out the makeself version number and exits. -.TP -.B -h, --help -Print out help information. -.TP -.B --tar-quietly -Suppress verbose output from the tar command -.TP -.B --quiet -Do not print any messages other than errors -.TP -.B --gzip -Compress using gzip (default if detected). -.TP -.B --bzip2 -Compress using bzip2. -.TP -.B --pbzip2 -Compress using pbzip2. -.TP -.B --xz -Compress using xz. -.TP -.B --lzo -Compress using lzop. -.TP -.B --lz4 -Compress using lz4. -.TP -.B --compress -Compress using the UNIX 'compress' command. -.TP -.B --nocomp -Do not compress the data. -.TP -.B --complevel lvl -Specify the compression level for gzip,bzip2,pbzui2,xz,lzo or lz4 -.TP -.B --notemp -The archive will create archive_dir in the current directory and -uncompress in ./archive_dir. -.TP -.B --copy -Upon extraction, the archive will first copy itself to a temporary directory. -.TP -.B --append -Append more files to an existing makeself archive. The label and startup scripts will then be ignored. -.TP -.B --current -Files will be extracted to the current directory. Both --current and --target dir imply --notemp. -.TP -.B --target dir -Extract directly to a target directory. Directory path can be either absolute or relative. -.TP -.B --header file -Specify location of the header script. -.TP -.B --cleanup file -Specify a cleanup script that executes on interrupt and when finished successfully. -.TP -.B --follow -Follow the symlinks in the archive. -.TP -.B --noprogress -Do not show the progress during the decompression. -.TP -.B --nox11 -Disable automatic spawn of an xterm if running in X11. -.TP -.B --nowait -Do not wait for user input after executing embedded program from an xterm. -.TP -.B --nomd5 -Do not create a MD5 checksum for the archive. -.TP -.B --nocrc -Do not create a CRC32 checksum for the archive. -.TP -.B --lsm file -LSM file describing the package. -.B --packaging-date date -Use provided string as the packaging date instead of the current date. -.SH "EXAMPLES" -Here is an example, assuming the user has a package image stored in a /home/joe/mysoft, -and he wants to generate a self-extracting package named mysoft.sh, which will launch -the "setup" script initially stored in /home/joe/mysoft: -.TP -makeself.sh /home/joe/mysoft mysoft.sh "Joe's Nice Software Package" ./setup -.TP -Here is also how I created the makeself.run archive which contains the Makeself distribution: -.TP -makeself.sh --notemp makeself makeself.run "Makeself by Stephane Peter" echo "Makeself has extracted itself" -.SH "AUTHORS" -Makeself has been written by Stéphane Peter . -.BR -This man page was originally written by Bartosz Fenski for the -Debian GNU/Linux distribution (but it may be used by others). diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.lsm b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.lsm deleted file mode 100755 index 3c4cea8c1..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.lsm +++ /dev/null @@ -1,16 +0,0 @@ -Begin3 -Title: makeself.sh -Version: 2.4.5 -Description: makeself.sh is a shell script that generates a self-extractable - tar.gz archive from a directory. The resulting file appears as a shell - script, and can be launched as is. The archive will then uncompress - itself to a temporary directory and an arbitrary command will be - executed (for example an installation script). This is pretty similar - to archives generated with WinZip Self-Extractor in the Windows world. -Keywords: Installation archive tar winzip -Author: Stephane Peter (megastep@megastep.org) -Maintained-by: Stephane Peter (megastep@megastep.org) -Original-site: https://makeself.io/ -Platform: Unix -Copying-policy: GPL -End diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.sh deleted file mode 100755 index c8ea56597..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/makeself.sh +++ /dev/null @@ -1,822 +0,0 @@ -#!/bin/sh -# -# Makeself version 2.4.x -# by Stephane Peter -# -# Utility to create self-extracting tar.gz archives. -# The resulting archive is a file holding the tar.gz archive with -# a small Shell script stub that uncompresses the archive to a temporary -# directory and then executes a given script from withing that directory. -# -# Makeself home page: https://makeself.io/ -# -# Version 2.0 is a rewrite of version 1.0 to make the code easier to read and maintain. -# -# Version history : -# - 1.0 : Initial public release -# - 1.1 : The archive can be passed parameters that will be passed on to -# the embedded script, thanks to John C. Quillan -# - 1.2 : Package distribution, bzip2 compression, more command line options, -# support for non-temporary archives. Ideas thanks to Francois Petitjean -# - 1.3 : More patches from Bjarni R. Einarsson and Francois Petitjean: -# Support for no compression (--nocomp), script is no longer mandatory, -# automatic launch in an xterm, optional verbose output, and -target -# archive option to indicate where to extract the files. -# - 1.4 : Improved UNIX compatibility (Francois Petitjean) -# Automatic integrity checking, support of LSM files (Francois Petitjean) -# - 1.5 : Many bugfixes. Optionally disable xterm spawning. -# - 1.5.1 : More bugfixes, added archive options -list and -check. -# - 1.5.2 : Cosmetic changes to inform the user of what's going on with big -# archives (Quake III demo) -# - 1.5.3 : Check for validity of the DISPLAY variable before launching an xterm. -# More verbosity in xterms and check for embedded command's return value. -# Bugfix for Debian 2.0 systems that have a different "print" command. -# - 1.5.4 : Many bugfixes. Print out a message if the extraction failed. -# - 1.5.5 : More bugfixes. Added support for SETUP_NOCHECK environment variable to -# bypass checksum verification of archives. -# - 1.6.0 : Compute MD5 checksums with the md5sum command (patch from Ryan Gordon) -# - 2.0 : Brand new rewrite, cleaner architecture, separated header and UNIX ports. -# - 2.0.1 : Added --copy -# - 2.1.0 : Allow multiple tarballs to be stored in one archive, and incremental updates. -# Added --nochown for archives -# Stopped doing redundant checksums when not necesary -# - 2.1.1 : Work around insane behavior from certain Linux distros with no 'uncompress' command -# Cleaned up the code to handle error codes from compress. Simplified the extraction code. -# - 2.1.2 : Some bug fixes. Use head -n to avoid problems. -# - 2.1.3 : Bug fixes with command line when spawning terminals. -# Added --tar for archives, allowing to give arbitrary arguments to tar on the contents of the archive. -# Added --noexec to prevent execution of embedded scripts. -# Added --nomd5 and --nocrc to avoid creating checksums in archives. -# Added command used to create the archive in --info output. -# Run the embedded script through eval. -# - 2.1.4 : Fixed --info output. -# Generate random directory name when extracting files to . to avoid problems. (Jason Trent) -# Better handling of errors with wrong permissions for the directory containing the files. (Jason Trent) -# Avoid some race conditions (Ludwig Nussel) -# Unset the $CDPATH variable to avoid problems if it is set. (Debian) -# Better handling of dot files in the archive directory. -# - 2.1.5 : Made the md5sum detection consistent with the header code. -# Check for the presence of the archive directory -# Added --encrypt for symmetric encryption through gpg (Eric Windisch) -# Added support for the digest command on Solaris 10 for MD5 checksums -# Check for available disk space before extracting to the target directory (Andreas Schweitzer) -# Allow extraction to run asynchronously (patch by Peter Hatch) -# Use file descriptors internally to avoid error messages (patch by Kay Tiong Khoo) -# - 2.1.6 : Replaced one dot per file progress with a realtime progress percentage and a spining cursor (Guy Baconniere) -# Added --noprogress to prevent showing the progress during the decompression (Guy Baconniere) -# Added --target dir to allow extracting directly to a target directory (Guy Baconniere) -# - 2.2.0 : Many bugfixes, updates and contributions from users. Check out the project page on Github for the details. -# - 2.3.0 : Option to specify packaging date to enable byte-for-byte reproducibility. (Marc Pawlowsky) -# - 2.4.0 : Optional support for SHA256 checksums in archives. -# - 2.4.2 : Add support for threads for several compressors. (M. Limber) -# Added zstd support. -# - 2.4.3 : Make explicit POSIX tar archives for increased compatibility. -# - 2.4.5 : Added --tar-format to override ustar tar archive format -# -# (C) 1998-2021 by Stephane Peter -# -# This software is released under the terms of the GNU GPL version 2 and above -# Please read the license at http://www.gnu.org/copyleft/gpl.html -# Self-extracting archives created with this script are explictly NOT released under the term of the GPL -# - -MS_VERSION=2.4.5 -MS_COMMAND="$0" -unset CDPATH - -for f in ${1+"$@"}; do - MS_COMMAND="$MS_COMMAND \\\\ - \\\"$f\\\"" -done - -# For Solaris systems -if test -d /usr/xpg4/bin; then - PATH=/usr/xpg4/bin:$PATH - export PATH -fi - -# Procedures - -MS_Usage() -{ - echo "Usage: $0 [args] archive_dir file_name label startup_script [script_args]" - echo "args can be one or more of the following :" - echo " --version | -v : Print out Makeself version number and exit" - echo " --help | -h : Print out this help message" - echo " --tar-quietly : Suppress verbose output from the tar command" - echo " --quiet | -q : Do not print any messages other than errors." - echo " --gzip : Compress using gzip (default if detected)" - echo " --pigz : Compress with pigz" - echo " --zstd : Compress with zstd" - echo " --bzip2 : Compress using bzip2 instead of gzip" - echo " --pbzip2 : Compress using pbzip2 instead of gzip" - echo " --xz : Compress using xz instead of gzip" - echo " --lzo : Compress using lzop instead of gzip" - echo " --lz4 : Compress using lz4 instead of gzip" - echo " --compress : Compress using the UNIX 'compress' command" - echo " --complevel lvl : Compression level for gzip pigz zstd xz lzo lz4 bzip2 and pbzip2 (default 9)" - echo " --threads thds : Number of threads to be used by compressors that support parallelization." - echo " Omit to use compressor's default. Most useful (and required) for opting" - echo " into xz's threading, usually with '--threads=0' for all available cores." - echo " pbzip2 and pigz are parallel by default, and setting this value allows" - echo " limiting the number of threads they use." - echo " --base64 : Instead of compressing, encode the data using base64" - echo " --gpg-encrypt : Instead of compressing, encrypt the data using GPG" - echo " --gpg-asymmetric-encrypt-sign" - echo " : Instead of compressing, asymmetrically encrypt and sign the data using GPG" - echo " --gpg-extra opt : Append more options to the gpg command line" - echo " --ssl-encrypt : Instead of compressing, encrypt the data using OpenSSL" - echo " --ssl-passwd pass : Use the given password to encrypt the data using OpenSSL" - echo " --ssl-pass-src src : Use the given src as the source of password to encrypt the data" - echo " using OpenSSL. See \"PASS PHRASE ARGUMENTS\" in man openssl." - echo " If this option is not supplied, the user will be asked to enter" - echo " encryption password on the current terminal." - echo " --ssl-no-md : Do not use \"-md\" option not supported by older OpenSSL." - echo " --nochown : Do not give the target folder to the current user (default)" - echo " --chown : Give the target folder to the current user recursively" - echo " --nocomp : Do not compress the data" - echo " --notemp : The archive will create archive_dir in the" - echo " current directory and uncompress in ./archive_dir" - echo " --needroot : Check that the root user is extracting the archive before proceeding" - echo " --copy : Upon extraction, the archive will first copy itself to" - echo " a temporary directory" - echo " --append : Append more files to an existing Makeself archive" - echo " The label and startup scripts will then be ignored" - echo " --target dir : Extract directly to a target directory" - echo " directory path can be either absolute or relative" - echo " --nooverwrite : Do not extract the archive if the specified target directory exists" - echo " --current : Files will be extracted to the current directory" - echo " Both --current and --target imply --notemp" - echo " --tar-format opt : Specify a tar archive format (default is ustar)" - echo " --tar-extra opt : Append more options to the tar command line" - echo " --untar-extra opt : Append more options to the during the extraction of the tar archive" - echo " --nomd5 : Don't calculate an MD5 for archive" - echo " --nocrc : Don't calculate a CRC for archive" - echo " --sha256 : Compute a SHA256 checksum for the archive" - echo " --header file : Specify location of the header script" - echo " --cleanup file : Specify a cleanup script that executes on interrupt and when finished successfully." - echo " --follow : Follow the symlinks in the archive" - echo " --noprogress : Do not show the progress during the decompression" - echo " --nox11 : Disable automatic spawn of a xterm" - echo " --nowait : Do not wait for user input after executing embedded" - echo " program from an xterm" - echo " --sign passphrase : Signature private key to sign the package with" - echo " --lsm file : LSM file describing the package" - echo " --license file : Append a license file" - echo " --help-header file : Add a header to the archive's --help output" - echo " --packaging-date date" - echo " : Use provided string as the packaging date" - echo " instead of the current date." - echo - echo " --keep-umask : Keep the umask set to shell default, rather than overriding when executing self-extracting archive." - echo " --export-conf : Export configuration variables to startup_script" - echo - echo "Do not forget to give a fully qualified startup script name" - echo "(i.e. with a ./ prefix if inside the archive)." - exit 1 -} - -# Default settings -if type gzip >/dev/null 2>&1; then - COMPRESS=gzip -elif type compress >/dev/null 2>&1; then - COMPRESS=compress -else - echo "ERROR: missing commands: gzip, compress" >&2 - MS_Usage -fi -ENCRYPT=n -PASSWD="" -PASSWD_SRC="" -OPENSSL_NO_MD=n -COMPRESS_LEVEL=9 -DEFAULT_THREADS=123456 # Sentinel value -THREADS=$DEFAULT_THREADS -KEEP=n -CURRENT=n -NOX11=n -NOWAIT=n -APPEND=n -TAR_QUIETLY=n -KEEP_UMASK=n -QUIET=n -NOPROGRESS=n -COPY=none -NEED_ROOT=n -TAR_ARGS=rvf -TAR_FORMAT=ustar -TAR_EXTRA="" -GPG_EXTRA="" -DU_ARGS=-ks -HEADER=`dirname "$0"`/makeself-header.sh -SIGNATURE="" -TARGETDIR="" -NOOVERWRITE=n -DATE=`LC_ALL=C date` -EXPORT_CONF=n -SHA256=n -OWNERSHIP=n -SIGN=n -GPG_PASSPHRASE="" - -# LSM file stuff -LSM_CMD="echo No LSM. >> \"\$archname\"" - -while true -do - case "$1" in - --version | -v) - echo Makeself version $MS_VERSION - exit 0 - ;; - --pbzip2) - COMPRESS=pbzip2 - shift - ;; - --bzip2) - COMPRESS=bzip2 - shift - ;; - --gzip) - COMPRESS=gzip - shift - ;; - --pigz) - COMPRESS=pigz - shift - ;; - --zstd) - COMPRESS=zstd - shift - ;; - --xz) - COMPRESS=xz - shift - ;; - --lzo) - COMPRESS=lzo - shift - ;; - --lz4) - COMPRESS=lz4 - shift - ;; - --compress) - COMPRESS=compress - shift - ;; - --base64) - COMPRESS=base64 - shift - ;; - --gpg-encrypt) - COMPRESS=gpg - shift - ;; - --gpg-asymmetric-encrypt-sign) - COMPRESS=gpg-asymmetric - shift - ;; - --gpg-extra) - GPG_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-encrypt) - ENCRYPT=openssl - shift - ;; - --ssl-passwd) - PASSWD=$2 - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-pass-src) - PASSWD_SRC=$2 - shift 2 || { MS_Usage; exit 1; } - ;; - --ssl-no-md) - OPENSSL_NO_MD=y - shift - ;; - --nocomp) - COMPRESS=none - shift - ;; - --complevel) - COMPRESS_LEVEL="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --threads) - THREADS="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --nochown) - OWNERSHIP=n - shift - ;; - --chown) - OWNERSHIP=y - shift - ;; - --notemp) - KEEP=y - shift - ;; - --copy) - COPY=copy - shift - ;; - --current) - CURRENT=y - KEEP=y - shift - ;; - --tar-format) - TAR_FORMAT="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --tar-extra) - TAR_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --untar-extra) - UNTAR_EXTRA="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --target) - TARGETDIR="$2" - KEEP=y - shift 2 || { MS_Usage; exit 1; } - ;; - --sign) - SIGN=y - GPG_PASSPHRASE="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --nooverwrite) - NOOVERWRITE=y - shift - ;; - --needroot) - NEED_ROOT=y - shift - ;; - --header) - HEADER="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --cleanup) - CLEANUP_SCRIPT="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --license) - # We need to escape all characters having a special meaning in double quotes - LICENSE=$(sed 's/\\/\\\\/g; s/"/\\\"/g; s/`/\\\`/g; s/\$/\\\$/g' "$2") - shift 2 || { MS_Usage; exit 1; } - ;; - --follow) - TAR_ARGS=rvhf - DU_ARGS=-ksL - shift - ;; - --noprogress) - NOPROGRESS=y - shift - ;; - --nox11) - NOX11=y - shift - ;; - --nowait) - NOWAIT=y - shift - ;; - --nomd5) - NOMD5=y - shift - ;; - --sha256) - SHA256=y - shift - ;; - --nocrc) - NOCRC=y - shift - ;; - --append) - APPEND=y - shift - ;; - --lsm) - LSM_CMD="cat \"$2\" >> \"\$archname\"" - shift 2 || { MS_Usage; exit 1; } - ;; - --packaging-date) - DATE="$2" - shift 2 || { MS_Usage; exit 1; } - ;; - --help-header) - HELPHEADER=`sed -e "s/'/'\\\\\''/g" $2` - shift 2 || { MS_Usage; exit 1; } - [ -n "$HELPHEADER" ] && HELPHEADER="$HELPHEADER -" - ;; - --tar-quietly) - TAR_QUIETLY=y - shift - ;; - --keep-umask) - KEEP_UMASK=y - shift - ;; - --export-conf) - EXPORT_CONF=y - shift - ;; - -q | --quiet) - QUIET=y - shift - ;; - -h | --help) - MS_Usage - ;; - -*) - echo Unrecognized flag : "$1" - MS_Usage - ;; - *) - break - ;; - esac -done - -if test $# -lt 1; then - MS_Usage -else - if test -d "$1"; then - archdir="$1" - else - echo "Directory $1 does not exist." >&2 - exit 1 - fi -fi -archname="$2" - -if test "$QUIET" = "y" || test "$TAR_QUIETLY" = "y"; then - if test "$TAR_ARGS" = "rvf"; then - TAR_ARGS="rf" - elif test "$TAR_ARGS" = "rvhf"; then - TAR_ARGS="rhf" - fi -fi - -if test "$APPEND" = y; then - if test $# -lt 2; then - MS_Usage - fi - - # Gather the info from the original archive - OLDENV=`sh "$archname" --dumpconf` - if test $? -ne 0; then - echo "Unable to update archive: $archname" >&2 - exit 1 - else - eval "$OLDENV" - OLDSKIP=`expr $SKIP + 1` - fi -else - if test "$KEEP" = n -a $# = 3; then - echo "ERROR: Making a temporary archive with no embedded command does not make sense!" >&2 - echo >&2 - MS_Usage - fi - # We don't want to create an absolute directory unless a target directory is defined - if test "$CURRENT" = y; then - archdirname="." - elif test x"$TARGETDIR" != x; then - archdirname="$TARGETDIR" - else - archdirname=`basename "$1"` - fi - - if test $# -lt 3; then - MS_Usage - fi - - LABEL="$3" - SCRIPT="$4" - test "x$SCRIPT" = x || shift 1 - shift 3 - SCRIPTARGS="$*" -fi - -if test "$KEEP" = n -a "$CURRENT" = y; then - echo "ERROR: It is A VERY DANGEROUS IDEA to try to combine --notemp and --current." >&2 - exit 1 -fi - -case $COMPRESS in -gzip) - GZIP_CMD="gzip -c$COMPRESS_LEVEL" - GUNZIP_CMD="gzip -cd" - ;; -pigz) - GZIP_CMD="pigz -$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD --processes $THREADS" - fi - GUNZIP_CMD="gzip -cd" - ;; -zstd) - GZIP_CMD="zstd -$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD --threads=$THREADS" - fi - GUNZIP_CMD="zstd -cd" - ;; -pbzip2) - GZIP_CMD="pbzip2 -c$COMPRESS_LEVEL" - if test $THREADS -ne $DEFAULT_THREADS; then # Leave as the default if threads not indicated - GZIP_CMD="$GZIP_CMD -p$THREADS" - fi - GUNZIP_CMD="bzip2 -d" - ;; -bzip2) - GZIP_CMD="bzip2 -$COMPRESS_LEVEL" - GUNZIP_CMD="bzip2 -d" - ;; -xz) - GZIP_CMD="xz -c$COMPRESS_LEVEL" - # Must opt-in by specifying a value since not all versions of xz support threads - if test $THREADS -ne $DEFAULT_THREADS; then - GZIP_CMD="$GZIP_CMD --threads=$THREADS" - fi - GUNZIP_CMD="xz -d" - ;; -lzo) - GZIP_CMD="lzop -c$COMPRESS_LEVEL" - GUNZIP_CMD="lzop -d" - ;; -lz4) - GZIP_CMD="lz4 -c$COMPRESS_LEVEL" - GUNZIP_CMD="lz4 -d" - ;; -base64) - GZIP_CMD="base64" - GUNZIP_CMD="base64 --decode -i -" - ;; -gpg) - GZIP_CMD="gpg $GPG_EXTRA -ac -z$COMPRESS_LEVEL" - GUNZIP_CMD="gpg -d" - ENCRYPT="gpg" - ;; -gpg-asymmetric) - GZIP_CMD="gpg $GPG_EXTRA -z$COMPRESS_LEVEL -es" - GUNZIP_CMD="gpg --yes -d" - ENCRYPT="gpg" - ;; -compress) - GZIP_CMD="compress -fc" - GUNZIP_CMD="(type compress >/dev/null 2>&1 && compress -fcd || gzip -cd)" - ;; -none) - GZIP_CMD="cat" - GUNZIP_CMD="cat" - ;; -esac - -if test x"$ENCRYPT" = x"openssl"; then - if test x"$APPEND" = x"y"; then - echo "Appending to existing archive is not compatible with OpenSSL encryption." >&2 - fi - - ENCRYPT_CMD="openssl enc -aes-256-cbc -salt" - DECRYPT_CMD="openssl enc -aes-256-cbc -d" - - if test x"$OPENSSL_NO_MD" != x"y"; then - ENCRYPT_CMD="$ENCRYPT_CMD -md sha256" - DECRYPT_CMD="$DECRYPT_CMD -md sha256" - fi - - if test -n "$PASSWD_SRC"; then - ENCRYPT_CMD="$ENCRYPT_CMD -pass $PASSWD_SRC" - elif test -n "$PASSWD"; then - ENCRYPT_CMD="$ENCRYPT_CMD -pass pass:$PASSWD" - fi -fi - -tmpfile="${TMPDIR:-/tmp}/mkself$$" - -if test -f "$HEADER"; then - oldarchname="$archname" - archname="$tmpfile" - # Generate a fake header to count its lines - SKIP=0 - . "$HEADER" - SKIP=`cat "$tmpfile" |wc -l` - # Get rid of any spaces - SKIP=`expr $SKIP` - rm -f "$tmpfile" - if test "$QUIET" = "n"; then - echo "Header is $SKIP lines long" >&2 - fi - archname="$oldarchname" -else - echo "Unable to open header file: $HEADER" >&2 - exit 1 -fi - -if test "$QUIET" = "n"; then - echo -fi - -if test "$APPEND" = n; then - if test -f "$archname"; then - echo "WARNING: Overwriting existing file: $archname" >&2 - fi -fi - -USIZE=`du $DU_ARGS "$archdir" | awk '{print $1}'` - -if test "." = "$archdirname"; then - if test "$KEEP" = n; then - archdirname="makeself-$$-`date +%Y%m%d%H%M%S`" - fi -fi - -test -d "$archdir" || { echo "Error: $archdir does not exist."; rm -f "$tmpfile"; exit 1; } -if test "$QUIET" = "n"; then - echo "About to compress $USIZE KB of data..." - echo "Adding files to archive named \"$archname\"..." -fi - -# See if we have GNU tar -TAR=`exec <&- 2>&-; which gtar || command -v gtar || type gtar` -test -x "$TAR" || TAR=tar - -tmparch="${TMPDIR:-/tmp}/mkself$$.tar" -( - if test "$APPEND" = "y"; then - tail -n "+$OLDSKIP" "$archname" | eval "$GUNZIP_CMD" > "$tmparch" - fi - cd "$archdir" - # "Determining if a directory is empty" - # https://www.etalabs.net/sh_tricks.html - find . \ - \( \ - ! -type d \ - -o \ - \( -links 2 -exec sh -c ' - is_empty () ( - cd "$1" - set -- .[!.]* ; test -f "$1" && return 1 - set -- ..?* ; test -f "$1" && return 1 - set -- * ; test -f "$1" && return 1 - return 0 - ) - is_empty "$0"' {} \; \ - \) \ - \) -print \ - | LC_ALL=C sort \ - | sed 's/./\\&/g' \ - | xargs $TAR $TAR_EXTRA --format $TAR_FORMAT -$TAR_ARGS "$tmparch" -) || { - echo "ERROR: failed to create temporary archive: $tmparch" - rm -f "$tmparch" "$tmpfile" - exit 1 -} - -USIZE=`du $DU_ARGS "$tmparch" | awk '{print $1}'` - -eval "$GZIP_CMD" <"$tmparch" >"$tmpfile" || { - echo "ERROR: failed to create temporary file: $tmpfile" - rm -f "$tmparch" "$tmpfile" - exit 1 -} -rm -f "$tmparch" - -if test x"$ENCRYPT" = x"openssl"; then - echo "About to encrypt archive \"$archname\"..." - { eval "$ENCRYPT_CMD -in $tmpfile -out ${tmpfile}.enc" && mv -f ${tmpfile}.enc $tmpfile; } || \ - { echo Aborting: could not encrypt temporary file: "$tmpfile".; rm -f "$tmpfile"; exit 1; } -fi - -fsize=`cat "$tmpfile" | wc -c | tr -d " "` - -# Compute the checksums - -shasum=0000000000000000000000000000000000000000000000000000000000000000 -md5sum=00000000000000000000000000000000 -crcsum=0000000000 - -if test "$NOCRC" = y; then - if test "$QUIET" = "n"; then - echo "skipping crc at user request" - fi -else - crcsum=`CMD_ENV=xpg4 cksum < "$tmpfile" | sed -e 's/ /Z/' -e 's/ /Z/' | cut -dZ -f1` - if test "$QUIET" = "n"; then - echo "CRC: $crcsum" - fi -fi - -if test "$SHA256" = y; then - SHA_PATH=`exec <&- 2>&-; which shasum || command -v shasum || type shasum` - if test -x "$SHA_PATH"; then - shasum=`eval "$SHA_PATH -a 256" < "$tmpfile" | cut -b-64` - else - SHA_PATH=`exec <&- 2>&-; which sha256sum || command -v sha256sum || type sha256sum` - shasum=`eval "$SHA_PATH" < "$tmpfile" | cut -b-64` - fi - if test "$QUIET" = "n"; then - if test -x "$SHA_PATH"; then - echo "SHA256: $shasum" - else - echo "SHA256: none, SHA command not found" - fi - fi -fi -if test "$NOMD5" = y; then - if test "$QUIET" = "n"; then - echo "Skipping md5sum at user request" - fi -else - # Try to locate a MD5 binary - OLD_PATH=$PATH - PATH=${GUESS_MD5_PATH:-"$OLD_PATH:/bin:/usr/bin:/sbin:/usr/local/ssl/bin:/usr/local/bin:/opt/openssl/bin"} - MD5_ARG="" - MD5_PATH=`exec <&- 2>&-; which md5sum || command -v md5sum || type md5sum` - test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which md5 || command -v md5 || type md5` - test -x "$MD5_PATH" || MD5_PATH=`exec <&- 2>&-; which digest || command -v digest || type digest` - PATH=$OLD_PATH - if test -x "$MD5_PATH"; then - if test `basename ${MD5_PATH}`x = digestx; then - MD5_ARG="-a md5" - fi - md5sum=`eval "$MD5_PATH $MD5_ARG" < "$tmpfile" | cut -b-32` - if test "$QUIET" = "n"; then - echo "MD5: $md5sum" - fi - else - if test "$QUIET" = "n"; then - echo "MD5: none, MD5 command not found" - fi - fi -fi -if test "$SIGN" = y; then - GPG_PATH=`exec <&- 2>&-; which gpg || command -v gpg || type gpg` - if test -x "$GPG_PATH"; then - SIGNATURE=`$GPG_PATH --pinentry-mode=loopback --batch --yes --passphrase "$GPG_PASSPHRASE" --output - --detach-sig $tmpfile | base64 | tr -d \\\\n` - if test "$QUIET" = "n"; then - echo "Signature: $SIGNATURE" - fi - else - echo "Missing gpg command" >&2 - fi -fi - -totalsize=0 -for size in $fsize; -do - totalsize=`expr $totalsize + $size` -done - -if test "$APPEND" = y; then - mv "$archname" "$archname".bak || exit - - # Prepare entry for new archive - filesizes="$fsize" - CRCsum="$crcsum" - MD5sum="$md5sum" - SHAsum="$shasum" - Signature="$SIGNATURE" - # Generate the header - . "$HEADER" - # Append the new data - cat "$tmpfile" >> "$archname" - - chmod +x "$archname" - rm -f "$archname".bak - if test "$QUIET" = "n"; then - echo "Self-extractable archive \"$archname\" successfully updated." - fi -else - filesizes="$fsize" - CRCsum="$crcsum" - MD5sum="$md5sum" - SHAsum="$shasum" - Signature="$SIGNATURE" - - # Generate the header - . "$HEADER" - - # Append the compressed tar data after the stub - if test "$QUIET" = "n"; then - echo - fi - cat "$tmpfile" >> "$archname" - chmod +x "$archname" - if test "$QUIET" = "n"; then - echo Self-extractable archive \"$archname\" successfully created. - fi -fi -rm -f "$tmpfile" diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/run-tests.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/run-tests.sh deleted file mode 100755 index 31ee16511..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/makeself/run-tests.sh +++ /dev/null @@ -1,8 +0,0 @@ -#!/bin/sh -# Run every available test - Bash needed -cd test -for test in *test; -do - echo "Running test $test ..." - bash $test || { echo "*** ERROR: Test '$test' failed!"; exit 1; } -done diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/merge_aicpu_info_json.sh b/atb_operator/AtbTrilSample/TrilOp/cmake/util/merge_aicpu_info_json.sh deleted file mode 100755 index a977bd51d..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/merge_aicpu_info_json.sh +++ /dev/null @@ -1,31 +0,0 @@ -#!/bin/bash - -project_path=$1 -build_path=$2 -vendor_name=customize -echo $@ -if [[ ! -d "$project_path" ]]; then - echo "[ERROR] No projcet path is provided" - exit 1 -fi - -if [[ ! -d "$build_path" ]]; then - echo "[ERROR] No build path is provided" - exit 1 -fi - -if [[ ! -d "$ASCEND_OPP_PATH" ]]; then - echo "[ERROR] No opp install path is provided" - exit 1 -fi -custom_exist_info_json=$ASCEND_OPP_PATH/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json -custom_new_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/cust_aicpu_kernel.json -temp_info_json=$build_path/makepkg/packages/vendors/$vendor_name/op_impl/cpu/config/temp_cust_aicpu_kernel.json - -if [[ -f "$custom_exist_info_json" ]] && [[ -f "$custom_new_info_json" ]]; then - cp -f $custom_exist_info_json $temp_info_json - chmod +w $temp_info_json - python3 ${project_path}/cmake/util/insert_op_info.py ${custom_new_info_json} ${temp_info_json} - cp -f $temp_info_json $custom_new_info_json - rm -f $temp_info_json -fi diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/opdesc_parser.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/opdesc_parser.py deleted file mode 100755 index bc39462b8..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/opdesc_parser.py +++ /dev/null @@ -1,277 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import sys -import os - - -OP_ALL = '__ALLOP__' -SOC_ALL = '__ALLSOC__' -SOC_TO_SHORT_SOC_MAP = { - "ascend910a": "ascend910", - "ascend910proa": "ascend910", - "ascend910b": "ascend910", - "ascend910prob": "ascend910", - "ascend910premiuma": "ascend910", - "ascend910b1": "ascend910b", - "ascend910b2": "ascend910b", - "ascend910b2c": "ascend910b", - "ascend910b3": "ascend910b", - "ascend910b4": "ascend910b", - "ascend910c1": "ascend910c", - "ascend910c2": "ascend910c", - "ascend910c3": "ascend910c", - "ascend910c4": "ascend910c", - "ascend310p1": "ascend310p", - "ascend310p3": "ascend310p", - "ascend310p3vir01": "ascend310p", - "ascend310p3vir02": "ascend310p", - "ascend310p3vir04": "ascend310p", - "ascend310p3vir08": "ascend310p", - "ascend310b1": "ascend310b", - "bs9sx1aa": "bs9sx1a" -} - - -class OpDesc: - def __init__(self: any, op_type: str): - self.op_type = op_type - self.attr_list = [] - self.attr_val = {} - self.input_name = [] - self.input_type = [] - self.input_dtype = [] - self.input_fmt = [] - self.output_name = [] - self.output_type = [] - self.output_dtype = [] - self.output_fmt = [] - self.op_fmt_sel = False - self.op_chk_support = False - self.op_intf = '' - self.kern_name = '' - self.op_file = '' - self.op_replay_flag = False - self.op_replay_batch = False - self.input_idx = -1 - self.output_idx = -1 - self.max_block_dim = 32 - self.max_shape_size = 268435456 - self.dynamic_shape = False - self.op_range_limit = '' - self.custom_compile_options = {} - self.custom_all_compile_options = {} - - @staticmethod - def _parse_digit(conf: str) -> int: - return int(conf.split('=')[1]) - - @staticmethod - def _parse_flag(conf: str) -> bool: - if 'true' == conf.split('=')[1]: - return True - return False - - @staticmethod - def _parse_str(conf: str) -> str: - return conf.split('=')[1] - - @staticmethod - def _parse_list(conf: str) -> list: - return conf.split('=')[1].split(',') - - def parse_input(self: any, conf: str): - if conf.startswith('input{}.name'.format(int(self.input_idx) + 1)): - self.input_idx += 1 - self.input_name.append(self._parse_str(conf) + '_in__') - elif conf.startswith('input{}.paramType'.format(int(self.input_idx))): - self.input_type.append(self._parse_str(conf)) - elif conf.startswith('input{}.dtype'.format(int(self.input_idx))): - self.input_dtype.append(self._parse_str(conf)) - elif conf.startswith('input{}.format'.format(int(self.input_idx))): - self.input_fmt.append(self._parse_str(conf)) - else: - return - - def parse_output(self: any, conf: str): - if conf.startswith('output{}.name'.format(int(self.output_idx) + 1)): - self.output_idx += 1 - self.output_name.append(self._parse_str(conf) + '_out_') - elif conf.startswith('output{}.paramType'.format(int(self.output_idx))): - self.output_type.append(self._parse_str(conf)) - elif conf.startswith('output{}.dtype'.format(int(self.output_idx))): - self.output_dtype.append(self._parse_str(conf)) - elif conf.startswith('output{}.format'.format(int(self.output_idx))): - self.output_fmt.append(self._parse_str(conf)) - else: - return - - def parse_op_format(self: any, conf: str): - self.op_fmt_sel = self._parse_flag(conf) - - def parse_check_support(self: any, conf: str): - self.op_chk_support = self._parse_flag(conf) - - def parse_range_limit(self: any, conf: str): - self.op_range_limit = self._parse_str(conf) - - def parse_kern_name(self: any, conf: str): - self.kern_name = self._parse_str(conf) - - def parse_op_intf(self: any, conf: str): - self.op_intf = self._parse_str(conf) - - def parse_op_file(self: any, conf: str): - self.op_file = self._parse_str(conf) - - def parse_dynamic_shape(self: any, conf: str): - self.dynamic_shape = self._parse_flag(conf) - - def parse_attr_list(self: any, conf: str): - self.attr_list = self._parse_list(conf) - - @staticmethod - def _camel_to_snake(camel_case_str: str): - snake_case_str = '' - for i, c in enumerate(camel_case_str): - if i == 0: - snake_case_str += c.lower() - elif c.isupper(): - snake_case_str += '_' + c.lower() - else: - snake_case_str += c - return snake_case_str - - def parse_attr_val(self: any, conf: str): - for attr in self.attr_list: - if self.attr_val.get(attr) is None: - self.attr_val[attr] = {} - if conf.startswith('attr_{}.type'.format(attr)): - self.attr_val.get(attr)['type'] = self._camel_to_snake(self._parse_str(conf)) - elif conf.startswith('attr_{}.paramType'.format(attr)): - self.attr_val.get(attr)['paramType'] = self._parse_str(conf) - elif conf.startswith('attr_{}.defaultValue'.format(attr)): - self.attr_val.get(attr)['defaultValue'] = self._parse_str(conf) - - def parse_replay_val(self: any, batch_list: list, iterator_list: list): - if self.op_type in batch_list: - self.op_replay_flag = True - self.op_replay_batch = True - elif self.op_type in iterator_list: - self.op_replay_flag = True - self.op_replay_batch = False - - -def _is_op_type_in_opdesc(op_descs: list, op_type: str): - for op in op_descs: - if op_type == op.op_type: - return True - return False - - -def _set_all_options_to_opdescs(op_descs, soc_ver_compile_options): - for op in op_descs: - op.custom_all_compile_options = soc_ver_compile_options - - -def _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options): - for op in op_descs: - if op.op_type != op_type: - continue - op.custom_compile_options = soc_ver_compile_options - - -def _trans_soc_ver_to_short(soc_ver: str): - low_soc_ver = soc_ver.lower() - if low_soc_ver not in SOC_TO_SHORT_SOC_MAP: - print(f'WARNING: caution: {soc_ver} will trans into ascend910, if not your intention,' - f'use ascend910b1~4 instead') - return SOC_TO_SHORT_SOC_MAP[low_soc_ver] - - -def _get_op_custom_options(op_descs: list, auto_gen_dir: str): - if auto_gen_dir is None: - return {} - file = os.path.join(auto_gen_dir, "custom_compile_options.ini") - if not os.path.exists(file): - print(f'WARNING: cannot find {auto_gen_dir}/custom_compile_options.ini') - return {} - with open (file, 'r') as fd: - lines = fd.readlines() - for line in lines: - param_list = str.split(line.rstrip('\n'), ',') - if len(param_list) != 3: - raise Exception(f'ERROR: custom compile option {param_list} len is not 3') - op_type = param_list[0] - if op_type.upper() == 'ALL': - op_type = OP_ALL - if op_type != OP_ALL and _is_op_type_in_opdesc(op_descs, op_type) == False: - print(f'WARNING: op: {op_type} are not exists in this project') - continue - soc_ver_compile_options = {} - soc_ver = param_list[1] - options_str = param_list[2] - options = str.split(options_str, ';') - if soc_ver == '': - soc_ver_compile_options[SOC_ALL] = options - else: - soc_ver_list = str.split(soc_ver, ';') - for ver in soc_ver_list: - short_ver = _trans_soc_ver_to_short(ver) - soc_ver_compile_options[short_ver] = options - if op_type == OP_ALL: - _set_all_options_to_opdescs(op_descs, soc_ver_compile_options) - else: - _set_options_to_opdesc(op_descs, op_type, soc_ver_compile_options) - - -def get_op_desc(file: str, batch_list: list, iterator_list: list, builder: any, - op_type: list, auto_gen_dir: str = None) -> list: - op_descs = [] - op_match = False - with open (file, 'r') as fd: - lines = fd.readlines() - for line in lines: - line = line.strip() - if line.startswith('['): - name = line[1:-1] - if op_type is None or name in op_type: - op_match = True - op_desc = builder(name) - op_desc.parse_replay_val(batch_list, iterator_list) - op_descs.append(op_desc) - else: - op_match = False - if op_type is not None and len(op_descs) == len(op_type): - return op_descs - continue - if not op_match: - continue - if line.startswith('input'): - op_desc.parse_input(line) - elif line.startswith('output'): - op_desc.parse_output(line) - elif line.startswith('dynamicFormat.flag'): - op_desc.parse_op_format(line) - elif line.startswith('needCheckSupport.flag'): - op_desc.parse_check_support(line) - elif line.startswith('rangeLimit.value'): - op_desc.parse_range_limit(line) - elif line.startswith('opInterface.value'): - op_desc.parse_op_intf(line) - elif line.startswith('kernel.name'): - op_desc.parse_kern_name(line) - elif line.startswith('opFile.value'): - op_desc.parse_op_file(line) - elif line.startswith('dynamicShapeSupport.flag'): - op_desc.parse_dynamic_shape(line) - elif line.startswith('attr.list'): - op_desc.parse_attr_list(line) - elif line.startswith('attr_'): - op_desc.parse_attr_val(line) - _get_op_custom_options(op_descs, auto_gen_dir) - return op_descs diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/parse_ini_to_json.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/parse_ini_to_json.py deleted file mode 100755 index 7a505aa38..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/parse_ini_to_json.py +++ /dev/null @@ -1,338 +0,0 @@ -# Copyright 2020-2021 Huawei Technologies Co., Ltd -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -""" -parser ini to json -""" - -import json -import os -import stat -import sys - - -ATTR_TYPE_LIST = ["int", "float", "bool", "str", "listInt", "listFloat", "listBool", "listStr", "listListInt", - "type", "listType", "tensor", "listTensor"] -ATTR_PARAMTYPE_LIST = ["optional", "required"] -BOOL_FLAG_KEY = ["dynamicFormat", "dynamicShapeSupport", "dynamicRankSupport", "precision_reduce", "heavyOp", - "needCheckSupport"] -BOOL_LIST = ["true", "false"] -DTYPE_LIST = ["float16", "float", "float32", "int8", "int16", "int32", "uint8", "uint16", "uint32", "bool", - "int64", "uint64", "qint8", "qint16", "qint32", "quint8", "quint16", "double", "complex64", - "complex128", "string", "resource", "dual", "dual_sub_int8", "dual_sub_uint8", "string_ref", - "int4", "bfloat16", "uint1"] -FORMAT_LIST = ["NCHW", "NHWC", "ND", "NC1HWC0", "FRACTAL_Z", "NC1C0HWPAD", "NHWC1C0", "FSR_NCHW", "FRACTAL_DECONV", - "C1HWNC0", "FRACTAL_DECONV_TRANSPOSE", "FRACTAL_DECONV_SP_STRIDE_TRANS", "NC1HWC0_C04", - "FRACTAL_Z_C04", "CHWN", "FRACTAL_DECONV_SP_STRIDE8_TRANS", "HWCN", "NC1KHKWHWC0", "BN_WEIGHT", - "FILTER_HWCK", "HASHTABLE_LOOKUP_LOOKUPS", "HASHTABLE_LOOKUP_KEYS", "HASHTABLE_LOOKUP_VALUE", - "HASHTABLE_LOOKUP_OUTPUT", "HASHTABLE_LOOKUP_HITS", "C1HWNCoC0", "MD", "NDHWC", "FRACTAL_ZZ", - "FRACTAL_NZ", "NCDHW", "DHWCN", "NDC1HWC0", "FRACTAL_Z_3D", "CN", "NC", "DHWNC", - "FRACTAL_Z_3D_TRANSPOSE", "FRACTAL_ZN_LSTM", "FRACTAL_ZN_RNN", "FRACTAL_Z_G", "NULL"] - - -def parse_ini_files(ini_files): - """ - parse ini files to json - Parameters: - ---------------- - ini_files:input file list - return:ops_info - ---------------- - """ - tbe_ops_info = {} - for ini_file in ini_files: - check_file_size(ini_file) - parse_ini_to_obj(ini_file, tbe_ops_info) - return tbe_ops_info - - -def check_file_size(input_file): - try: - file_size = os.path.getsize(input_file) - except OSError as os_error: - print('[ERROR] Failed to open "%s". %s' % (input_file, str(os_error))) - raise OSError from os_error - if file_size > 10*1024*1024: - print('[WARN] The size of %s exceeds 10MB, it may take more time to run, please wait.' % input_file) - - -def parse_ini_to_obj(ini_file, tbe_ops_info): - """ - parse ini file to json obj - Parameters: - ---------------- - ini_file:ini file path - tbe_ops_info:ops_info - ---------------- - """ - with open(ini_file) as ini_file: - lines = ini_file.readlines() - op_dict = {} - op_name = "" - find_op_type = False - for line in lines: - line = line.rstrip() - if line == "": - continue - if line.startswith("["): - if line.endswith("]"): - op_name = line[1:-1] - op_dict = {} - tbe_ops_info[op_name] = op_dict - find_op_type = True - elif "=" in line: - key1 = line[:line.index("=")] - key2 = line[line.index("=")+1:] - key1_0, key1_1 = key1.split(".") - if key1_0 not in op_dict: - op_dict[key1_0] = {} - if key1_1 in op_dict.get(key1_0): - raise RuntimeError("Op:" + op_name + " " + key1_0 + " " + - key1_1 + " is repeated!") - dic_key = op_dict.get(key1_0) - dic_key[key1_1] = key2 - else: - continue - if not find_op_type: - raise RuntimeError("Not find OpType in .ini file.") - - -def check_output_exist(op_dict, is_valid): - """ - Function Description: - Check output is exist - Parameter: op_dict - Parameter: is_valid - """ - if "output0" in op_dict: - output0_dict = op_dict.get("output0") - if output0_dict.get("name", None) is None: - is_valid = False - print("output0.name is required in .ini file!") - else: - is_valid = False - print("output0 is required in .ini file!") - return is_valid - - -def check_attr_dict(attr_dict, is_valid, attr): - """ - Function Description: - Check attr_dict - Parameter: attr_dict - Parameter: is_valid - Parameter: attr - """ - attr_type = attr_dict.get("type") - value = attr_dict.get("value") - param_type = attr_dict.get("paramType") - if attr_type is None or value is None: - is_valid = False - print("If attr.list is exist, {0}.type and {0}.value is required".format(attr)) - if param_type and param_type not in ATTR_PARAMTYPE_LIST: - is_valid = False - print("{0}.paramType only support {1}.".format(attr, ATTR_PARAMTYPE_LIST)) - if attr_type and attr_type not in ATTR_TYPE_LIST: - is_valid = False - print("{0}.type only support {1}.".format(attr, ATTR_TYPE_LIST)) - return is_valid - - -def check_attr(op_dict, is_valid): - """ - Function Description: - Check attr - Parameter: op_dict - Parameter: is_valid - """ - if "attr" in op_dict: - attr_dict = op_dict.get("attr") - attr_list_str = attr_dict.get("list", None) - if attr_list_str is None: - is_valid = False - print("attr.list is required in .ini file!") - else: - attr_list = attr_list_str.split(",") - for attr_name in attr_list: - attr = "attr_" + attr_name.strip() - attr_dict = op_dict.get(attr) - if attr_dict: - is_valid = check_attr_dict(attr_dict, is_valid, attr) - else: - is_valid = False - print("%s is required in .ini file, when attr.list is %s!" % (attr, attr_list_str)) - return is_valid - - -def check_bool_flag(op_dict, is_valid): - """ - Function Description: - check_bool_flag - Parameter: op_dict - Parameter: is_valid - """ - for key in BOOL_FLAG_KEY: - if key in op_dict: - op_bool_key = op_dict.get(key) - if op_bool_key.get("flag").strip() not in BOOL_LIST: - is_valid = False - print("{0}.flag only support {1}.".format(key, BOOL_LIST)) - return is_valid - - -def check_type_format(op_info, is_valid, op_info_key): - """ - Function Description: - Check type and format - Parameter: op_info - Parameter: is_valid - Parameter: op_info_key - """ - op_info_dtype_str = op_info.get("dtype") - op_info_dtype_num = 0 - op_info_format_num = 0 - if op_info_dtype_str: - op_info_dtype = op_info_dtype_str.split(",") - op_info_dtype_num = len(op_info_dtype) - for dtype in op_info_dtype: - if dtype.strip() not in DTYPE_LIST: - is_valid = False - print("{0}.dtype not support {1}.".format(op_info_key, dtype)) - op_info_format_str = op_info.get("format") - if op_info_format_str: - op_info_format = op_info_format_str.split(",") - op_info_format_num = len(op_info_format) - for op_format in op_info_format: - if op_format.strip() not in FORMAT_LIST: - is_valid = False - print("{0}.format not support {1}.".format(op_info_key, op_format)) - if op_info_dtype_num > 0 and op_info_format_num > 0: - if op_info_dtype_num != op_info_format_num: - is_valid = False - print("The number of {0}.dtype not match the number of {0}.format.".format(op_info_key)) - return is_valid - - -def check_op_info(tbe_ops): - """ - Function Description: - Check info. - Parameter: tbe_ops - Return Value: is_valid - """ - print("\n\n==============check valid for ops info start==============") - required_op_input_info_keys = ["paramType", "name"] - required_op_output_info_keys = ["paramType", "name"] - param_type_valid_value = ["dynamic", "optional", "required"] - is_valid = True - for op_key in tbe_ops: - op_dict = tbe_ops[op_key] - for op_info_key in op_dict: - if op_info_key.startswith("input"): - op_input_info = op_dict[op_info_key] - missing_keys = [] - for required_op_input_info_key in required_op_input_info_keys: - if required_op_input_info_key not in op_input_info: - missing_keys.append(required_op_input_info_key) - if len(missing_keys) > 0: - print("op: " + op_key + " " + op_info_key + " missing: " + - ",".join(missing_keys)) - is_valid = False - else: - if not op_input_info["paramType"] in param_type_valid_value: - print("op: " + op_key + " " + op_info_key + \ - " paramType not valid, valid key:[dynamic, " - "optional, required]") - is_valid = False - is_valid = check_type_format(op_input_info, is_valid, op_info_key) - if op_info_key.startswith("output"): - op_input_info = op_dict[op_info_key] - missing_keys = [] - for required_op_input_info_key in required_op_output_info_keys: - if required_op_input_info_key not in op_input_info: - missing_keys.append(required_op_input_info_key) - if len(missing_keys) > 0: - print("op: " + op_key + " " + op_info_key + " missing: " + - ",".join(missing_keys)) - is_valid = False - else: - if not op_input_info["paramType"] in param_type_valid_value: - print("op: " + op_key + " " + op_info_key + - " paramType not valid, valid key:[dynamic, " - "optional, required]") - is_valid = False - is_valid = check_type_format(op_input_info, is_valid, op_info_key) - is_valid = check_attr(op_dict, is_valid) - is_valid = check_bool_flag(op_dict, is_valid) - print("==============check valid for ops info end================\n\n") - return is_valid - - -def write_json_file(tbe_ops_info, json_file_path): - """ - Save info to json file - Parameters: - ---------------- - tbe_ops_info: ops_info - json_file_path: json file path - ---------------- - """ - json_file_real_path = os.path.realpath(json_file_path) - wr_flag = os.O_WRONLY | os.O_CREAT - wr_mode = stat.S_IWUSR | stat.S_IRUSR - with os.fdopen(os.open(json_file_real_path, wr_flag, wr_mode), 'w') as file_path: - # The owner have all rights,group only have read rights - os.chmod(json_file_real_path, stat.S_IWUSR + stat.S_IRGRP - + stat.S_IRUSR) - json.dump(tbe_ops_info, file_path, sort_keys=True, indent=4, - separators=(',', ':')) - print("Compile op info cfg successfully.") - - -def parse_ini_to_json(ini_file_paths, outfile_path): - """ - parse ini files to json file - Parameters: - ---------------- - ini_file_paths: list of ini file path - outfile_path: output file path - ---------------- - """ - tbe_ops_info = parse_ini_files(ini_file_paths) - if not check_op_info(tbe_ops_info): - print("Compile op info cfg failed.") - return False - write_json_file(tbe_ops_info, outfile_path) - return True - - -if __name__ == '__main__': - args = sys.argv - - OUTPUT_FILE_PATH = "tbe_ops_info.json" - ini_file_path_list = [] - - for arg in args: - if arg.endswith("ini"): - ini_file_path_list.append(arg) - OUTPUT_FILE_PATH = arg.replace(".ini", ".json") - if arg.endswith("json"): - OUTPUT_FILE_PATH = arg - - if len(ini_file_path_list) == 0: - ini_file_path_list.append("tbe_ops_info.ini") - - if not parse_ini_to_json(ini_file_path_list, OUTPUT_FILE_PATH): - sys.exit(1) - sys.exit(0) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/preset_parse.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/preset_parse.py deleted file mode 100755 index 8f1124b1d..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/preset_parse.py +++ /dev/null @@ -1,23 +0,0 @@ -import json -import sys -import os - - -def get_config_opts(file): - src_dir = os.path.abspath(os.path.dirname(file)) - opts = '' - with open(file, 'r') as fd: - config = json.load(fd) - for conf in config: - if conf == 'configurePresets': - for node in config[conf]: - macros = node.get('cacheVariables') - if macros is not None: - for key in macros: - opts += '-D{}={} '.format(key, macros[key]['value']) - opts = opts.replace('${sourceDir}', src_dir) - print(opts) - - -if __name__ == "__main__": - get_config_opts(sys.argv[1]) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_codegen.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_codegen.py deleted file mode 100755 index 1baa364ef..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_codegen.py +++ /dev/null @@ -1,105 +0,0 @@ -#!/usr/bin/env python -# -*- coding: UTF-8 -*- -""" -Created on Feb 28 20:56:45 2020 -Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. -""" - -import os -import stat -import collections -import kernel_entry as keb -from tiling_data_def_build import gen_tiling -import code_channel_infer -import const_var - -PYF_PATH = os.path.dirname(__file__) - -ReplayCodeGenParams = collections.namedtuple('ReplayCodeGenParams',\ -['op_type', 'impl', 'tiling_file', 'kernel', 'entry', 'argn', 'op_replay_batch', 'max_block_dim', 'max_shape_size']) - - -class ReplayCodeGen: - def __init__(self, replayCodeGenParams): - self.op_type = replayCodeGenParams.op_type - self.impl = replayCodeGenParams.impl - self.tiling_file = replayCodeGenParams.tiling_file - self.tiling_data_file = '' - self.kernel = replayCodeGenParams.kernel - self.entry = replayCodeGenParams.entry - self.argn = replayCodeGenParams.argn - self.batch = False - self.outdir = '' - self.data_type = 'uint8_t' - self.blknum = 32 - self.op_replay_batch = replayCodeGenParams.op_replay_batch - self.max_block_dim = replayCodeGenParams.max_block_dim - self.max_shape_size = replayCodeGenParams.max_shape_size - - def set_batch(self, is_batch): - self.batch = is_batch - - def set_outdir(self, outdir): - self.outdir = outdir - - def gen_replay(self, ops_product: str): - kerentry = os.path.join(self.outdir, self.kernel + '_entry.cce') - kerimpl = os.path.join(self.outdir, self.kernel + '_impl.cpp') - replayimpl = os.path.join(self.outdir, self.kernel + '_replay.cpp') - if self.batch: - reptmp = os.path.join(PYF_PATH, 'batch_replay_impl.temp') - else: - reptmp = os.path.join(PYF_PATH, 'replay_impl.temp') - kertmp = os.path.join(PYF_PATH, 'kernel_impl.temp') - self._gen_kentry(kerentry) - self._gen_kimpl_code(kerimpl, kertmp) - self._gen_tiling_data_header() - self._gen_replay_code(replayimpl, reptmp, ops_product) - - def _gen_tiling_data_header(self): - self.tiling_data_file = os.path.join(self.outdir, self.kernel + '_tiling_data.h') - gen_tiling(self.tiling_file, self.tiling_data_file) - - def _gen_kimpl_code(self, src, tmpfile): - with open(tmpfile, 'r') as fd: - temp = fd.read() - temp = temp.replace('__CCE_FILE__', self.impl) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(temp) - - def _gen_replay_code(self, src, tmpfile, ops_product: str): - with open(tmpfile, 'r') as fd: - temp = fd.read() - temp = temp.replace('__ARG_NUM__', str(self.argn)) - argdef = [] - kargs = [] - for i in range(0, self.argn): - argdef.append('{} *'.format(self.data_type)) - kargs.append('({} *)GetArg({})'.format(self.data_type, i)) - temp = temp.replace('__ARGS_DEF__', ', '.join(argdef)) - temp = temp.replace('__KERNEL_ARGS__', ', '.join(kargs)) - temp = temp.replace('__KERNEL_FUN__', self.entry) - core_type_infer = 'core_type' - code_channel = code_channel_infer.infer_code_channel(code_channel_infer.InfoCodeChanelParams(self.impl,\ - self.tiling_data_file, self.kernel, self.outdir, ops_product, None)) - if code_channel == code_channel_infer.CODE_VEC: - core_type_infer = '0' - elif code_channel == code_channel_infer.CODE_CUBE: - core_type_infer = '1' - temp = temp.replace('__CORE_TYPE__', core_type_infer) - # regist function - temp = temp.replace('__OPS_PRODUCT__', ops_product) - temp = temp.replace('__OPTYPE__', self.op_type) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(temp) - - def _gen_kentry(self, src): - kf = '' - pre_alloc_str = 'A' * 256 - if self.batch: - kf += keb.batch_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str), self.argn, self.data_type) - else: - kf += keb.mc_code_gen("K{:02d}_{}{}".format(0, self.entry, pre_alloc_str),\ - self.argn, self.data_type, self.blknum) - with os.fdopen(os.open(src, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(kf) diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_impl.temp b/atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_impl.temp deleted file mode 100755 index 1d30dd865..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/replay_impl.temp +++ /dev/null @@ -1,120 +0,0 @@ -#include -#include -#include -#include -#include -#include -#include "replay_def.h" -#include "code_gen.h" -#include "replay_fun.h" -#include "register/op_check.h" -#define __ASCENDC_REPLAY_CODE__ -using namespace std; -using namespace optiling; -using namespace AscendCReplay; - -extern "C" void __KERNEL_FUN__ (__ARGS_DEF__, const char *); -extern "C" int elf_append(char *elf, uint32_t elfSize, char *jit, int kernum, int blknum[], char *atext[], - int alen[], int atlen, const char* kernelname[]); - -#define KERNEL_N 1 -#define ARG_N (__ARG_NUM__) -#define MAX_L (1024 * 1024 * 100) -#define MAX_E (1024 * 1024) - -int __KERNEL_FUN___replay___OPS_PRODUCT__(ReplayFuncParam& param, const int core_type) -{ - // gen type 1 : direct call codes 0: load .o file - if (param.gentype < 0 || param.gentype > 1) { - printf("Error: call replay gen type is %d, should only be 1 or 0\n", param.gentype); - return 0; - } else if (param.gentype == 1 && param.objptr == nullptr) { - printf("Error: call replay with direct call mode, but code obj addr is null\n"); - return 0; - } else if (param.gentype == 0 && param.output_kernel_file == nullptr) { - printf("Error: call replay with object file mode, but object file path is null\n"); - return 0; - } - // core_type 0:MIX 1:CUBE 2:VEC - if (core_type < 0 || core_type > 2) { - printf("Error: call replay core type is %d !\n", core_type); - return 0; - } - g_coreType = __CORE_TYPE__; - g_taskRation = param.task_ration; - g_tilingKey = param.tiling_key; - - unsigned char *buf, *jit; - char *kernel[KERNEL_N * 32]; - int len[KERNEL_N * 32]; - int blknum[KERNEL_N]; - int max; - block_num = param.block_dim; - g_ubBase = block_num; - uint8_t *code = (uint8_t *)malloc(MAX_L); - uint8_t *pos = code; - struct timespec tp1, tp2; - - clock_gettime(CLOCK_MONOTONIC, &tp1); - if (block_num > 32) { - printf("Error: block_num > 32\n"); - return 0; - } - //__OP_FOPEN__ - for (int i = 0; i < KERNEL_N; i++) { - for (int j = 0; j < ARG_N; j++) - AddArg(j, ARG_STEP * (j + 1)); - for (block_idx = 0; block_idx < block_num; block_idx++) { - //__OP_SET_KERNEL__ - int code_idx = i * block_num + block_idx; -#ifdef FP_CEILING - SetCtrlFloatEnable(); -#else - SetCtrlFloatDisable(); -#endif - CodeInit(pos, false); - __KERNEL_FUN__(__KERNEL_ARGS__, param.tiling_data); - CodeEnd(); - kernel[code_idx] = (char *)pos; - len[code_idx] = CodeLen(); - pos += len[code_idx]; - printf("kernel %d core %ld code generated len %d\n", i, block_idx, len[code_idx]); - } - blknum[i] = block_num; - } - //__OP_FCLOSE__ - clock_gettime(CLOCK_MONOTONIC, &tp2); - buf = (unsigned char *)malloc(MAX_E); - int fd = open(param.entry_file, O_RDONLY); - if (fd < 0) { - printf("[error]: cannot find entry.o : %s\n", param.entry_file); - return 0; - } - uint32_t bufSize = read(fd, buf, MAX_E); - if (bufSize <= 0) { - printf("[error]: entry.o : %s is too small ! \n", param.entry_file); - } - close(fd); - jit = (unsigned char *)malloc(MAX_L); - printf("total code generated %ld\n", pos - code); - int sz = elf_append((char *)buf, bufSize, (char *)jit, KERNEL_N, blknum, kernel, len, pos - code, ¶m.kernel_name); - if (tp1.tv_sec != tp2.tv_sec) { - printf("%ld NS\n", tp2.tv_nsec + 1000000000 - tp1.tv_nsec); - } else { - printf("%ld NS\n", tp2.tv_nsec - tp1.tv_nsec); - } - printf("new elf size %d\n", sz); - if (param.gentype == 0) { - fd = open(param.output_kernel_file, O_RDWR | O_CREAT | O_TRUNC, S_IRUSR | S_IWUSR); - (void)write(fd, jit, sz); - close(fd); - free(jit); - } else if (param.gentype == 1) { - *param.objptr = (char*)jit; - } - free(buf); - free(code); - return sz; -} - -REG_REPLAY_FUNC(__OPTYPE__, __OPS_PRODUCT__, __KERNEL_FUN___replay___OPS_PRODUCT__); diff --git a/atb_operator/AtbTrilSample/TrilOp/cmake/util/tiling_data_def_build.py b/atb_operator/AtbTrilSample/TrilOp/cmake/util/tiling_data_def_build.py deleted file mode 100755 index 4c5b099b7..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/cmake/util/tiling_data_def_build.py +++ /dev/null @@ -1,87 +0,0 @@ -#!/usr/bin/env python -# coding=utf-8 -""" -Function: -The replay funtion entry -Copyright Information: -Huawei Technologies Co., Ltd. All Rights Reserved © 2020 -""" - -import sys -import os -import stat -import re -import const_var - - -def gen_tiling(tiling_header_file: str, tiling_file_out: str): - if not os.path.exists(tiling_header_file): - print("warning: no userdef tiling header file: ", tiling_header_file) - return - print("generate tiling def header file: ", tiling_file_out) - tmp_name = os.path.splitext(os.path.basename(tiling_header_file))[0].upper() - tiling_source = '#ifndef __{}_H__\n'.format(tmp_name) - tiling_source += '#define __{}_H__\n\n'.format(tmp_name) - tiling_source += '#include \n' - tiling_source += '#include \n\n' - tiling_source += '#include "kernel_tiling/kernel_tiling.h"\n\n' - end_source = "" - pattern = re.compile(r'[(](.*)[)]', re.S) - with open(tiling_header_file, 'r') as fd: - lines = fd.readlines() - for line in lines: - line = line.strip() - if (line.startswith('BEGIN_TILING_DATA_DEF')): - tiling_source += '#pragma pack(1)\n' - tiling_source += 'struct ' - struct_def = re.findall(pattern, line)[0] - tiling_source += struct_def + ' {\n' - elif (line.startswith('TILING_DATA_FIELD_DEF_ARR')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {}[{}] = {{}};\n'.format(fds[0].strip(), fds[2].strip(), fds[1].strip()) - elif (line.startswith('TILING_DATA_FIELD_DEF_STRUCT')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {};\n'.format(fds[0].strip(), fds[1].strip()) - elif (line.startswith('TILING_DATA_FIELD_DEF')): - field_params = re.findall(pattern, line)[0] - fds = field_params.split(',') - tiling_source += ' {} {} = 0;\n'.format(fds[0].strip(), fds[1].strip()) - elif (line.startswith('END_TILING_DATA_DEF')): - tiling_source += '};\n' - tiling_source += '#pragma pack()\n\n' - tiling_source += '#ifdef __NPU_TILING__\n' - tiling_source += \ - 'inline [aicore] void Init{stru}(const __gm__ uint8_t* tiling, {stru}* const_data)\n'\ - .format(stru=struct_def) - tiling_source += '{\n' - tiling_source += ' const __gm__ uint32_t *src = (const __gm__ uint32_t *)tiling;\n' - tiling_source += ' uint32_t *dst = (uint32_t *)const_data;\n' - tiling_source += ' for (auto i = 0; i < sizeof({}) / 4; i++) *(dst + i) = *(src + i);\n'\ - .format(struct_def) - tiling_source += '}\n' - tiling_source += '#else\n' - tiling_source += 'inline void Init{stru}(uint8_t* tiling, {stru}* const_data)\n'.format(stru=struct_def) - tiling_source += '{\n' - tiling_source += ' uint64_t *src = (uint64_t *)tiling;\n' - tiling_source += ' uint64_t *dst = (uint64_t *)const_data;\n' - tiling_source += ' for (auto i = 0; i < sizeof({}) / 8; i++) *(dst + i) = *(src + i);\n'\ - .format(struct_def) - tiling_source += '}\n' - tiling_source += '#endif\n\n' - end_source = ''' -#define GET_TILING_DATA(tiling_data, tiling_arg) \\ -{stru} tiling_data; \\ -Init{stru}(tiling_arg, &tiling_data)\n -'''.format(stru=struct_def) - tiling_source += end_source - tiling_source += '#endif' - with os.fdopen(os.open(tiling_file_out, const_var.WFLAGS, const_var.WMODES), 'w') as ofd: - ofd.write(tiling_source) - - -if __name__ == '__main__': - if len(sys.argv) <= 2: - raise RuntimeError('arguments must greater than 2') - gen_tiling(sys.argv[1], sys.argv[2]) diff --git a/atb_operator/AtbTrilSample/TrilOp/framework/CMakeLists.txt b/atb_operator/AtbTrilSample/TrilOp/framework/CMakeLists.txt deleted file mode 100755 index b6be9b492..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/framework/CMakeLists.txt +++ /dev/null @@ -1,11 +0,0 @@ -if(NOT EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/mindspore") - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/caffe_plugin") - add_subdirectory(caffe_plugin) - endif() - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/tf_plugin") - add_subdirectory(tf_plugin) - endif() - if(EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/onnx_plugin") - add_subdirectory(onnx_plugin) - endif() -endif() diff --git a/atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/CMakeLists.txt b/atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/CMakeLists.txt deleted file mode 100755 index a6aba5c20..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/CMakeLists.txt +++ /dev/null @@ -1,14 +0,0 @@ - -aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} plugin_srcs) -add_library(cust_tf_parsers SHARED ${plugin_srcs}) -target_compile_definitions(cust_tf_parsers PRIVATE google=ascend_private) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_tf_parsers PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_tf_parsers PRIVATE intf_pub graph) -install(TARGETS cust_tf_parsers - LIBRARY DESTINATION packages/vendors/${vendor_name}/framework/tensorflow -) diff --git a/atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/tensorflow_tril_plugin.cc b/atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/tensorflow_tril_plugin.cc deleted file mode 100755 index 9b71ea121..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/framework/tf_plugin/tensorflow_tril_plugin.cc +++ /dev/null @@ -1,20 +0,0 @@ -/* Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. - * - * This program is free software; you can redistribute it and/or modify - * it under the terms of the Apache License Version 2.0. - * You may not use this file except in compliance with the License. - * - * This program is distributed in the hope that it will be useful, - * but WITHOUT ANY WARRANTY; without even the implied warranty of - * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. - */ - -#include "register/register.h" - -namespace domi { -// register op info to GE -REGISTER_CUSTOM_OP("Tril") - .FrameworkType(TENSORFLOW) // type: CAFFE, TENSORFLOW - .OriginOpType("Tril") // name in tf module - .ParseParamsByOperatorFn(AutoMappingByOpFn); -} // namespace domi diff --git a/atb_operator/AtbTrilSample/TrilOp/op_host/CMakeLists.txt b/atb_operator/AtbTrilSample/TrilOp/op_host/CMakeLists.txt deleted file mode 100755 index 40dd51cfa..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/op_host/CMakeLists.txt +++ /dev/null @@ -1,82 +0,0 @@ - -aux_source_directory(${CMAKE_CURRENT_SOURCE_DIR} ops_srcs) - -opbuild(OPS_SRC ${ops_srcs} - OUT_DIR ${ASCEND_AUTOGEN_PATH} -) - -add_library(cust_op_proto SHARED ${ops_srcs} ${ASCEND_AUTOGEN_PATH}/op_proto.cc) -target_compile_definitions(cust_op_proto PRIVATE OP_PROTO_LIB) -target_compile_options(cust_op_proto PRIVATE - -fvisibility=hidden -) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_op_proto PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_op_proto PRIVATE - intf_pub - exe_graph - register - tiling_api - -Wl,--whole-archive - rt2_registry - -Wl,--no-whole-archive -) -set_target_properties(cust_op_proto PROPERTIES OUTPUT_NAME - cust_opsproto_rt2.0 -) -add_library(cust_optiling SHARED ${ops_srcs}) -target_compile_definitions(cust_optiling PRIVATE OP_TILING_LIB) -target_compile_options(cust_optiling PRIVATE - -fvisibility=hidden -) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_optiling PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_optiling PRIVATE - intf_pub - exe_graph - register - tiling_api - -Wl,--whole-archive - rt2_registry - -Wl,--no-whole-archive -) -set_target_properties(cust_optiling PROPERTIES OUTPUT_NAME - cust_opmaster_rt2.0 -) - -file(GLOB aclnn_src ${ASCEND_AUTOGEN_PATH}/aclnn_*.cpp) -file(GLOB aclnn_inc ${ASCEND_AUTOGEN_PATH}/aclnn_*.h) -add_library(cust_opapi SHARED ${aclnn_src}) -if(ENABLE_CROSS_COMPILE) - target_link_directories(cust_opapi PRIVATE - ${CMAKE_COMPILE_COMPILER_LIBRARY} - ${CMAKE_COMPILE_RUNTIME_LIBRARY} - ) -endif() -target_link_libraries(cust_opapi PRIVATE intf_pub ascendcl nnopbase) - -add_custom_target(optiling_compat ALL - COMMAND ln -sf lib/linux/${CMAKE_SYSTEM_PROCESSOR}/$ - ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so -) - -install(TARGETS cust_op_proto - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_proto/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) -install(FILES ${ASCEND_AUTOGEN_PATH}/op_proto.h - DESTINATION packages/vendors/${vendor_name}/op_proto/inc) -install(TARGETS cust_optiling - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling/lib/linux/${CMAKE_SYSTEM_PROCESSOR}) -install(FILES ${CMAKE_CURRENT_BINARY_DIR}/liboptiling.so - DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/op_tiling) -install(TARGETS cust_opapi - LIBRARY DESTINATION packages/vendors/${vendor_name}/op_api/lib) -install(FILES ${aclnn_inc} - DESTINATION packages/vendors/${vendor_name}/op_api/include) diff --git a/atb_operator/AtbTrilSample/TrilOp/op_host/tril.cpp b/atb_operator/AtbTrilSample/TrilOp/op_host/tril.cpp deleted file mode 100755 index 60b0bb3b0..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/op_host/tril.cpp +++ /dev/null @@ -1,215 +0,0 @@ -/* Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. - * - * This program is free software; you can redistribute it and/or modify - * it under the terms of the Apache License Version 2.0. - * You may not use this file except in compliance with the License. - * - * This program is distributed in the hope that it will be useful, - * but WITHOUT ANY WARRANTY; without even the implied warranty of - * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. - */ - -#include "tril_tiling.h" -#include "register/op_def_registry.h" -#include "tiling/platform/platform_ascendc.h" - -namespace optiling { - constexpr int minNum = 1; - - constexpr int keyOne = 1; - constexpr int keyTwo = 2; - constexpr int keyThree = 3; - constexpr int keyFour = 4; - - constexpr int bufferFour = 4; - constexpr int BlockSize = 32; - constexpr int computeBatchSize = 256; - constexpr int sizeHalf = 2; - constexpr int VAL_ZRRO = 0; - - uint32_t typeSize = VAL_ZRRO; - uint64_t key = keyOne; - // buffer for queue - uint64_t UB_SHARING_NUM = 2; - int64_t rowLength = VAL_ZRRO; - int64_t columnLength = VAL_ZRRO; - int64_t matrixNum = 1, matrixSize = 1; - int64_t diagVal = VAL_ZRRO; - - uint32_t ALIGN_NUM = VAL_ZRRO; - uint32_t totalLengthAligned = VAL_ZRRO; - uint64_t loopCnt = VAL_ZRRO, fullTileLength = VAL_ZRRO, lastTileLength = VAL_ZRRO; - int32_t fullCnt = VAL_ZRRO, lastCnt = VAL_ZRRO; - - static int setShapeInfo(gert::TilingContext *context){ - const auto inputDataType = context->GetInputTensor(0)->GetDataType(); - - switch (inputDataType){ - case ge::DT_FLOAT: - typeSize = sizeof(float); - break; - case ge::DT_FLOAT16: - typeSize = sizeHalf; - break; - default: - typeSize = sizeof(float); - break; - } - - const auto inputShape = context->GetInputTensor(0)->GetOriginShape(); - // class Shape: size_t dim_num_; int64_t dims_[]; - int64_t dimSize = inputShape.GetDimNum(), i = 0; - // The number 2 is to preserve the last two dimensions - for (i = 0; i < dimSize - 2; i++){ - matrixNum *= inputShape.GetDim(i); - } - rowLength = inputShape.GetDim(i); - i++; - columnLength = inputShape.GetDim(i); - matrixSize = rowLength * columnLength; - - const auto runtime_attrs = context->GetAttrs(); - const int64_t *diagPtr = runtime_attrs->GetInt(0); - diagVal = *diagPtr; - if (diagVal < columnLength - 1 && diagVal > -rowLength){ - // Regular - key = keyOne; - }else if (diagVal <= -rowLength){ - // The result is itself, TQueBind is enough - key = keyTwo; - }else{ - // All zero, just copyIn, Sub and copyOut - key = keyThree; - } - return 0; - } - - static int setTilingInfo(gert::TilingContext *context,uint64_t ub_size){ - loopCnt = VAL_ZRRO; - fullTileLength = VAL_ZRRO; - lastTileLength = VAL_ZRRO; - fullCnt = VAL_ZRRO; - lastCnt = VAL_ZRRO; - uint64_t ub_length = ((ub_size / typeSize / UB_SHARING_NUM) / ALIGN_NUM * ALIGN_NUM) - ALIGN_NUM; - if (key == keyOne && diagVal <= 0 && columnLength % (computeBatchSize / typeSize) == 0){ - // A faster method for aligned processing only - key = keyFour; - // Double buffer setting - UB_SHARING_NUM = bufferFour; - // The result would not be the expected - if (columnLength == 0){ - columnLength = minNum; - } - ub_length = ((ub_size) / typeSize / UB_SHARING_NUM) / columnLength * columnLength; - loopCnt = (matrixSize + ub_length - 1) / ub_length; - if (loopCnt == 1){ - fullCnt = 0; - lastCnt = rowLength; - }else{ - // The result would not be the expected - if (columnLength == 0){ - columnLength = minNum; - } - fullCnt = ub_length / columnLength; - lastCnt = rowLength - fullCnt * (loopCnt - 1); - } - // Already aligned - fullTileLength = fullCnt * columnLength; - lastTileLength = lastCnt * columnLength; - }else if (key == keyThree){ - loopCnt = (totalLengthAligned + ub_length - 1) / ub_length; - UB_SHARING_NUM = bufferFour; - ub_length = ((ub_size / typeSize / UB_SHARING_NUM) / ALIGN_NUM * ALIGN_NUM) - ALIGN_NUM; - fullTileLength = ub_length; - lastTileLength = (totalLengthAligned - fullTileLength * (loopCnt - 1) + ALIGN_NUM - 1) / ALIGN_NUM * ALIGN_NUM; - if (loopCnt == 1){ fullTileLength = 0; } - }else{ - loopCnt = (totalLengthAligned + ub_length - 1) / ub_length; - fullTileLength = ub_length; - lastTileLength = (totalLengthAligned - fullTileLength * (loopCnt - 1) + ALIGN_NUM - 1) / ALIGN_NUM * ALIGN_NUM; - if (loopCnt == 1){ fullTileLength = 0; } - } - return 0; - } - -static ge::graphStatus TilingFunc(gert::TilingContext* context) -{ - TrilTilingData tiling; - auto ascendcPlatform = platform_ascendc::PlatformAscendC(context->GetPlatformInfo()); - auto coreNum = ascendcPlatform.GetCoreNum(); - auto BLOCK_DIM = 1; - context->SetBlockDim(BLOCK_DIM); - - setShapeInfo(context); - - ALIGN_NUM = BlockSize / typeSize; - totalLengthAligned = (matrixNum * matrixSize + ALIGN_NUM - 1) / ALIGN_NUM * ALIGN_NUM; - uint64_t ub_size=0; - ascendcPlatform.GetCoreMemSize(platform_ascendc::CoreMemType::UB, ub_size); - - setTilingInfo(context,ub_size); - - tiling.set_totalLengthAligned(totalLengthAligned); - tiling.set_matrixNum(matrixNum); - tiling.set_matrixSize(matrixSize); - tiling.set_rowLength(rowLength); - tiling.set_columnLength(columnLength); - tiling.set_diagVal(diagVal); - - tiling.set_loopCnt(loopCnt); - tiling.set_fullTileLength(fullTileLength); - tiling.set_lastTileLength(lastTileLength); - tiling.set_fullCnt(fullCnt); - tiling.set_lastCnt(lastCnt); - - tiling.set_alignNum(ALIGN_NUM); - tiling.set_typeSize(typeSize); - - tiling.SaveToBuffer(context->GetRawTilingData()->GetData(), context->GetRawTilingData()->GetCapacity()); - context->GetRawTilingData()->SetDataSize(tiling.GetDataSize()); - context->SetTilingKey(key); - size_t *currentWorkspace = context->GetWorkspaceSizes(1); - currentWorkspace[0] = 0; - - return ge::GRAPH_SUCCESS; -} -} - -namespace ge { -static ge::graphStatus InferShape(gert::InferShapeContext* context) -{ - const gert::Shape* x1_shape = context->GetInputShape(0); - gert::Shape* y_shape = context->GetOutputShape(0); - *y_shape = *x1_shape; - return GRAPH_SUCCESS; -} -} - -namespace ops { -class Tril : public OpDef { -public: - explicit Tril(const char* name) : OpDef(name) - { - this->Input("x") - .ParamType(REQUIRED) - .DataType({ge::DT_FLOAT16, ge::DT_FLOAT}) - .Format({ge::FORMAT_ND, ge::FORMAT_ND}) - .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND}); - this->Output("y") - .ParamType(REQUIRED) - .DataType({ge::DT_FLOAT16, ge::DT_FLOAT}) - .Format({ge::FORMAT_ND, ge::FORMAT_ND}) - .UnknownShapeFormat({ge::FORMAT_ND, ge::FORMAT_ND}); - this->Attr("diagonal").AttrType(OPTIONAL).Int(0); - - this->SetInferShape(ge::InferShape); - - this->AICore() - .SetTiling(optiling::TilingFunc); - this->AICore().AddConfig("ascend310b") - .AddConfig("ascend910b"); - } -}; - -OP_ADD(Tril); -} diff --git a/atb_operator/AtbTrilSample/TrilOp/op_host/tril_tiling.h b/atb_operator/AtbTrilSample/TrilOp/op_host/tril_tiling.h deleted file mode 100755 index fbf7749d2..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/op_host/tril_tiling.h +++ /dev/null @@ -1,34 +0,0 @@ -/* Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. - * - * This program is free software; you can redistribute it and/or modify - * it under the terms of the Apache License Version 2.0. - * You may not use this file except in compliance with the License. - * - * This program is distributed in the hope that it will be useful, - * but WITHOUT ANY WARRANTY; without even the implied warranty of - * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. - */ -#ifndef TRIL_TILING_H -#define TRIL_TILING_H -#include "register/tilingdata_base.h" - -namespace optiling { -BEGIN_TILING_DATA_DEF(TrilTilingData) - TILING_DATA_FIELD_DEF(uint32_t, totalLengthAligned); - TILING_DATA_FIELD_DEF(int32_t, matrixNum); - TILING_DATA_FIELD_DEF(int32_t, matrixSize); - TILING_DATA_FIELD_DEF(int32_t, rowLength); - TILING_DATA_FIELD_DEF(int32_t, columnLength); - TILING_DATA_FIELD_DEF(int32_t, diagVal); - TILING_DATA_FIELD_DEF(int32_t, loopCnt); - TILING_DATA_FIELD_DEF(uint32_t, fullTileLength); - TILING_DATA_FIELD_DEF(uint32_t, lastTileLength); - TILING_DATA_FIELD_DEF(int32_t, fullCnt); - TILING_DATA_FIELD_DEF(int32_t, lastCnt); - TILING_DATA_FIELD_DEF(uint32_t, alignNum); - TILING_DATA_FIELD_DEF(uint32_t, typeSize); -END_TILING_DATA_DEF; - -REGISTER_TILING_DATA_CLASS(Tril, TrilTilingData) -} -#endif \ No newline at end of file diff --git a/atb_operator/AtbTrilSample/TrilOp/op_kernel/CMakeLists.txt b/atb_operator/AtbTrilSample/TrilOp/op_kernel/CMakeLists.txt deleted file mode 100755 index 8c94a952d..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/op_kernel/CMakeLists.txt +++ /dev/null @@ -1,68 +0,0 @@ -# set custom compile options -if ("${CMAKE_BUILD_TYPE}x" STREQUAL "Debugx") - add_ops_compile_options(ALL OPTIONS -g -O0) -endif() - -foreach(compute_unit ${ASCEND_COMPUTE_UNIT}) - - # generate aic-${compute_unit}-ops-info.json - add_ops_info_target(TARGET ops_info_gen_${compute_unit} - OUTPUT ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core/${compute_unit}/aic-${compute_unit}-ops-info.json - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/config/${compute_unit} - ) - - # generate ascendc impl py once - if (NOT TARGET ascendc_impl_gen) - add_ops_impl_target(TARGET ascendc_impl_gen - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl - ) - endif() - - # dynamic shape binary compile - if (${ENABLE_BINARY_PACKAGE} AND NOT ${ENABLE_CROSS_COMPILE}) - add_bin_compile_target(TARGET ascendc_bin_${compute_unit} - OPS_INFO ${ASCEND_AUTOGEN_PATH}/aic-${compute_unit}-ops-info.ini - IMPL_DIR ${CMAKE_CURRENT_SOURCE_DIR} - ADP_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/dynamic - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/binary/${compute_unit} - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/kernel - COMPUTE_UNIT ${compute_unit} - ) - add_dependencies(ascendc_bin_${compute_unit} ascendc_impl_gen) - endif() - - if (${ENABLE_CROSS_COMPILE} AND ${ENABLE_BINARY_PACKAGE}) - add_cross_compile_target( - TARGET bin_${compute_unit} - OUT_DIR ${CMAKE_CURRENT_SOURCE_DIR}/../kernel - INSTALL_DIR packages/vendors/${vendor_name}/op_impl/ai_core/tbe/ - ) - endif() -endforeach() - -# generate npu_supported_ops.json -add_npu_support_target(TARGET npu_supported_ops - OPS_INFO_DIR ${ASCEND_AUTOGEN_PATH} - OUT_DIR ${CMAKE_CURRENT_BINARY_DIR}/tbe/op_info_cfg/ai_core - INSTALL_DIR packages/vendors/${vendor_name}/framework/${ASCEND_FRAMEWORK_TYPE} -) - -if(ENABLE_TEST AND EXISTS ${CMAKE_CURRENT_SOURCE_DIR}/testcases) - add_subdirectory(testcases) -endif() - -# install kernel file -if (${ENABLE_SOURCE_PACKAGE}) - file(GLOB KERNEL_FILES - ${CMAKE_CURRENT_SOURCE_DIR}/*.cpp - ${CMAKE_CURRENT_SOURCE_DIR}/*.h - ${CMAKE_CURRENT_SOURCE_DIR}/*.py - ) - install(FILES ${KERNEL_FILES} - DESTINATION packages/vendors/${vendor_name}/op_impl/ai_core/tbe/${vendor_name}_impl/dynamic - ) -endif() diff --git a/atb_operator/AtbTrilSample/TrilOp/op_kernel/tril.cpp b/atb_operator/AtbTrilSample/TrilOp/op_kernel/tril.cpp deleted file mode 100755 index 71711b6fa..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/op_kernel/tril.cpp +++ /dev/null @@ -1,278 +0,0 @@ -/* Copyright (c) Huawei Technologies Co., Ltd. 2020-2021. All rights reserved. - * - * This program is free software; you can redistribute it and/or modify - * it under the terms of the Apache License Version 2.0. - * You may not use this file except in compliance with the License. - * - * This program is distributed in the hope that it will be useful, - * but WITHOUT ANY WARRANTY; without even the implied warranty of - * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. - */ - -#include "kernel_operator.h" -using namespace AscendC; -constexpr int32_t BUFFER_NUM = 2; // tensor num for each queue -constexpr int32_t minNum = 1; - -constexpr int keyOne = 1; -constexpr int keyTwo = 2; -constexpr int keyThree = 3; -constexpr int keyFour = 4; - -constexpr int computeBatchSize = 256; - -struct IntegrateParam{ - uint32_t totalLengthAligned; - int32_t matrixNum; - int32_t matrixSize; - int32_t rowLength; - int32_t columnLength; - int32_t diagVal; - int32_t loopCnt; - uint32_t fullTileLength; - uint32_t lastTileLength; - int32_t fullCnt; - int32_t lastCnt; - uint32_t alignNum; - uint32_t typeSize; -}; - -class KernelTril { -public: - __aicore__ inline KernelTril() {} - //only pass the length this one is assigned to - __aicore__ inline void Init(GM_ADDR x, GM_ADDR y, IntegrateParam& paramList, uint32_t key) - { - this->matrixNum = paramList.matrixNum; - this->matrixSize = paramList.matrixSize; - this->rowLength = paramList.rowLength; - this->columnLength = paramList.columnLength; - this->diagVal = paramList.diagVal; - this->fullCnt = paramList.fullCnt; - this->lastCnt = paramList.lastCnt; - if(paramList.columnLength==0){ - paramList.columnLength = minNum; - } - this->fullRowInc = paramList.fullTileLength / paramList.columnLength; - this->initLength = 1; - // The result would not be the expected - if(paramList.typeSize==0){ - paramList.typeSize = sizeof(float); - } - - this->typeSize = paramList.typeSize; - - this->key=key; - - uint64_t gmBuffer=paramList.totalLengthAligned; - - xGm.SetGlobalBuffer((__gm__ DTYPE_X*)x, gmBuffer); - yGm.SetGlobalBuffer((__gm__ DTYPE_X*)y, gmBuffer); - - this->loopCnt = paramList.loopCnt; - this->fullTileLength = paramList.fullTileLength; - this->lastTileLength = paramList.lastTileLength; - - uint32_t singleBuffer = paramList.fullTileLength; - if(singleBuffer < paramList.lastTileLength){ - singleBuffer = paramList.lastTileLength; - } - if(key==keyThree || key==keyFour){ - pipe.InitBuffer(inQueueX, BUFFER_NUM, singleBuffer * this->typeSize); - pipe.InitBuffer(outQueueY, BUFFER_NUM, singleBuffer * this->typeSize); - }else{ - pipe.InitBuffer(queBind, BUFFER_NUM, singleBuffer * this->typeSize); - } - } - - __aicore__ inline void Process() - { - if(this->key==keyOne){ - NaivePath(); - }else if(this->key==keyTwo){ - SheerDup(); - }else if(this->key==keyThree){ - SheerZero(); - }else if(key==keyFour){ - FastPath(); - } - } - -private: - __aicore__ inline void SheerDup() - { - uint32_t GmOffset=0; - for (int i = 0; i < this->loopCnt-1; i++, GmOffset+=this->fullTileLength) { - auto bindLocal = queBind.AllocTensor(); - DataCopy(bindLocal, xGm[GmOffset], this->fullTileLength); - queBind.EnQue(bindLocal); - bindLocal = queBind.DeQue(); - DataCopy(yGm[GmOffset], bindLocal, this->fullTileLength); - queBind.FreeTensor(bindLocal); - } - auto bindLocal = queBind.AllocTensor(); - DataCopy(bindLocal, xGm[GmOffset], this->lastTileLength); - queBind.EnQue(bindLocal); - bindLocal = queBind.DeQue(); - DataCopy(yGm[GmOffset], bindLocal, this->lastTileLength); - queBind.FreeTensor(bindLocal); - } - - __aicore__ inline void SheerZero(){ - uint32_t GmOffset=0; - for (int i = 0; i < this->loopCnt-1; i++, GmOffset+=this->fullTileLength) { - CopyIn(GmOffset,this->fullTileLength); - AllZero(this->fullTileLength); - CopyOut(GmOffset,this->fullTileLength); - } - CopyIn(GmOffset,this->lastTileLength); - AllZero(this->lastTileLength); - CopyOut(GmOffset,this->lastTileLength); - } - - __aicore__ inline void NaivePath(){ - int32_t cnt=0; - for(int32_t i=0;imatrixNum;i++){ - for(int32_t j=0;jrowLength;j++){ - int32_t k=0; - while(kcolumnLength && k-j<=this->diagVal){ - DTYPE_X curr=xGm.GetValue(cnt); - yGm.SetValue(cnt,curr); - k++; - cnt++; - } - while(kcolumnLength){ - yGm.SetValue(cnt,(DTYPE_X)0); - k++; - cnt++; - } - } - } - } - - __aicore__ inline void FastPath(){ - uint32_t GmOffset=0; - int32_t init_row = 0; - for(int num=0;nummatrixNum;num++){ - uint32_t calLength=this->initLength; - if(this->diagVal<=0){ - init_row = -diagVal; - } - for (int32_t i = 0; i < this->loopCnt-1; i++) { - CopyIn(GmOffset,this->fullTileLength); - Compute(this->fullCnt, calLength, init_row); - CopyOut(GmOffset,this->fullTileLength); - if(init_row>0){ - init_row-=this->fullRowInc; - if(init_row<0){ - calLength-=init_row; - } - }else{ - calLength+=this->fullRowInc; - } - GmOffset+=this->fullTileLength; - } - CopyIn(GmOffset,this->lastTileLength); - Compute(this->lastCnt, calLength, init_row); - CopyOut(GmOffset,this->lastTileLength); - GmOffset+=this->lastTileLength; - } - } - - __aicore__ inline void CopyIn(uint32_t GmOffset, uint32_t tileLength){ - auto xLocal = inQueueX.AllocTensor(); - DataCopy(xLocal, xGm[GmOffset], tileLength); - inQueueX.EnQue(xLocal); - } - - __aicore__ inline void CopyOut(uint32_t GmOffset, uint32_t tileLength){ - auto yLocal=outQueueY.DeQue(); - DataCopy(yGm[GmOffset], yLocal, tileLength); - outQueueY.FreeTensor(yLocal); - } - - __aicore__ inline void Compute(int32_t cnt, uint32_t initLength, int32_t adjust){ - auto xLocal = inQueueX.DeQue(); - auto yLocal = outQueueY.AllocTensor(); - uint32_t localOffset=0; - uint32_t currLength=initLength; - DTYPE_X scalarZero=0; - for(int32_t i=0;icolumnLength); - if(i>=adjust){ - Adds(yLocal[localOffset],xLocal[localOffset],scalarZero,currLength); - currLength++; - } - localOffset+=this->columnLength; - } - outQueueY.EnQue(yLocal); - inQueueX.FreeTensor(xLocal); - } - - __aicore__ inline void AllZero(uint32_t tileLength){ - auto xLocal = inQueueX.DeQue(); - auto yLocal = outQueueY.AllocTensor(); - Sub(yLocal,xLocal,xLocal,tileLength); - outQueueY.EnQue(yLocal); - inQueueX.FreeTensor(xLocal); - } - -private: - TPipe pipe; - //queue for simple duplication - TQueBind queBind; // Use TQueBind to replace QueI,QueO if needed - - TQue inQueueX; - TQue outQueueY; - - GlobalTensor xGm; - GlobalTensor yGm; - - int32_t matrixNum; - int32_t matrixSize; - int32_t rowLength; - int32_t columnLength; - int32_t diagVal; - int32_t fullCnt; - int32_t lastCnt; - - int32_t loopCnt; - uint32_t fullTileLength; - uint32_t lastTileLength; - uint32_t fullRowInc; - uint32_t initLength; - - uint32_t typeSize; - uint32_t alignNum; - uint32_t key; -}; - -extern "C" __global__ __aicore__ void tril(GM_ADDR x, GM_ADDR y, GM_ADDR workspace, GM_ADDR tiling) { - GET_TILING_DATA(tiling_data, tiling); - KernelTril op; - IntegrateParam paramList = { - .totalLengthAligned=tiling_data.totalLengthAligned, - .matrixNum=tiling_data.matrixNum, - .matrixSize=tiling_data.matrixSize, - .rowLength=tiling_data.rowLength, - .columnLength=tiling_data.columnLength, - .diagVal=tiling_data.diagVal, - .loopCnt=tiling_data.loopCnt, - .fullTileLength=tiling_data.fullTileLength, - .lastTileLength=tiling_data.lastTileLength, - .fullCnt=tiling_data.fullCnt, - .lastCnt=tiling_data.lastCnt, - .alignNum=tiling_data.alignNum, - .typeSize=tiling_data.typeSize - }; - if(TILING_KEY_IS(1)){ - op.Init(x, y, paramList, 1); - }else if(TILING_KEY_IS(2)){ - op.Init(x, y, paramList, 2); - }else if(TILING_KEY_IS(3)){ - op.Init(x, y, paramList, 3); - }else if(TILING_KEY_IS(4)){ - op.Init(x, y, paramList, 4); - } - op.Process(); -} diff --git a/atb_operator/AtbTrilSample/TrilOp/scripts/install.sh b/atb_operator/AtbTrilSample/TrilOp/scripts/install.sh deleted file mode 100755 index 8468c5a25..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/scripts/install.sh +++ /dev/null @@ -1,318 +0,0 @@ -#!/bin/bash -vendor_name=customize -targetdir=/usr/local/Ascend/opp -target_custom=0 - -sourcedir=$PWD/packages -vendordir=vendors/$vendor_name - -QUIET="y" - -while true -do - case $1 in - --quiet) - QUIET="y" - shift - ;; - --install-path=*) - INSTALL_PATH=$(echo $1 | cut -d"=" -f2-) - INSTALL_PATH=${INSTALL_PATH%*/} - shift - ;; - --*) - shift - ;; - *) - break - ;; - esac -done - -log() { - cur_date=`date +"%Y-%m-%d %H:%M:%S"` - echo "[runtime] [$cur_date] "$1 -} - -if [ -n "${INSTALL_PATH}" ]; then - if [[ ! "${INSTALL_PATH}" = /* ]]; then - log "[ERROR] use absolute path for --install-path argument" - exit 1 - fi - if [ ! -d ${INSTALL_PATH} ]; then - mkdir ${INSTALL_PATH} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${INSTALL_PATH} failed" - exit 1 - fi - fi - targetdir=${INSTALL_PATH} -elif [ -n "${ASCEND_CUSTOM_OPP_PATH}" ]; then - if [ ! -d ${ASCEND_CUSTOM_OPP_PATH} ]; then - mkdir -p ${ASCEND_CUSTOM_OPP_PATH} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${ASCEND_CUSTOM_OPP_PATH} failed" - fi - fi - targetdir=${ASCEND_CUSTOM_OPP_PATH} -else - if [ "x${ASCEND_OPP_PATH}" == "x" ]; then - log "[ERROR] env ASCEND_OPP_PATH no exist" - exit 1 - fi - targetdir="${ASCEND_OPP_PATH}" -fi - -if [ ! -d $targetdir ];then - log "[ERROR] $targetdir no exist" - exit 1 -fi - -upgrade() -{ - if [ ! -d ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 files" - return 0 - fi - - if [ ! -d ${targetdir}/$vendordir/$1 ];then - log "[INFO] create ${targetdir}/$vendordir/$1." - mkdir -p ${targetdir}/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/$1 failed" - return 1 - fi - else - has_same_file=-1 - for file_a in ${sourcedir}/$vendordir/$1/*; do - file_b=${file_a##*/}; - if [ "ls ${targetdir}/$vendordir/$1" = "" ]; then - log "[INFO] ${targetdir}/$vendordir/$1 is empty !!" - return 1 - fi - grep -q $file_b <<<`ls ${targetdir}/$vendordir/$1`; - if [[ $? -eq 0 ]]; then - echo -n "${file_b} " - has_same_file=0 - fi - done - if [ 0 -eq $has_same_file ]; then - if test $QUIET = "n"; then - echo "[INFO]: has old version in ${targetdir}/$vendordir/$1, \ - you want to Overlay Installation , please enter:[o]; \ - or replace directory installation , please enter: [r]; \ - or not install , please enter:[n]." - - while true - do - read orn - if [ "$orn" = n ]; then - return 0 - elif [ "$orn" = m ]; then - break; - elif [ "$0rn" = r ]; then - [ -n "${targetdir}/$vendordir/$1/" ] && rm -rf "${targetdir}/$vendordir/$1"/* - break; - else - echo "[ERROR] input error, please input again!" - fi - done - fi - fi - log "[INFO] replace or merge old ops $1 files .g....." - fi - - log "copy new ops $1 files ......" - if [ -d ${targetdir}/$vendordir/$1/ ]; then - chmod -R +w "$targetdir/$vendordir/$1/" >/dev/null 2>&1 - fi - cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 files failed" - return 1 - fi - - return 0 -} -upgrade_proto() -{ - if [ ! -f ${sourcedir}/$vendordir/custom.proto ]; then - log "[INFO] no need to upgrade custom.proto files" - return 0 - fi - if [ ! -d ${targetdir}/$vendordir/framework/caffe ];then - log "[INFO] create ${targetdir}/$vendordir/framework/caffe." - mkdir -p ${targetdir}/$vendordir/framework/caffe - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/framework/caffe failed" - return 1 - fi - else - if [ -f ${targetdir}/$vendordir/framework/caffe/custom.proto ]; then - # 有老版本,判断是否要覆盖式安装 - if test $QUIET = "n"; then - echo "[INFO] ${targetdir}/$vendordir/framework/caffe has old version"\ - "custom.proto file. Do you want to replace? [y/n] " - - while true - do - read yn - if [ "$yn" = n ]; then - return 0 - elif [ "$yn" = y ]; then - break; - else - echo "[ERROR] input error, please input again!" - fi - done - fi - fi - log "[INFO] replace old caffe.proto files ......" - fi - chmod -R +w "$targetdir/$vendordir/framework/caffe/" >/dev/null 2>&1 - cp -rf ${sourcedir}/$vendordir/custom.proto ${targetdir}/$vendordir/framework/caffe/ - if [ $? -ne 0 ];then - log "[ERROR] copy new custom.proto failed" - return 1 - fi - log "[INFO] copy custom.proto success" - - return 0 -} - -upgrade_file() -{ - if [ ! -e ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 file" - return 0 - fi - - log "copy new $1 files ......" - cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 file failed" - return 1 - fi - - return 0 -} - -delete_optiling_file() -{ - if [ ! -d ${targetdir}/vendors ];then - log "[INFO] $1 not exist, no need to uninstall" - return 0 - fi - sys_info=$(uname -m) - if [ ! -d ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/lib/linux/${sys_info} ];then - rm -rf ${sourcedir}/$vendordir/$1/ai_core/tbe/op_tiling/liboptiling.so - fi - return 0 -} - -log "[INFO] copy uninstall sh success" - -if [ ! -d ${targetdir}/vendors ];then - log "[INFO] create ${targetdir}/vendors." - mkdir -p ${targetdir}/vendors - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/vendors failed" - return 1 - fi -fi -chmod u+w ${targetdir}/vendors - -echo "[ops_custom]upgrade framework" -upgrade framework -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op proto" -upgrade op_proto -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade version.info" -upgrade_file version.info -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op impl" -delete_optiling_file op_impl -upgrade op_impl -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op api" -upgrade op_api -if [ $? -ne 0 ];then - exit 1 -fi - -upgrade_proto -if [ $? -ne 0 ];then - exit 1 -fi - -# set the set_env.bash -if [ -n "${INSTALL_PATH}" ] && [ -d ${INSTALL_PATH} ]; then - _ASCEND_CUSTOM_OPP_PATH=${targetdir}/${vendordir} - bin_path="${_ASCEND_CUSTOM_OPP_PATH}/bin" - set_env_variable="#!/bin/bash\nexport ASCEND_CUSTOM_OPP_PATH=${_ASCEND_CUSTOM_OPP_PATH}:\${ASCEND_CUSTOM_OPP_PATH}" - if [ ! -d ${bin_path} ]; then - mkdir -p ${bin_path} >> /dev/null 2>&1 - if [ $? -ne 0 ]; then - log "[ERROR] create ${bin_path} failed" - exit 1 - fi - fi - echo -e ${set_env_variable} > ${bin_path}/set_env.bash - if [ $? -ne 0 ]; then - log "[ERROR] write ASCEND_CUSTOM_OPP_PATH to set_env.bash failed" - exit 1 - else - log "[INFO] using requirements: when custom module install finished or before you run the custom module, \ - execute the command [ source ${bin_path}/set_env.bash ] to set the environment path" - fi -else - config_file=${targetdir}/vendors/config.ini - if [ ! -f ${config_file} ]; then - touch ${config_file} - chmod 640 ${config_file} - echo "load_priority=$vendor_name" > ${config_file} - if [ $? -ne 0 ];then - echo "echo load_priority failed" - exit 1 - fi - else - found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" - found_vendor=$(echo $found_vendors | sed "s/$vendor_name//g" | tr ',' ' ') - vendor=$(echo $found_vendor | tr -s ' ' ',') - if [ "$vendor" != "" ]; then - sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" - fi - fi -fi - -chmod u-w ${targetdir}/vendors - -if [ -d ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/ ]; then - chmod -R 440 ${targetdir}/$vendordir/op_impl/cpu/aicpu_kernel/impl/* >/dev/null 2>&1 -fi -if [ -f ${targetdir}/ascend_install.info ]; then - chmod -R 440 ${targetdir}/ascend_install.info -fi -if [ -f ${targetdir}/scene.info ]; then - chmod -R 440 ${targetdir}/scene.info -fi -if [ -f ${targetdir}/version.info ]; then - chmod -R 440 ${targetdir}/version.info -fi - -echo "SUCCESS" -exit 0 - diff --git a/atb_operator/AtbTrilSample/TrilOp/scripts/upgrade.sh b/atb_operator/AtbTrilSample/TrilOp/scripts/upgrade.sh deleted file mode 100755 index e09173485..000000000 --- a/atb_operator/AtbTrilSample/TrilOp/scripts/upgrade.sh +++ /dev/null @@ -1,151 +0,0 @@ -#!/bin/bash -vendor_name=customize -targetdir=/usr/local/Ascend/opp -target_custom=0 - -sourcedir=$PWD/packages -vendordir=vendors/$vendor_name - -log() { - cur_date=`date +"%Y-%m-%d %H:%M:%S"` - echo "[runtime] [$cur_date] "$1 -} - -if [[ "x${ASCEND_OPP_PATH}" == "x" ]];then - log "[ERROR] env ASCEND_OPP_PATH no exist" - exit 1 -fi - -targetdir=${ASCEND_OPP_PATH} - -if [ ! -d $targetdir ];then - log "[ERROR] $targetdir no exist" - exit 1 -fi - -upgrade() -{ - if [ ! -d ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 files" - return 0 - fi - - if [ ! -d ${targetdir}/$vendordir/$1 ];then - log "[INFO] create ${targetdir}/$vendordir/$1." - mkdir -p ${targetdir}/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] create ${targetdir}/$vendordir/$1 failed" - return 1 - fi - else - vendor_installed_dir=$(ls "$targetdir/vendors" 2> /dev/null) - for i in $vendor_installed_dir;do - vendor_installed_file=$(ls "$vendor_installed_dir/$vendor_name/$i" 2> /dev/null) - if [ "$i" = "$vendor_name" ] && [ "$vendor_installed_file" != "" ]; then - echo "[INFO]: $vendor_name custom opp package has been installed on the path $vendor_installed_dir, \ - you want to Overlay Installation , please enter:[o]; \ - or replace directory installation , please enter: [r]; \ - or not install , please enter:[n]." - fi - while true - do - read mrn - if [ "$mrn" = m ]; then - break - elif [ "$mrn" = r ]; then - [ -n "$vendor_installed_file"] && rm -rf "$vendor_installed_file" - break - elif [ "$mrn" = n ]; then - return 0 - else - echo "[WARNING]: Input error, please input m or r or n to choose!" - fi - done - done - log "[INFO] replace old ops $1 files ......" - fi - - log "copy new ops $1 files ......" - cp -rf ${sourcedir}/$vendordir/$1/* $targetdir/$vendordir/$1/ - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 files failed" - return 1 - fi - - return 0 -} - -upgrade_file() -{ - if [ ! -e ${sourcedir}/$vendordir/$1 ]; then - log "[INFO] no need to upgrade ops $1 file" - return 0 - fi - - log "copy new $1 files ......" - cp -f ${sourcedir}/$vendordir/$1 $targetdir/$vendordir/$1 - if [ $? -ne 0 ];then - log "[ERROR] copy new $1 file failed" - return 1 - fi - - return 0 -} - -log "[INFO] copy uninstall sh success" - -echo "[ops_custom]upgrade framework" -upgrade framework -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op proto" -upgrade op_proto -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op impl" -upgrade op_impl -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade op api" -upgrade op_api -if [ $? -ne 0 ];then - exit 1 -fi - -echo "[ops_custom]upgrade version.info" -upgrade_file version.info -if [ $? -ne 0 ];then - exit 1 -fi - -config_file=${targetdir}/vendors/config.ini -found_vendors="$(grep -w "load_priority" "$config_file" | cut --only-delimited -d"=" -f2-)" -found_vendor=$(echo $found_vendors | sed "s/$vendor_name//g" | tr ',' ' ') -vendor=$(echo $found_vendor | tr -s ' ' ',') -if [ "$vendor" != "" ]; then - sed -i "/load_priority=$found_vendors/s@load_priority=$found_vendors@load_priority=$vendor_name,$vendor@g" "$config_file" -fi - -changemode() -{ - if [ -d ${targetdir} ];then - chmod -R 550 ${targetdir}>/dev/null 2>&1 - fi - - return 0 -} -echo "[ops_custom]changemode..." -#changemode -if [ $? -ne 0 ];then - exit 1 -fi - -echo "SUCCESS" -exit 0 - -- Gitee From 8a7d58ca22c56bda0d007ceffefcfde964e74e41 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Tue, 1 Apr 2025 03:16:47 +0000 Subject: [PATCH 10/26] update atb_operator/AtbAddSample/readme.md. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: 万祖涛 --- atb_operator/AtbAddSample/readme.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/atb_operator/AtbAddSample/readme.md b/atb_operator/AtbAddSample/readme.md index c1d8752db..8997e89d7 100644 --- a/atb_operator/AtbAddSample/readme.md +++ b/atb_operator/AtbAddSample/readme.md @@ -15,7 +15,7 @@ ### Add AscendC自定义算子部署 -参照[cann-ops](https://gitee.com/ascend/cann-ops/blob/master/README.md) +参照[cann-ops仓add_custom算子](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)" **算子包编译部署** "章节 ### AddOperation ATB插件部署 -- Gitee From 6aea8cd26a2473ac9cbea2fbe17a8c1c5e9356c5 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Tue, 1 Apr 2025 03:17:18 +0000 Subject: [PATCH 11/26] update atb_operator/AtbAddSample/readme.md. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: 万祖涛 --- atb_operator/AtbAddSample/readme.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/atb_operator/AtbAddSample/readme.md b/atb_operator/AtbAddSample/readme.md index 8997e89d7..e0a38627c 100644 --- a/atb_operator/AtbAddSample/readme.md +++ b/atb_operator/AtbAddSample/readme.md @@ -15,7 +15,7 @@ ### Add AscendC自定义算子部署 -参照[cann-ops仓add_custom算子](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)" **算子包编译部署** "章节 +参照cann-ops仓[add_custom算子](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)" **算子包编译部署** "章节 ### AddOperation ATB插件部署 -- Gitee From 6d31f27b2d4d26a4600c2e66b261f98268f47598 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Tue, 1 Apr 2025 03:31:45 +0000 Subject: [PATCH 12/26] =?UTF-8?q?=E9=87=8D=E5=91=BD=E5=90=8D=20atb=5Fopera?= =?UTF-8?q?tor=20=E4=B8=BA=20atb=5Fplugin?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .../AtbAddSample/AddOperationATBPlugin/CMakeLists.txt | 0 .../AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp | 0 .../AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h | 0 .../AtbAddSample/AddOperationATBPlugin/build.sh | 0 .../AtbAddSample/AddOperationTest/CMakeLists.txt | 0 .../AtbAddSample/AddOperationTest/script/gen_data.py | 0 .../AtbAddSample/AddOperationTest/script/run.sh | 0 .../AtbAddSample/AddOperationTest/script/verify_result.py | 0 .../AtbAddSample/AddOperationTest/src/main.cpp | 0 .../AtbAddSample/AddOperationTest/src/main.h | 0 {atb_operator => atb_plugin}/AtbAddSample/readme.md | 0 {atb_operator => atb_plugin}/README.md | 0 12 files changed, 0 insertions(+), 0 deletions(-) rename {atb_operator => atb_plugin}/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt (100%) rename {atb_operator => atb_plugin}/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp (100%) rename {atb_operator => atb_plugin}/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h (100%) rename {atb_operator => atb_plugin}/AtbAddSample/AddOperationATBPlugin/build.sh (100%) rename {atb_operator => atb_plugin}/AtbAddSample/AddOperationTest/CMakeLists.txt (100%) rename {atb_operator => atb_plugin}/AtbAddSample/AddOperationTest/script/gen_data.py (100%) rename {atb_operator => atb_plugin}/AtbAddSample/AddOperationTest/script/run.sh (100%) rename {atb_operator => atb_plugin}/AtbAddSample/AddOperationTest/script/verify_result.py (100%) rename {atb_operator => atb_plugin}/AtbAddSample/AddOperationTest/src/main.cpp (100%) rename {atb_operator => atb_plugin}/AtbAddSample/AddOperationTest/src/main.h (100%) rename {atb_operator => atb_plugin}/AtbAddSample/readme.md (100%) rename {atb_operator => atb_plugin}/README.md (100%) diff --git a/atb_operator/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt b/atb_plugin/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt similarity index 100% rename from atb_operator/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt rename to atb_plugin/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt diff --git a/atb_operator/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp b/atb_plugin/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp similarity index 100% rename from atb_operator/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp rename to atb_plugin/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp diff --git a/atb_operator/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h b/atb_plugin/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h similarity index 100% rename from atb_operator/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h rename to atb_plugin/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h diff --git a/atb_operator/AtbAddSample/AddOperationATBPlugin/build.sh b/atb_plugin/AtbAddSample/AddOperationATBPlugin/build.sh similarity index 100% rename from atb_operator/AtbAddSample/AddOperationATBPlugin/build.sh rename to atb_plugin/AtbAddSample/AddOperationATBPlugin/build.sh diff --git a/atb_operator/AtbAddSample/AddOperationTest/CMakeLists.txt b/atb_plugin/AtbAddSample/AddOperationTest/CMakeLists.txt similarity index 100% rename from atb_operator/AtbAddSample/AddOperationTest/CMakeLists.txt rename to atb_plugin/AtbAddSample/AddOperationTest/CMakeLists.txt diff --git a/atb_operator/AtbAddSample/AddOperationTest/script/gen_data.py b/atb_plugin/AtbAddSample/AddOperationTest/script/gen_data.py similarity index 100% rename from atb_operator/AtbAddSample/AddOperationTest/script/gen_data.py rename to atb_plugin/AtbAddSample/AddOperationTest/script/gen_data.py diff --git a/atb_operator/AtbAddSample/AddOperationTest/script/run.sh b/atb_plugin/AtbAddSample/AddOperationTest/script/run.sh similarity index 100% rename from atb_operator/AtbAddSample/AddOperationTest/script/run.sh rename to atb_plugin/AtbAddSample/AddOperationTest/script/run.sh diff --git a/atb_operator/AtbAddSample/AddOperationTest/script/verify_result.py b/atb_plugin/AtbAddSample/AddOperationTest/script/verify_result.py similarity index 100% rename from atb_operator/AtbAddSample/AddOperationTest/script/verify_result.py rename to atb_plugin/AtbAddSample/AddOperationTest/script/verify_result.py diff --git a/atb_operator/AtbAddSample/AddOperationTest/src/main.cpp b/atb_plugin/AtbAddSample/AddOperationTest/src/main.cpp similarity index 100% rename from atb_operator/AtbAddSample/AddOperationTest/src/main.cpp rename to atb_plugin/AtbAddSample/AddOperationTest/src/main.cpp diff --git a/atb_operator/AtbAddSample/AddOperationTest/src/main.h b/atb_plugin/AtbAddSample/AddOperationTest/src/main.h similarity index 100% rename from atb_operator/AtbAddSample/AddOperationTest/src/main.h rename to atb_plugin/AtbAddSample/AddOperationTest/src/main.h diff --git a/atb_operator/AtbAddSample/readme.md b/atb_plugin/AtbAddSample/readme.md similarity index 100% rename from atb_operator/AtbAddSample/readme.md rename to atb_plugin/AtbAddSample/readme.md diff --git a/atb_operator/README.md b/atb_plugin/README.md similarity index 100% rename from atb_operator/README.md rename to atb_plugin/README.md -- Gitee From 0c596fbac504b8b451ffdfc15e68ec808ee1ebe7 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E5=88=98=E5=96=9C=E5=BC=BA?= Date: Tue, 1 Apr 2025 07:07:47 +0000 Subject: [PATCH 13/26] category update --- .../AddOperationATBPlugin/CMakeLists.txt | 21 ++ .../aclnn_add_operation.cpp | 142 ++++++++++++ .../aclnn_add_operation.h | 56 +++++ .../AddOperationATBPlugin/build.sh | 33 +++ .../AddOperationTest/CMakeLists.txt | 40 ++++ .../AddOperationTest/script/gen_data.py | 20 ++ .../AddOperationTest/script/run.sh | 52 +++++ .../AddOperationTest/script/verify_result.py | 28 +++ .../AddOperationTest/src/main.cpp | 217 ++++++++++++++++++ .../ATBInvocation/AddOperationTest/src/main.h | 55 +++++ .../24_add_custom/ATBInvocation/readme.md | 70 ++++++ 11 files changed, 734 insertions(+) create mode 100644 operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/CMakeLists.txt create mode 100644 operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/aclnn_add_operation.cpp create mode 100644 operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/aclnn_add_operation.h create mode 100755 operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/build.sh create mode 100644 operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/CMakeLists.txt create mode 100644 operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/gen_data.py create mode 100644 operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/run.sh create mode 100644 operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/verify_result.py create mode 100644 operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp create mode 100644 operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.h create mode 100644 operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/CMakeLists.txt b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/CMakeLists.txt new file mode 100644 index 000000000..c89578fcd --- /dev/null +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/CMakeLists.txt @@ -0,0 +1,21 @@ +# Copyright 2024 Tencent Inc. All rights reserved. +# +# ============================================================================== +cmake_minimum_required(VERSION 3.14) +set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") +file(GLOB_RECURSE atb_plugin_operations_SRCS "./*.cpp") +set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") +message(STATUS "atb_plugin_operations_SRCS: ${atb_plugin_operations_SRCS}") +include_directories( + $ENV{ATB_HOME_PATH}/include + $ENV{ASCEND_HOME_PATH}/include + ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 + ${CMAKE_CURRENT_SOURCE_DIR} + ) +link_directories( + $ENV{ATB_HOME_PATH}/lib + $ENV{ASCEND_HOME_PATH}/lib64 + ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 + ) +add_library(atb_plugin_operations STATIC ${atb_plugin_operations_SRCS}) +target_link_libraries(atb_plugin_operations PRIVATE ${ACL_SHARED_LIBS} ${ACL_SHARED_LIBS} ascendcl) \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/aclnn_add_operation.cpp b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/aclnn_add_operation.cpp new file mode 100644 index 000000000..468b84aff --- /dev/null +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/aclnn_add_operation.cpp @@ -0,0 +1,142 @@ +#include "aclnn_add_operation.h" +#include "aclnn_add_custom.h" + +AddOperation::AddOperation(const std::string &name, AddAttrParam param){ + attrParam = param; + opName_ = name; +} + +atb::SVector GetCopyTensorStride(atb::Dims &tensorDims) +{ + atb::SVector tmpStrides(tensorDims.dimNum, 1); + if (tensorDims.dimNum > 8) { // 8: tensor最大维度数量 + printf("tensor's dimNum is larger than 8, GetCopyTensorStride failed."); + return tmpStrides; + } + for (int64_t i = static_cast(tensorDims.dimNum) - 2; i >= 0; i--) { + tmpStrides[i] = (tensorDims.dims[i + 1] * tmpStrides[i + 1]); + } + return tmpStrides; +} + +std::shared_ptr AddOperation::CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx) +{ + auto aclnnTensor = std::make_shared(); + aclnnTensor->tensorIdx = static_cast(tensorIdx); + aclnnTensor->needUpdateTensorDataPtr = true; + aclnnTensor->atbTensor = atbTensor; + aclnnTensor->strides = GetCopyTensorStride(atbTensor.desc.shape); + + // 创建Aclnn tensor + aclnnTensor->tensor = aclCreateTensor(atbTensor.desc.shape.dims, + atbTensor.desc.shape.dimNum, + atbTensor.desc.dtype, + aclnnTensor->strides.data(), + 0, + atbTensor.desc.format, + atbTensor.desc.shape.dims, + atbTensor.desc.shape.dimNum, + atbTensor.deviceData); + return aclnnTensor; +} + +atb::Status AddOperation::UpdateAclnnVariantPack(const atb::VariantPack &variantPack) +{ + // 更新inTensor的device地址 + for (size_t i = 0; i < aclInTensors_.size(); ++i) { + int ret = -1; + if (!aclInTensors_[i]->needUpdateTensorDataPtr) { + continue; + } + aclInTensors_[i]->atbTensor = variantPack.inTensors.at(i); + ret = aclSetInputTensorAddr(aclExecutor_, + aclInTensors_[i]->tensorIdx, + aclInTensors_[i]->tensor, + aclInTensors_[i]->atbTensor.deviceData); + + if (ret != 0) { + printf("set input fail"); + return atb::ERROR_CANN_ERROR; + } + } + + // 更新outTensor的device地址 + for (size_t i = 0; i < aclOutTensors_.size(); ++i) { + int ret = -1; + if (!aclOutTensors_[i]->needUpdateTensorDataPtr) { + continue; + } + aclOutTensors_[i]->atbTensor = variantPack.outTensors.at(i); + ret = aclSetOutputTensorAddr(aclExecutor_, + aclOutTensors_[i]->tensorIdx, + aclOutTensors_[i]->tensor, + aclOutTensors_[i]->atbTensor.deviceData); + + if (ret != 0) { + printf("set output fail"); + return atb::ERROR_CANN_ERROR; + } + } + return atb::NO_ERROR; +} + +atb::Status AddOperation::Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) { + + aclInTensors_.resize(GetInputNum()); + for (size_t i = 0; i < aclInTensors_.size(); ++i) { + auto aclnnTensor = CreateAclnnTensor(variantPack.inTensors.at(i), i); + if (aclnnTensor->tensor == nullptr) { + printf("creat input tensor %ld fail", i); + return atb::ERROR_INTERNAL_ERROR; + } + aclInTensors_[i] = aclnnTensor; + } + + aclOutTensors_.resize(GetOutputNum()); + for (size_t i = 0; i < aclOutTensors_.size(); ++i) { + auto aclnnTensor = CreateAclnnTensor(variantPack.outTensors.at(i), i); + if (aclnnTensor->tensor == nullptr) { + printf("creat output tensor %ld fail", i); + return atb::ERROR_INTERNAL_ERROR; + } + aclOutTensors_[i] = aclnnTensor; + } + + + auto ret = aclnnAddCustomGetWorkspaceSize(aclInTensors_.at(0)->tensor, + aclInTensors_.at(1)->tensor, + aclOutTensors_.at(0)->tensor, + &workspaceSize_, + &aclExecutor_); + + workspaceSize = workspaceSize_; + return ret; + +} + +atb::Status AddOperation::Execute(const atb::VariantPack &variantPack, uint8_t *workspace, uint64_t workspaceSize, atb::Context *context) { + + + + aclrtStream stream = context->GetExecuteStream(); + if (!stream) { + printf("get stream fail"); + return atb::ERROR_INVALID_PARAM; + } + // 更新数据传入的地址 + int ret = UpdateAclnnVariantPack(variantPack); + if (ret != 0) { + printf("UpdateAclnnVariantPack fail"); + return atb::ERROR_CANN_ERROR; + } + ret = aclnnAddCustom(workspace, workspaceSize_, aclExecutor_, stream); + + return ret; +} + +atb::Status AddOperation::InferShape( + const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const +{ + outTensorDesc.at(0) = inTensorDesc.at(0); + return atb::NO_ERROR; +} diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/aclnn_add_operation.h b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/aclnn_add_operation.h new file mode 100644 index 000000000..8670088d3 --- /dev/null +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/aclnn_add_operation.h @@ -0,0 +1,56 @@ +#include +#include +#include +#include +#include +#include "atb/infer_op_params.h" + + +struct AddAttrParam +{ + // add没属性,此处空 +}; + +struct AclnnTensor +{ +public: + atb::Tensor atbTensor; // + aclTensor *tensor = nullptr; + int tensorIdx = -1; // aclTensor在aclExecutor中的index + bool needUpdateTensorDataPtr = false; + atb::SVector strides = {}; +}; + +class AddOperation: public atb::Operation{ +public: + AddOperation(const std::string &name, AddAttrParam param); + atb::Status Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) override; + atb::Status Execute(const atb::VariantPack &variantPack, uint8_t *workspace, + uint64_t workspaceSize, atb::Context *context) override; + atb::Status InferShape( + const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const; + std::shared_ptr CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx); + atb::Status UpdateAclnnVariantPack(const atb::VariantPack &variantPack); + + uint32_t GetInputNum() const + { + return 2; // 算子入参个数 + } + + uint32_t GetOutputNum() const + { + return 1; // 算子出参个数 + } + std::string GetName() const + { + return opName_; + } + + aclOpExecutor *aclExecutor_ = nullptr; + AddAttrParam attrParam; + std::string opName_; + uint64_t workspaceSize_; + + atb::SVector> aclInTensors_; + atb::SVector> aclOutTensors_; +}; \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/build.sh b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/build.sh new file mode 100755 index 000000000..ba28cc143 --- /dev/null +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/build.sh @@ -0,0 +1,33 @@ +#!/bin/bash + +# 定义构建目录 +BUILD_DIR="build" + +# 创建构建目录 +mkdir -p "$BUILD_DIR" +cd "$BUILD_DIR" + +# 运行 CMake 配置和编译 +cmake .. +make + +# 查找生成的 .a 文件 +A_FILE=$(find . -name "*.a" -type f) + +# 检查是否找到了 .a 文件 +if [ -z "$A_FILE" ]; then + echo "未找到 .a 文件,编译可能失败。" + exit 1 +fi + +# 复制头文件到 /usr/include +HEADER_FILES=$(find .. -name "*.h" -type f) +for header in $HEADER_FILES; do + cp "$header" /usr/include/ +done + +# 复制 .a 文件到 /usr/local/lib +cp "$A_FILE" /usr/local/lib/ + +echo "构建完成,头文件和 .a 文件已复制到目标目录。" + diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/CMakeLists.txt b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/CMakeLists.txt new file mode 100644 index 000000000..b9c2cec46 --- /dev/null +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/CMakeLists.txt @@ -0,0 +1,40 @@ +cmake_minimum_required(VERSION 3.16) +project("test_model") +option(USE_CXX11_ABI "USE_CXX11_ABI" OFF) +set(CMAKE_BUILD_TYPE Debug) +if(USE_CXX11_ABI) + set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=1") +else() + set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") +endif() + +set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -std=c++17") + +set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") + +include_directories( + $ENV{ATB_HOME_PATH}/include + $ENV{ASCEND_HOME_PATH}/include + ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 + ${CMAKE_CURRENT_SOURCE_DIR} + ./src/ + ) + +message(".h path = ${CUST_PKG_PATH}/include") +link_directories( + $ENV{ATB_HOME_PATH}/lib + $ENV{ASCEND_HOME_PATH}/lib64 + ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 + ) + + +link_libraries(atb_plugin_operations) +add_executable(test_model ./src/main.cpp) + +# 列出所有的头文件目录 +target_link_libraries(test_model PRIVATE atb ascendcl cust_opapi opapi nnopbase pthread) # 添加自定义算子so及适配so + + + + + diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/gen_data.py b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/gen_data.py new file mode 100644 index 000000000..83ecafbce --- /dev/null +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/gen_data.py @@ -0,0 +1,20 @@ +import torch +import os +def gen_golden_data_simple(): + dtype = "float" + input1 = torch.randn(133, 4095, dtype=torch.float) + input2 = torch.randn(133, 4095, dtype=torch.float) + + + golden = input1 + input2 + os.system("mkdir -p input") + os.system("mkdir -p output") + input1.numpy().tofile('./script/input/input0.bin') + input2.numpy().tofile('./script/input/input1.bin') + golden.numpy().tofile("./script/output/golden0.bin") + + with open("./output/meta", "w") as fp: + print(dtype, file=fp) + +if __name__ == "__main__": + gen_golden_data_simple() diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/run.sh b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/run.sh new file mode 100644 index 000000000..450b717f2 --- /dev/null +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/run.sh @@ -0,0 +1,52 @@ +#!/bin/bash + + +# 步骤1: 运行gen_data.py生成输入bin文件和golden标杆输出数据 +echo "正在生成输入数据和golden标杆数据..." +mkdir -p script/input +mkdir -p script/output +python3 script/gen_data.py +if [ $? -ne 0 ]; then + echo "生成数据失败,脚本终止。" + exit 1 +fi + +# 步骤2: 创建构建目录并进入 +mkdir -p build +cd build +if [ $? -ne 0 ]; then + echo "无法进入构建目录,脚本终止。" + exit 1 +fi + +# 步骤3: 使用CMake配置项目 +echo "正在配置CMake项目..." +cmake .. +if [ $? -ne 0 ]; then + echo "CMake配置失败,脚本终止。" + exit 1 +fi + +# 步骤4: 编译代码 +echo "正在编译代码..." +make +if [ $? -ne 0 ]; then + echo "编译失败,脚本终止。" + exit 1 +fi + +mv test_model ../ +cd .. + +# 步骤5: 运行可执行文件生成实际输出文件 +echo "正在运行可执行文件生成实际输出..." +./test_model +if [ $? -ne 0 ]; then + echo "运行可执行文件失败,脚本终止。" + exit 1 +fi + +# 步骤6: 调用verify_result.py进行golden标杆数据和实际输出数据的比对 +echo "正在验证结果..." +python3 script/verify_result.py script/output/output_0.bin script/output/golden0.bin + diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/verify_result.py b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/verify_result.py new file mode 100644 index 000000000..629845e13 --- /dev/null +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/verify_result.py @@ -0,0 +1,28 @@ +import os +import sys +import numpy + +loss = 1e-3 # 容忍偏差,一般fp16要求绝对误差和相对误差均不超过千分之一 +minimum = 10e-10 + +def verify_result(real_result, golden): + with open("output/meta", "r") as fp: + dtype_str = fp.readline().strip() + dtype = eval(dtype_str) + real_result = numpy.fromfile(real_result, dtype=dtype) # 从bin文件读取实际运算结果 + golden = numpy.fromfile(golden, dtype=dtype) # 从bin文件读取预期运算结果 + print("=" * 50, real_result[:5], golden[:5], "=" * 50, sep='\n', end='\n', file=sys.stderr) + result = numpy.abs(real_result - golden) # 计算运算结果和预期结果偏差 + deno = numpy.maximum(numpy.abs(real_result), numpy.abs(golden)) # 获取最大值并组成新数组 + result_atol = numpy.less_equal(result, loss) # 计算绝对误差 + result_rtol = numpy.less_equal(result / numpy.add(deno, minimum), loss) # 计算相对误差 + if not result_rtol.all() and not result_atol.all(): + if numpy.sum(result_rtol == False) > real_result.size * loss and numpy.sum(result_atol == False) > real_result.size * loss: # 误差超出预期时返回打印错误,返回对比失败 + print("[ERROR] result error") + return False + print("test Operation success!") + return True + + +if __name__ == '__main__': + verify_result(sys.argv[1],sys.argv[2]) diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp new file mode 100644 index 000000000..258c0cb3c --- /dev/null +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp @@ -0,0 +1,217 @@ +/* + * Copyright (c) Huawei Technologies Co., Ltd. 2023. All rights reserved. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +#include "main.h" + +bool SetInputData(std::vector &inputData){ + char *xPath = "./script/input/input0.bin"; + char *yPath = "./script/input/input1.bin"; + InputData inputX; + InputData inputY; + inputX.data = ReadBinFile(xPath,inputX.size); + inputY.data = ReadBinFile(yPath,inputY.size); + inputData.push_back(inputX); + inputData.push_back(inputY); + return true; +} + +bool SetOperationInputDesc(atb::SVector &intensorDescs){ + atb::TensorDesc xDesc; + xDesc.dtype = ACL_FLOAT; + xDesc.format = ACL_FORMAT_ND; + xDesc.shape.dimNum = 2; + xDesc.shape.dims[0] = 133; + xDesc.shape.dims[1] = 4095; + + atb::TensorDesc yDesc; + yDesc.dtype = ACL_FLOAT; + yDesc.format = ACL_FORMAT_ND; + yDesc.shape.dimNum = 2; + yDesc.shape.dims[0] = 133; + yDesc.shape.dims[1] = 4095; + + intensorDescs.at(0) = xDesc; + intensorDescs.at(1) = yDesc; +} + + + +static void SetCurrentDevice() +{ + const int deviceId = 0; + std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; + int ret = aclrtSetDevice(deviceId); + if (ret != 0) { + std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; + return; + } + std::cout << "[INFO]: aclrtSetDevice success" << std::endl; +} + + +static void FreeTensor(atb::Tensor &tensor) +{ + if (tensor.deviceData) { + int ret = aclrtFree(tensor.deviceData); + if (ret != 0) { + std::cout << "[ERROR]: aclrtFree fail" << std::endl; + } + tensor.deviceData = nullptr; + tensor.dataSize = 0; + } + if (tensor.hostData) { + int ret = aclrtFreeHost(tensor.hostData); + if (ret != 0) { + std::cout << "[ERROR]: aclrtFreeHost fail, ret = " << ret << std::endl; + } + tensor.hostData = nullptr; + tensor.dataSize = 0; + } +} + +static void FreeTensors(atb::SVector &inTensors, atb::SVector &outTensors) +{ + for (size_t i = 0; i < inTensors.size(); ++i) { + FreeTensor(inTensors.at(i)); + } + for (size_t i = 0; i < outTensors.size(); ++i) { + FreeTensor(outTensors.at(i)); + } +} +bool SaveMemoryToBinFile(void* memoryAddress, size_t memorySize, size_t i) { + // 创建 output 目录(如果不存在) + std::filesystem::create_directories("output"); + + // 生成文件名 + std::string filename = "script/output/output_" + std::to_string(i) + ".bin"; + + // 打开文件以二进制写入模式 + std::ofstream file(filename, std::ios::binary); + if (!file) { + std::cerr << "无法打开文件: " << filename << std::endl; + return false; + } + + // 写入数据 + file.write(static_cast(memoryAddress), memorySize); + if (!file) { + std::cerr << "写入文件时出错: " << filename << std::endl; + file.close(); + return false; + } + + // 关闭文件 + file.close(); + std::cout << "数据已成功保存到: " << filename << std::endl; + return true; +} + +int main(int argc, const char *argv[]) +{ + const int deviceId = 0; + std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; + int ret = aclrtSetDevice(deviceId); + if (ret != 0) { + std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; + return 1; + } + std::cout << "[INFO]: aclrtSetDevice success" << std::endl; + atb::Context *context = nullptr; + ret = atb::CreateContext(&context); + void *stream = nullptr; + ret = aclrtCreateStream(&stream); + if (ret != 0) { + std::cout << "[ERROR]: AsdRtStreamCreate fail, ret:" << ret << std::endl; + return 1; + } + context->SetExecuteStream(stream); + + std::vector input; + SetInputData(input); + + AddAttrParam addAttrParam; + AddOperation *op = new AddOperation("Add",addAttrParam); + std::cout << "[INFO]: complete CreateOp!" << std::endl; + + if(input.size() != op->GetInputNum()) std::cout << "[ERROR]: Operation actual input num is not equal to GetInputNum()"; + + atb::SVector intensorDescs; + atb::SVector outtensorDescs; + intensorDescs.resize(op->GetInputNum()); + outtensorDescs.resize(op->GetOutputNum()); + SetOperationInputDesc(intensorDescs); + atb::Status st = op->InferShape(intensorDescs,outtensorDescs); + if (st != 0) { + std::cout << "[ERROR]: Operation InferShape fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation InferShape success" << std::endl; + atb::VariantPack variantPack; + variantPack.inTensors.resize(op->GetInputNum()); + variantPack.outTensors.resize(op->GetOutputNum()); + for(size_t i=0;iGetInputNum();i++){ + variantPack.inTensors.at(i).desc = intensorDescs.at(i); + variantPack.inTensors.at(i).hostData = input[i].data; + variantPack.inTensors.at(i).dataSize = input[i].size; + CheckAcl(aclrtMalloc(&variantPack.inTensors.at(i).deviceData, input[i].size, ACL_MEM_MALLOC_HUGE_FIRST)); + CheckAcl(aclrtMemcpy(variantPack.inTensors.at(i).deviceData, input[i].size, input[i].data, input[i].size, ACL_MEMCPY_HOST_TO_DEVICE)); + } + std::cout << "[INFO]: Operation Input prepare sucess" << std::endl; + for(size_t i=0;iGetOutputNum();i++){ + int64_t *dims = new int64_t[outtensorDescs.at(i).shape.dimNum]; + for(size_t j=0;jSetup(variantPack, workspaceSize, context); + if (st != 0) { + std::cout << "[ERROR]: Operation setup fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation setup success" << std::endl; + void *workspace = nullptr; + if (workspaceSize > 0) { + aclrtMalloc(&workspace, workspaceSize, ACL_MEM_MALLOC_HUGE_FIRST); + } + + std::cout << "[INFO]: Operation execute start" << std::endl; + st = op->Execute(variantPack, (uint8_t*)workspace, workspaceSize, context); + if (st != 0) { + std::cout << "[ERROR]: Operation execute fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation execute success" << std::endl; + for(size_t i = 0; i < op->GetOutputNum(); i++){ + CheckAcl(aclrtMemcpy(variantPack.outTensors.at(i).hostData, variantPack.outTensors.at(i).dataSize, variantPack.outTensors.at(0).deviceData, + variantPack.outTensors.at(i).dataSize, ACL_MEMCPY_DEVICE_TO_HOST)); + SaveMemoryToBinFile(variantPack.outTensors.at(i).hostData,variantPack.outTensors.at(i).dataSize,i); + } + + FreeTensors(variantPack.inTensors, variantPack.outTensors); + st = atb::DestroyContext(context); + CheckAcl(aclrtDestroyStream(stream)); + CheckAcl(aclrtResetDevice(0)); + CheckAcl(aclFinalize()); + return atb::ErrorType::NO_ERROR; +} diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.h b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.h new file mode 100644 index 000000000..44a94e9ec --- /dev/null +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.h @@ -0,0 +1,55 @@ +#include +#include +#include +#include +#include +#include +#include +#include "securec.h" +#include "atb/atb_infer.h" +#include "aclnn_add_operation.h" +#include + +struct InputData{ + void* data; + uint64_t size; +}; +aclError CheckAcl(aclError ret) +{ + if (ret != ACL_ERROR_NONE) { + std::cerr << __FILE__ << ":" << __LINE__ << " aclError:" << ret << std::endl; + } + return ret; +} +void* ReadBinFile(const char* filename, size_t& size) { + std::ifstream file(filename, std::ios::binary | std::ios::ate); + if (!file) { + std::cerr << "无法打开文件: " << filename << std::endl; + return nullptr; + } + + // 获取文件大小 + size = file.tellg(); + file.seekg(0, std::ios::beg); + + // 分配内存 + void* buffer; + int ret = aclrtMallocHost(&buffer,size); + if (!buffer) { + std::cerr << "内存分配失败" << std::endl; + file.close(); + return nullptr; + } + + // 读取文件内容到内存 + file.read(static_cast(buffer), size); + if (!file) { + std::cerr << "读取文件失败" << std::endl; + delete[] static_cast(buffer); + file.close(); + return nullptr; + } + + file.close(); + return buffer; +} \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md new file mode 100644 index 000000000..e0a38627c --- /dev/null +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md @@ -0,0 +1,70 @@ +## 概述 + +本样例基于AscendC自定义[Add](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)算子,开发了ATB插件并进行了插件调用测试. + +## 项目结构介绍 +``` + +├── AddAscendC //Add AscendC自定义算子工程 + +├── AddOperationATBPlugin //AddOperation ATB插件代码 + +├── AddOperationTest //AddOperation 测试代码 +``` +## 样例运行 + +### Add AscendC自定义算子部署 + +参照cann-ops仓[add_custom算子](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)" **算子包编译部署** "章节 + +### AddOperation ATB插件部署 + +- 运行编译脚本完成部署(脚本会生成静态库.a文件,同时将头文件拷贝到/usr/include,.a文件拷贝到/usr/local/lib下) + + ``` + cd AddOperationATBPlugin + bash build.sh + ``` + +### AddOperation测试 + +- 运行脚本完成算子测试 + + ```shell + cd AddOperationTest + bash run.sh + ``` + +## AddOperation算子介绍 + +### 功能 + +实现两个输入张量相加 + +### 定义 + +``` +struct AddParam { +}; +``` + +### 参数列表 + +该算子参数为空 + +### 输入 + +| **参数** | **维度** | **数据类型** | **格式** | 描述 | +| -------- | -------------------------- | --------------------- | -------- | ---------- | +| x | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | +| y | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | + +### 输出 + +| **参数** | **维度** | **数据类型** | **格式** | 描述 | +| -------- | -------------------------- | --------------------- | -------- | ---------------------------------------- | +| output | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输出tensor。数据类型和shape与x保持一致。 | + +### 规格约束 + +暂无 \ No newline at end of file -- Gitee From dad6832175a7f22e254cb5c23aa178fce2dbdf2d Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E5=88=98=E5=96=9C=E5=BC=BA?= Date: Tue, 1 Apr 2025 07:40:24 +0000 Subject: [PATCH 14/26] update --- .../24_add_custom/AddCustom/README.md | 3 ++ .../4_best_practices/24_add_custom/README.md | 49 +++++++++++++++++++ 2 files changed, 52 insertions(+) create mode 100644 operator/ascendc/4_best_practices/24_add_custom/AddCustom/README.md create mode 100644 operator/ascendc/4_best_practices/24_add_custom/README.md diff --git a/operator/ascendc/4_best_practices/24_add_custom/AddCustom/README.md b/operator/ascendc/4_best_practices/24_add_custom/AddCustom/README.md new file mode 100644 index 000000000..25c5ffc7e --- /dev/null +++ b/operator/ascendc/4_best_practices/24_add_custom/AddCustom/README.md @@ -0,0 +1,3 @@ +### 概述 + +具体算子工程,编译部署请见[AddCustom](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom) \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/24_add_custom/README.md b/operator/ascendc/4_best_practices/24_add_custom/README.md new file mode 100644 index 000000000..16780e8fa --- /dev/null +++ b/operator/ascendc/4_best_practices/24_add_custom/README.md @@ -0,0 +1,49 @@ +### 概述 + +本样例介基于AddCustom算子工程,介绍了单算子工程以及ATB插件调用方法 + +### 目录结构介绍 + +``` +├── 24_add_custom // 使用框架调用的方式调用AllGatherMatmul自定义算子 +│ ├── ATBInvocation // 通过ATB调用的方式调用 +│ ├── AddCustom // AddCustom算子工程,引用cann-ops仓的AddCustom工程 +``` + +### 算子描述 + +`AddCustom`算子返回两个数据相加的结果。 + +### 算子规格描述 + +| 算子类型(OpType) | AddCustom | | | | +| ---------------- | ---------- | -------- | --------- | ------ | +| 算子输入 | name | shape | data type | format | +| x | 8 * 2048 | float16 | ND | | +| y | 8 * 2048 | float16 | ND | | +| 算子输出 | z | 8 * 2048 | float16 | ND | +| 核函数名 | add_custom | | | | + +### 支持的产品型号 + +本样例支持如下产品型号: + +- Atlas 训练系列产品 +- Atlas 推理系列产品 +- Atlas A2训练系列产品 +- Atlas 800I A2推理产品 +- Atlas 200I/500 A2推理产品 + +### 算子编译部署 + +- [AddCustom编译部署(需要下载cann-ops仓)](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom) + +### 算子ATB插件调用 + +- [AddOperation编译调用](https://gitee.com/youmoxiao/samples/tree/atb_op/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation) + +## 更新说明 + +| 时间 | 更新事项 | +| -------- | -------- | +| 2025/4/1 | 新增样例 | \ No newline at end of file -- Gitee From 3dfa797272f0409f03295bba787d06456ff16729 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E5=88=98=E5=96=9C=E5=BC=BA?= Date: Tue, 1 Apr 2025 11:52:05 +0000 Subject: [PATCH 15/26] update eye --- .../EyeOperationATBPlugin/CMakeLists.txt | 21 ++ .../aclnn_eye_operation.cpp | 133 ++++++++++++ .../aclnn_eye_operation.h | 60 ++++++ .../EyeOperationATBPlugin/build.sh | 33 +++ .../EyeOperationTest/CMakeLists.txt | 40 ++++ .../EyeOperationTest/script/gen_data.py | 16 ++ .../EyeOperationTest/script/run.sh | 52 +++++ .../EyeOperationTest/script/verify_result.py | 25 +++ .../EyeOperationTest/src/main.cpp | 195 ++++++++++++++++++ .../ATBInvocation/EyeOperationTest/src/main.h | 55 +++++ .../25_Eye_custom/ATBInvocation/readme.md | 70 +++++++ .../25_Eye_custom/EyeCustom/README.md | 3 + .../4_best_practices/25_Eye_custom/README.md | 49 +++++ .../EyeOperationATBPlugin/CMakeLists.txt | 21 ++ .../aclnn_eye_operation.cpp | 133 ++++++++++++ .../aclnn_eye_operation.h | 60 ++++++ .../EyeOperationATBPlugin/build.sh | 33 +++ .../EyeOperationTest/CMakeLists.txt | 40 ++++ .../EyeOperationTest/script/gen_data.py | 16 ++ .../EyeOperationTest/script/run.sh | 52 +++++ .../EyeOperationTest/script/verify_result.py | 25 +++ .../EyeOperationTest/src/main.cpp | 195 ++++++++++++++++++ .../ATBInvocation/EyeOperationTest/src/main.h | 55 +++++ .../25_eye/ATBInvocation/readme.md | 70 +++++++ .../25_eye/EyeCustom/README.md | 3 + .../ascendc/4_best_practices/25_eye/README.md | 49 +++++ 26 files changed, 1504 insertions(+) create mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt create mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp create mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h create mode 100755 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/build.sh create mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/CMakeLists.txt create mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/gen_data.py create mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/run.sh create mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/verify_result.py create mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.cpp create mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.h create mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/readme.md create mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/EyeCustom/README.md create mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/README.md create mode 100644 operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt create mode 100644 operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp create mode 100644 operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h create mode 100755 operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/build.sh create mode 100644 operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/CMakeLists.txt create mode 100644 operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/gen_data.py create mode 100644 operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/run.sh create mode 100644 operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/verify_result.py create mode 100644 operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/src/main.cpp create mode 100644 operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/src/main.h create mode 100644 operator/ascendc/4_best_practices/25_eye/ATBInvocation/readme.md create mode 100644 operator/ascendc/4_best_practices/25_eye/EyeCustom/README.md create mode 100644 operator/ascendc/4_best_practices/25_eye/README.md diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt new file mode 100644 index 000000000..c89578fcd --- /dev/null +++ b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt @@ -0,0 +1,21 @@ +# Copyright 2024 Tencent Inc. All rights reserved. +# +# ============================================================================== +cmake_minimum_required(VERSION 3.14) +set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") +file(GLOB_RECURSE atb_plugin_operations_SRCS "./*.cpp") +set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") +message(STATUS "atb_plugin_operations_SRCS: ${atb_plugin_operations_SRCS}") +include_directories( + $ENV{ATB_HOME_PATH}/include + $ENV{ASCEND_HOME_PATH}/include + ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 + ${CMAKE_CURRENT_SOURCE_DIR} + ) +link_directories( + $ENV{ATB_HOME_PATH}/lib + $ENV{ASCEND_HOME_PATH}/lib64 + ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 + ) +add_library(atb_plugin_operations STATIC ${atb_plugin_operations_SRCS}) +target_link_libraries(atb_plugin_operations PRIVATE ${ACL_SHARED_LIBS} ${ACL_SHARED_LIBS} ascendcl) \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp new file mode 100644 index 000000000..a4f156c4b --- /dev/null +++ b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp @@ -0,0 +1,133 @@ +#include "aclnn_eye_operation.h" +#include "aclnn_eye.h" + +EyeOperation::EyeOperation(const std::string &name, EyeAttrParam param){ + attrParam = param; + opName_ = name; +} + +atb::SVector GetCopyTensorStride(atb::Dims &tensorDims) +{ + atb::SVector tmpStrides(tensorDims.dimNum, 1); + if (tensorDims.dimNum > 8) { // 8: tensor最大维度数量 + printf("tensor's dimNum is larger than 8, GetCopyTensorStride failed."); + return tmpStrides; + } + for (int64_t i = static_cast(tensorDims.dimNum) - 2; i >= 0; i--) { + tmpStrides[i] = (tensorDims.dims[i + 1] * tmpStrides[i + 1]); + } + return tmpStrides; +} + +std::shared_ptr EyeOperation::CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx) +{ + auto aclnnTensor = std::make_shared(); + aclnnTensor->tensorIdx = static_cast(tensorIdx); + aclnnTensor->needUpdateTensorDataPtr = true; + aclnnTensor->atbTensor = atbTensor; + aclnnTensor->strides = GetCopyTensorStride(atbTensor.desc.shape); + + // 创建Aclnn tensor + aclnnTensor->tensor = aclCreateTensor(atbTensor.desc.shape.dims, + atbTensor.desc.shape.dimNum, + atbTensor.desc.dtype, + aclnnTensor->strides.data(), + 0, + atbTensor.desc.format, + atbTensor.desc.shape.dims, + atbTensor.desc.shape.dimNum, + atbTensor.deviceData); + return aclnnTensor; +} + +atb::Status EyeOperation::UpdateAclnnVariantPack(const atb::VariantPack &variantPack) +{ + // 更新inTensor的device地址 + for (size_t i = 0; i < aclInTensors_.size(); ++i) { + int ret = -1; + if (!aclInTensors_[i]->needUpdateTensorDataPtr) { + continue; + } + aclInTensors_[i]->atbTensor = variantPack.inTensors.at(i); + ret = aclSetInputTensorAddr(aclExecutor_, + aclInTensors_[i]->tensorIdx, + aclInTensors_[i]->tensor, + aclInTensors_[i]->atbTensor.deviceData); + + if (ret != 0) { + printf("set input fail"); + return atb::ERROR_CANN_ERROR; + } + } + + // 更新outTensor的device地址 + for (size_t i = 0; i < aclOutTensors_.size(); ++i) { + int ret = -1; + if (!aclOutTensors_[i]->needUpdateTensorDataPtr) { + continue; + } + aclOutTensors_[i]->atbTensor = variantPack.outTensors.at(i); + ret = aclSetOutputTensorAddr(aclExecutor_, + aclOutTensors_[i]->tensorIdx, + aclOutTensors_[i]->tensor, + aclOutTensors_[i]->atbTensor.deviceData); + + if (ret != 0) { + printf("set output fail"); + return atb::ERROR_CANN_ERROR; + } + } + return atb::NO_ERROR; +} + +atb::Status EyeOperation::Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) { + + aclInTensors_.resize(GetInputNum()); + for (size_t i = 0; i < aclInTensors_.size(); ++i) { + auto aclnnTensor = CreateAclnnTensor(variantPack.inTensors.at(i), i); + if (aclnnTensor->tensor == nullptr) { + printf("creat input tensor %ld fail", i); + return atb::ERROR_INTERNAL_ERROR; + } + aclInTensors_[i] = aclnnTensor; + } + + auto ret = aclnnEyeGetWorkspaceSize(aclInTensors_.at(0)->tensor, + attrParam.num_rows, + attrParam.num_columns, + attrParam.batch_shape, + attrParam.dtype, + &workspaceSize_, + &aclExecutor_); + + workspaceSize = workspaceSize_; + return ret; + +} + +atb::Status EyeOperation::Execute(const atb::VariantPack &variantPack, uint8_t *workspace, uint64_t workspaceSize, atb::Context *context) { + + + + aclrtStream stream = context->GetExecuteStream(); + if (!stream) { + printf("get stream fail"); + return atb::ERROR_INVALID_PARAM; + } + // 更新数据传入的地址 + int ret = UpdateAclnnVariantPack(variantPack); + if (ret != 0) { + printf("UpdateAclnnVariantPack fail"); + return atb::ERROR_CANN_ERROR; + } + ret = aclnnEye(workspace, workspaceSize_, aclExecutor_, stream); + + return ret; +} + +atb::Status EyeOperation::InferShape( + const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const +{ + outTensorDesc.at(0) = inTensorDesc.at(0); + return atb::NO_ERROR; +} diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h new file mode 100644 index 000000000..b09697135 --- /dev/null +++ b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h @@ -0,0 +1,60 @@ +#include +#include +#include +#include +#include +#include "atb/infer_op_params.h" +#include + +struct EyeAttrParam +{ + uint64_t num_rows; + uint64_t num_columns = 0; + std::vector batchShape = {1}; + aclIntArray* batch_shape = aclCreateIntArray(batchShape.data(),batchShape.size()); + uint64_t dtype = 0; +}; + +struct AclnnTensor +{ +public: + atb::Tensor atbTensor; // + aclTensor *tensor = nullptr; + int tensorIdx = -1; // aclTensor在aclExecutor中的index + bool needUpdateTensorDataPtr = false; + atb::SVector strides = {}; +}; + +class EyeOperation: public atb::Operation{ +public: + EyeOperation(const std::string &name, EyeAttrParam param); + atb::Status Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) override; + atb::Status Execute(const atb::VariantPack &variantPack, uint8_t *workspace, + uint64_t workspaceSize, atb::Context *context) override; + atb::Status InferShape( + const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const; + std::shared_ptr CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx); + atb::Status UpdateAclnnVariantPack(const atb::VariantPack &variantPack); + + uint32_t GetInputNum() const + { + return 1; // 算子入参个数 + } + + uint32_t GetOutputNum() const + { + return 1; // 算子出参个数 + } + std::string GetName() const + { + return opName_; + } + + aclOpExecutor *aclExecutor_ = nullptr; + EyeAttrParam attrParam; + std::string opName_; + uint64_t workspaceSize_; + + atb::SVector> aclInTensors_; + atb::SVector> aclOutTensors_; +}; \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/build.sh b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/build.sh new file mode 100755 index 000000000..ba28cc143 --- /dev/null +++ b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/build.sh @@ -0,0 +1,33 @@ +#!/bin/bash + +# 定义构建目录 +BUILD_DIR="build" + +# 创建构建目录 +mkdir -p "$BUILD_DIR" +cd "$BUILD_DIR" + +# 运行 CMake 配置和编译 +cmake .. +make + +# 查找生成的 .a 文件 +A_FILE=$(find . -name "*.a" -type f) + +# 检查是否找到了 .a 文件 +if [ -z "$A_FILE" ]; then + echo "未找到 .a 文件,编译可能失败。" + exit 1 +fi + +# 复制头文件到 /usr/include +HEADER_FILES=$(find .. -name "*.h" -type f) +for header in $HEADER_FILES; do + cp "$header" /usr/include/ +done + +# 复制 .a 文件到 /usr/local/lib +cp "$A_FILE" /usr/local/lib/ + +echo "构建完成,头文件和 .a 文件已复制到目标目录。" + diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/CMakeLists.txt b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/CMakeLists.txt new file mode 100644 index 000000000..b9c2cec46 --- /dev/null +++ b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/CMakeLists.txt @@ -0,0 +1,40 @@ +cmake_minimum_required(VERSION 3.16) +project("test_model") +option(USE_CXX11_ABI "USE_CXX11_ABI" OFF) +set(CMAKE_BUILD_TYPE Debug) +if(USE_CXX11_ABI) + set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=1") +else() + set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") +endif() + +set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -std=c++17") + +set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") + +include_directories( + $ENV{ATB_HOME_PATH}/include + $ENV{ASCEND_HOME_PATH}/include + ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 + ${CMAKE_CURRENT_SOURCE_DIR} + ./src/ + ) + +message(".h path = ${CUST_PKG_PATH}/include") +link_directories( + $ENV{ATB_HOME_PATH}/lib + $ENV{ASCEND_HOME_PATH}/lib64 + ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 + ) + + +link_libraries(atb_plugin_operations) +add_executable(test_model ./src/main.cpp) + +# 列出所有的头文件目录 +target_link_libraries(test_model PRIVATE atb ascendcl cust_opapi opapi nnopbase pthread) # 添加自定义算子so及适配so + + + + + diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/gen_data.py b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/gen_data.py new file mode 100644 index 000000000..b0a8ba36f --- /dev/null +++ b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/gen_data.py @@ -0,0 +1,16 @@ +import torch +import os +def gen_golden_data_simple(): + dtype = "float32" + input1 = torch.zeros(3 , 4, 133, 4095, dtype=torch.float) + golden = torch.eye(133,4095, dtype=torch.float) + golden = golden.unsqueeze(0).unsqueeze(0) + golden = golden.repeat(3,4,1,1) + input1.numpy().tofile('./script/input/input0.bin') + golden.numpy().tofile("./script/output/golden0.bin") + + with open("./script/output/meta", "w") as fp: + print(dtype, file=fp) + +if __name__ == "__main__": + gen_golden_data_simple() diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/run.sh b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/run.sh new file mode 100644 index 000000000..450b717f2 --- /dev/null +++ b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/run.sh @@ -0,0 +1,52 @@ +#!/bin/bash + + +# 步骤1: 运行gen_data.py生成输入bin文件和golden标杆输出数据 +echo "正在生成输入数据和golden标杆数据..." +mkdir -p script/input +mkdir -p script/output +python3 script/gen_data.py +if [ $? -ne 0 ]; then + echo "生成数据失败,脚本终止。" + exit 1 +fi + +# 步骤2: 创建构建目录并进入 +mkdir -p build +cd build +if [ $? -ne 0 ]; then + echo "无法进入构建目录,脚本终止。" + exit 1 +fi + +# 步骤3: 使用CMake配置项目 +echo "正在配置CMake项目..." +cmake .. +if [ $? -ne 0 ]; then + echo "CMake配置失败,脚本终止。" + exit 1 +fi + +# 步骤4: 编译代码 +echo "正在编译代码..." +make +if [ $? -ne 0 ]; then + echo "编译失败,脚本终止。" + exit 1 +fi + +mv test_model ../ +cd .. + +# 步骤5: 运行可执行文件生成实际输出文件 +echo "正在运行可执行文件生成实际输出..." +./test_model +if [ $? -ne 0 ]; then + echo "运行可执行文件失败,脚本终止。" + exit 1 +fi + +# 步骤6: 调用verify_result.py进行golden标杆数据和实际输出数据的比对 +echo "正在验证结果..." +python3 script/verify_result.py script/output/output_0.bin script/output/golden0.bin + diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/verify_result.py b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/verify_result.py new file mode 100644 index 000000000..a0798d50a --- /dev/null +++ b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/verify_result.py @@ -0,0 +1,25 @@ +import os +import sys +import numpy + +loss = 1e-3 # 容忍偏差,一般fp16要求绝对误差和相对误差均不超过千分之一 +minimum = 10e-10 + +def verify_result(real_result, golden): + real_result = numpy.fromfile(real_result, dtype=numpy.float32) # 从bin文件读取实际运算结果 + golden = numpy.fromfile(golden, dtype=numpy.float32) # 从bin文件读取预期运算结果 + print("=" * 50, real_result[:5], golden[:5], "=" * 50, sep='\n', end='\n', file=sys.stderr) + result = numpy.abs(real_result - golden) # 计算运算结果和预期结果偏差 + deno = numpy.maximum(numpy.abs(real_result), numpy.abs(golden)) # 获取最大值并组成新数组 + result_atol = numpy.less_equal(result, loss) # 计算绝对误差 + result_rtol = numpy.less_equal(result / numpy.add(deno, minimum), loss) # 计算相对误差 + if not result_rtol.all() and not result_atol.all(): + if numpy.sum(result_rtol == False) > real_result.size * loss and numpy.sum(result_atol == False) > real_result.size * loss: # 误差超出预期时返回打印错误,返回对比失败 + print("[ERROR] result error") + return False + print("test Operation success!") + return True + + +if __name__ == '__main__': + verify_result(sys.argv[1],sys.argv[2]) diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.cpp b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.cpp new file mode 100644 index 000000000..0bd444165 --- /dev/null +++ b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.cpp @@ -0,0 +1,195 @@ +/* + * Copyright (c) Huawei Technologies Co., Ltd. 2023. All rights reserved. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +#include "main.h" + +bool SetInputData(std::vector &inputData){ + char *xPath = "./script/input/input0.bin"; + InputData inputX; + inputX.data = ReadBinFile(xPath,inputX.size); + inputData.push_back(inputX); + return true; +} + +bool SetOperationInputDesc(atb::SVector &intensorDescs){ + atb::TensorDesc xDesc; + xDesc.dtype = ACL_FLOAT; + xDesc.format = ACL_FORMAT_ND; + xDesc.shape.dimNum = 4; + xDesc.shape.dims[0] = 3; + xDesc.shape.dims[1] = 4; + xDesc.shape.dims[2] = 133; + xDesc.shape.dims[3] = 4095; + intensorDescs.at(0) = xDesc; +} + + + +static void SetCurrentDevice() +{ + const int deviceId = 0; + std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; + int ret = aclrtSetDevice(deviceId); + if (ret != 0) { + std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; + return; + } + std::cout << "[INFO]: aclrtSetDevice success" << std::endl; +} + + +static void FreeTensor(atb::Tensor &tensor) +{ + if (tensor.deviceData) { + int ret = aclrtFree(tensor.deviceData); + if (ret != 0) { + std::cout << "[ERROR]: aclrtFree fail" << std::endl; + } + tensor.deviceData = nullptr; + tensor.dataSize = 0; + } + if (tensor.hostData) { + int ret = aclrtFreeHost(tensor.hostData); + if (ret != 0) { + std::cout << "[ERROR]: aclrtFreeHost fail, ret = " << ret << std::endl; + } + tensor.hostData = nullptr; + tensor.dataSize = 0; + } +} + +static void FreeTensors(atb::SVector &inTensors) +{ + for (size_t i = 0; i < inTensors.size(); ++i) { + FreeTensor(inTensors.at(i)); + } +} +bool SaveMemoryToBinFile(void* memoryAddress, size_t memorySize, size_t i) { + // 创建 output 目录(如果不存在) + std::filesystem::create_directories("output"); + + // 生成文件名 + std::string filename = "script/output/output_" + std::to_string(i) + ".bin"; + + // 打开文件以二进制写入模式 + std::ofstream file(filename, std::ios::binary); + if (!file) { + std::cerr << "无法打开文件: " << filename << std::endl; + return false; + } + + // 写入数据 + file.write(static_cast(memoryAddress), memorySize); + if (!file) { + std::cerr << "写入文件时出错: " << filename << std::endl; + file.close(); + return false; + } + + // 关闭文件 + file.close(); + std::cout << "数据已成功保存到: " << filename << std::endl; + return true; +} + +int main(int argc, const char *argv[]) +{ + const int deviceId = 0; + std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; + int ret = aclrtSetDevice(deviceId); + if (ret != 0) { + std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; + return 1; + } + std::cout << "[INFO]: aclrtSetDevice success" << std::endl; + atb::Context *context = nullptr; + ret = atb::CreateContext(&context); + void *stream = nullptr; + ret = aclrtCreateStream(&stream); + if (ret != 0) { + std::cout << "[ERROR]: AsdRtStreamCreate fail, ret:" << ret << std::endl; + return 1; + } + context->SetExecuteStream(stream); + + std::vector input; + SetInputData(input); + + EyeAttrParam eyeAttrParam; + eyeAttrParam.num_rows = 133; + eyeAttrParam.num_columns = 4095; + + std::vector batchShape = {3,4}; + eyeAttrParam.batch_shape = aclCreateIntArray(batchShape.data(),batchShape.size()); + eyeAttrParam.dtype = 0; + EyeOperation *op = new EyeOperation("Eye",eyeAttrParam); + std::cout << "[INFO]: complete CreateOp!" << std::endl; + + if(input.size() != op->GetInputNum()) std::cout << "[ERROR]: Operation actual input num is not equal to GetInputNum()"; + + atb::SVector intensorDescs; + atb::SVector outtensorDescs; + intensorDescs.resize(op->GetInputNum()); + outtensorDescs.resize(op->GetOutputNum()); + SetOperationInputDesc(intensorDescs); + atb::Status st = op->InferShape(intensorDescs,outtensorDescs); + if (st != 0) { + std::cout << "[ERROR]: Operation InferShape fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation InferShape success" << std::endl; + atb::VariantPack variantPack; + variantPack.inTensors.resize(op->GetInputNum()); + for(size_t i=0;iGetInputNum();i++){ + variantPack.inTensors.at(i).desc = intensorDescs.at(i); + variantPack.inTensors.at(i).hostData = input[i].data; + variantPack.inTensors.at(i).dataSize = input[i].size; + CheckAcl(aclrtMalloc(&variantPack.inTensors.at(i).deviceData, input[i].size, ACL_MEM_MALLOC_HUGE_FIRST)); + CheckAcl(aclrtMemcpy(variantPack.inTensors.at(i).deviceData, input[i].size, input[i].data, input[i].size, ACL_MEMCPY_HOST_TO_DEVICE)); + } + std::cout << "[INFO]: Operation Input prepare sucess" << std::endl; + + uint64_t workspaceSize = 0; + st = op->Setup(variantPack, workspaceSize, context); + if (st != 0) { + std::cout << "[ERROR]: Operation setup fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation setup success" << std::endl; + void *workspace = nullptr; + if (workspaceSize > 0) { + aclrtMalloc(&workspace, workspaceSize, ACL_MEM_MALLOC_HUGE_FIRST); + } + + std::cout << "[INFO]: Operation execute start" << std::endl; + st = op->Execute(variantPack, (uint8_t*)workspace, workspaceSize, context); + if (st != 0) { + std::cout << "[ERROR]: Operation execute fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation execute success" << std::endl; + + CheckAcl(aclrtMemcpy(variantPack.inTensors.at(0).hostData, variantPack.inTensors.at(0).dataSize, variantPack.inTensors.at(0).deviceData, + variantPack.inTensors.at(0).dataSize, ACL_MEMCPY_DEVICE_TO_HOST)); + SaveMemoryToBinFile(variantPack.inTensors.at(0).hostData,variantPack.inTensors.at(0).dataSize,0); + + + FreeTensors(variantPack.inTensors); + st = atb::DestroyContext(context); + CheckAcl(aclrtDestroyStream(stream)); + CheckAcl(aclrtResetDevice(0)); + CheckAcl(aclFinalize()); + return atb::ErrorType::NO_ERROR; +} diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.h b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.h new file mode 100644 index 000000000..015c779c3 --- /dev/null +++ b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.h @@ -0,0 +1,55 @@ +#include +#include +#include +#include +#include +#include +#include +#include "securec.h" +#include "atb/atb_infer.h" +#include "aclnn_eye_operation.h" +#include + +struct InputData{ + void* data; + uint64_t size; +}; +aclError CheckAcl(aclError ret) +{ + if (ret != ACL_ERROR_NONE) { + std::cerr << __FILE__ << ":" << __LINE__ << " aclError:" << ret << std::endl; + } + return ret; +} +void* ReadBinFile(const char* filename, size_t& size) { + std::ifstream file(filename, std::ios::binary | std::ios::ate); + if (!file) { + std::cerr << "无法打开文件: " << filename << std::endl; + return nullptr; + } + + // 获取文件大小 + size = file.tellg(); + file.seekg(0, std::ios::beg); + + // 分配内存 + void* buffer; + int ret = aclrtMallocHost(&buffer,size); + if (!buffer) { + std::cerr << "内存分配失败" << std::endl; + file.close(); + return nullptr; + } + + // 读取文件内容到内存 + file.read(static_cast(buffer), size); + if (!file) { + std::cerr << "读取文件失败" << std::endl; + delete[] static_cast(buffer); + file.close(); + return nullptr; + } + + file.close(); + return buffer; +} \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/readme.md b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/readme.md new file mode 100644 index 000000000..e0a38627c --- /dev/null +++ b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/readme.md @@ -0,0 +1,70 @@ +## 概述 + +本样例基于AscendC自定义[Add](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)算子,开发了ATB插件并进行了插件调用测试. + +## 项目结构介绍 +``` + +├── AddAscendC //Add AscendC自定义算子工程 + +├── AddOperationATBPlugin //AddOperation ATB插件代码 + +├── AddOperationTest //AddOperation 测试代码 +``` +## 样例运行 + +### Add AscendC自定义算子部署 + +参照cann-ops仓[add_custom算子](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)" **算子包编译部署** "章节 + +### AddOperation ATB插件部署 + +- 运行编译脚本完成部署(脚本会生成静态库.a文件,同时将头文件拷贝到/usr/include,.a文件拷贝到/usr/local/lib下) + + ``` + cd AddOperationATBPlugin + bash build.sh + ``` + +### AddOperation测试 + +- 运行脚本完成算子测试 + + ```shell + cd AddOperationTest + bash run.sh + ``` + +## AddOperation算子介绍 + +### 功能 + +实现两个输入张量相加 + +### 定义 + +``` +struct AddParam { +}; +``` + +### 参数列表 + +该算子参数为空 + +### 输入 + +| **参数** | **维度** | **数据类型** | **格式** | 描述 | +| -------- | -------------------------- | --------------------- | -------- | ---------- | +| x | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | +| y | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | + +### 输出 + +| **参数** | **维度** | **数据类型** | **格式** | 描述 | +| -------- | -------------------------- | --------------------- | -------- | ---------------------------------------- | +| output | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输出tensor。数据类型和shape与x保持一致。 | + +### 规格约束 + +暂无 \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/EyeCustom/README.md b/operator/ascendc/4_best_practices/25_Eye_custom/EyeCustom/README.md new file mode 100644 index 000000000..25c5ffc7e --- /dev/null +++ b/operator/ascendc/4_best_practices/25_Eye_custom/EyeCustom/README.md @@ -0,0 +1,3 @@ +### 概述 + +具体算子工程,编译部署请见[AddCustom](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom) \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/README.md b/operator/ascendc/4_best_practices/25_Eye_custom/README.md new file mode 100644 index 000000000..16780e8fa --- /dev/null +++ b/operator/ascendc/4_best_practices/25_Eye_custom/README.md @@ -0,0 +1,49 @@ +### 概述 + +本样例介基于AddCustom算子工程,介绍了单算子工程以及ATB插件调用方法 + +### 目录结构介绍 + +``` +├── 24_add_custom // 使用框架调用的方式调用AllGatherMatmul自定义算子 +│ ├── ATBInvocation // 通过ATB调用的方式调用 +│ ├── AddCustom // AddCustom算子工程,引用cann-ops仓的AddCustom工程 +``` + +### 算子描述 + +`AddCustom`算子返回两个数据相加的结果。 + +### 算子规格描述 + +| 算子类型(OpType) | AddCustom | | | | +| ---------------- | ---------- | -------- | --------- | ------ | +| 算子输入 | name | shape | data type | format | +| x | 8 * 2048 | float16 | ND | | +| y | 8 * 2048 | float16 | ND | | +| 算子输出 | z | 8 * 2048 | float16 | ND | +| 核函数名 | add_custom | | | | + +### 支持的产品型号 + +本样例支持如下产品型号: + +- Atlas 训练系列产品 +- Atlas 推理系列产品 +- Atlas A2训练系列产品 +- Atlas 800I A2推理产品 +- Atlas 200I/500 A2推理产品 + +### 算子编译部署 + +- [AddCustom编译部署(需要下载cann-ops仓)](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom) + +### 算子ATB插件调用 + +- [AddOperation编译调用](https://gitee.com/youmoxiao/samples/tree/atb_op/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation) + +## 更新说明 + +| 时间 | 更新事项 | +| -------- | -------- | +| 2025/4/1 | 新增样例 | \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt new file mode 100644 index 000000000..c89578fcd --- /dev/null +++ b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt @@ -0,0 +1,21 @@ +# Copyright 2024 Tencent Inc. All rights reserved. +# +# ============================================================================== +cmake_minimum_required(VERSION 3.14) +set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") +file(GLOB_RECURSE atb_plugin_operations_SRCS "./*.cpp") +set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") +message(STATUS "atb_plugin_operations_SRCS: ${atb_plugin_operations_SRCS}") +include_directories( + $ENV{ATB_HOME_PATH}/include + $ENV{ASCEND_HOME_PATH}/include + ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 + ${CMAKE_CURRENT_SOURCE_DIR} + ) +link_directories( + $ENV{ATB_HOME_PATH}/lib + $ENV{ASCEND_HOME_PATH}/lib64 + ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 + ) +add_library(atb_plugin_operations STATIC ${atb_plugin_operations_SRCS}) +target_link_libraries(atb_plugin_operations PRIVATE ${ACL_SHARED_LIBS} ${ACL_SHARED_LIBS} ascendcl) \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp new file mode 100644 index 000000000..a4f156c4b --- /dev/null +++ b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp @@ -0,0 +1,133 @@ +#include "aclnn_eye_operation.h" +#include "aclnn_eye.h" + +EyeOperation::EyeOperation(const std::string &name, EyeAttrParam param){ + attrParam = param; + opName_ = name; +} + +atb::SVector GetCopyTensorStride(atb::Dims &tensorDims) +{ + atb::SVector tmpStrides(tensorDims.dimNum, 1); + if (tensorDims.dimNum > 8) { // 8: tensor最大维度数量 + printf("tensor's dimNum is larger than 8, GetCopyTensorStride failed."); + return tmpStrides; + } + for (int64_t i = static_cast(tensorDims.dimNum) - 2; i >= 0; i--) { + tmpStrides[i] = (tensorDims.dims[i + 1] * tmpStrides[i + 1]); + } + return tmpStrides; +} + +std::shared_ptr EyeOperation::CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx) +{ + auto aclnnTensor = std::make_shared(); + aclnnTensor->tensorIdx = static_cast(tensorIdx); + aclnnTensor->needUpdateTensorDataPtr = true; + aclnnTensor->atbTensor = atbTensor; + aclnnTensor->strides = GetCopyTensorStride(atbTensor.desc.shape); + + // 创建Aclnn tensor + aclnnTensor->tensor = aclCreateTensor(atbTensor.desc.shape.dims, + atbTensor.desc.shape.dimNum, + atbTensor.desc.dtype, + aclnnTensor->strides.data(), + 0, + atbTensor.desc.format, + atbTensor.desc.shape.dims, + atbTensor.desc.shape.dimNum, + atbTensor.deviceData); + return aclnnTensor; +} + +atb::Status EyeOperation::UpdateAclnnVariantPack(const atb::VariantPack &variantPack) +{ + // 更新inTensor的device地址 + for (size_t i = 0; i < aclInTensors_.size(); ++i) { + int ret = -1; + if (!aclInTensors_[i]->needUpdateTensorDataPtr) { + continue; + } + aclInTensors_[i]->atbTensor = variantPack.inTensors.at(i); + ret = aclSetInputTensorAddr(aclExecutor_, + aclInTensors_[i]->tensorIdx, + aclInTensors_[i]->tensor, + aclInTensors_[i]->atbTensor.deviceData); + + if (ret != 0) { + printf("set input fail"); + return atb::ERROR_CANN_ERROR; + } + } + + // 更新outTensor的device地址 + for (size_t i = 0; i < aclOutTensors_.size(); ++i) { + int ret = -1; + if (!aclOutTensors_[i]->needUpdateTensorDataPtr) { + continue; + } + aclOutTensors_[i]->atbTensor = variantPack.outTensors.at(i); + ret = aclSetOutputTensorAddr(aclExecutor_, + aclOutTensors_[i]->tensorIdx, + aclOutTensors_[i]->tensor, + aclOutTensors_[i]->atbTensor.deviceData); + + if (ret != 0) { + printf("set output fail"); + return atb::ERROR_CANN_ERROR; + } + } + return atb::NO_ERROR; +} + +atb::Status EyeOperation::Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) { + + aclInTensors_.resize(GetInputNum()); + for (size_t i = 0; i < aclInTensors_.size(); ++i) { + auto aclnnTensor = CreateAclnnTensor(variantPack.inTensors.at(i), i); + if (aclnnTensor->tensor == nullptr) { + printf("creat input tensor %ld fail", i); + return atb::ERROR_INTERNAL_ERROR; + } + aclInTensors_[i] = aclnnTensor; + } + + auto ret = aclnnEyeGetWorkspaceSize(aclInTensors_.at(0)->tensor, + attrParam.num_rows, + attrParam.num_columns, + attrParam.batch_shape, + attrParam.dtype, + &workspaceSize_, + &aclExecutor_); + + workspaceSize = workspaceSize_; + return ret; + +} + +atb::Status EyeOperation::Execute(const atb::VariantPack &variantPack, uint8_t *workspace, uint64_t workspaceSize, atb::Context *context) { + + + + aclrtStream stream = context->GetExecuteStream(); + if (!stream) { + printf("get stream fail"); + return atb::ERROR_INVALID_PARAM; + } + // 更新数据传入的地址 + int ret = UpdateAclnnVariantPack(variantPack); + if (ret != 0) { + printf("UpdateAclnnVariantPack fail"); + return atb::ERROR_CANN_ERROR; + } + ret = aclnnEye(workspace, workspaceSize_, aclExecutor_, stream); + + return ret; +} + +atb::Status EyeOperation::InferShape( + const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const +{ + outTensorDesc.at(0) = inTensorDesc.at(0); + return atb::NO_ERROR; +} diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h new file mode 100644 index 000000000..b09697135 --- /dev/null +++ b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h @@ -0,0 +1,60 @@ +#include +#include +#include +#include +#include +#include "atb/infer_op_params.h" +#include + +struct EyeAttrParam +{ + uint64_t num_rows; + uint64_t num_columns = 0; + std::vector batchShape = {1}; + aclIntArray* batch_shape = aclCreateIntArray(batchShape.data(),batchShape.size()); + uint64_t dtype = 0; +}; + +struct AclnnTensor +{ +public: + atb::Tensor atbTensor; // + aclTensor *tensor = nullptr; + int tensorIdx = -1; // aclTensor在aclExecutor中的index + bool needUpdateTensorDataPtr = false; + atb::SVector strides = {}; +}; + +class EyeOperation: public atb::Operation{ +public: + EyeOperation(const std::string &name, EyeAttrParam param); + atb::Status Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) override; + atb::Status Execute(const atb::VariantPack &variantPack, uint8_t *workspace, + uint64_t workspaceSize, atb::Context *context) override; + atb::Status InferShape( + const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const; + std::shared_ptr CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx); + atb::Status UpdateAclnnVariantPack(const atb::VariantPack &variantPack); + + uint32_t GetInputNum() const + { + return 1; // 算子入参个数 + } + + uint32_t GetOutputNum() const + { + return 1; // 算子出参个数 + } + std::string GetName() const + { + return opName_; + } + + aclOpExecutor *aclExecutor_ = nullptr; + EyeAttrParam attrParam; + std::string opName_; + uint64_t workspaceSize_; + + atb::SVector> aclInTensors_; + atb::SVector> aclOutTensors_; +}; \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/build.sh b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/build.sh new file mode 100755 index 000000000..ba28cc143 --- /dev/null +++ b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/build.sh @@ -0,0 +1,33 @@ +#!/bin/bash + +# 定义构建目录 +BUILD_DIR="build" + +# 创建构建目录 +mkdir -p "$BUILD_DIR" +cd "$BUILD_DIR" + +# 运行 CMake 配置和编译 +cmake .. +make + +# 查找生成的 .a 文件 +A_FILE=$(find . -name "*.a" -type f) + +# 检查是否找到了 .a 文件 +if [ -z "$A_FILE" ]; then + echo "未找到 .a 文件,编译可能失败。" + exit 1 +fi + +# 复制头文件到 /usr/include +HEADER_FILES=$(find .. -name "*.h" -type f) +for header in $HEADER_FILES; do + cp "$header" /usr/include/ +done + +# 复制 .a 文件到 /usr/local/lib +cp "$A_FILE" /usr/local/lib/ + +echo "构建完成,头文件和 .a 文件已复制到目标目录。" + diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/CMakeLists.txt b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/CMakeLists.txt new file mode 100644 index 000000000..b9c2cec46 --- /dev/null +++ b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/CMakeLists.txt @@ -0,0 +1,40 @@ +cmake_minimum_required(VERSION 3.16) +project("test_model") +option(USE_CXX11_ABI "USE_CXX11_ABI" OFF) +set(CMAKE_BUILD_TYPE Debug) +if(USE_CXX11_ABI) + set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=1") +else() + set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") +endif() + +set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -std=c++17") + +set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") + +include_directories( + $ENV{ATB_HOME_PATH}/include + $ENV{ASCEND_HOME_PATH}/include + ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 + ${CMAKE_CURRENT_SOURCE_DIR} + ./src/ + ) + +message(".h path = ${CUST_PKG_PATH}/include") +link_directories( + $ENV{ATB_HOME_PATH}/lib + $ENV{ASCEND_HOME_PATH}/lib64 + ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 + ) + + +link_libraries(atb_plugin_operations) +add_executable(test_model ./src/main.cpp) + +# 列出所有的头文件目录 +target_link_libraries(test_model PRIVATE atb ascendcl cust_opapi opapi nnopbase pthread) # 添加自定义算子so及适配so + + + + + diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/gen_data.py b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/gen_data.py new file mode 100644 index 000000000..b0a8ba36f --- /dev/null +++ b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/gen_data.py @@ -0,0 +1,16 @@ +import torch +import os +def gen_golden_data_simple(): + dtype = "float32" + input1 = torch.zeros(3 , 4, 133, 4095, dtype=torch.float) + golden = torch.eye(133,4095, dtype=torch.float) + golden = golden.unsqueeze(0).unsqueeze(0) + golden = golden.repeat(3,4,1,1) + input1.numpy().tofile('./script/input/input0.bin') + golden.numpy().tofile("./script/output/golden0.bin") + + with open("./script/output/meta", "w") as fp: + print(dtype, file=fp) + +if __name__ == "__main__": + gen_golden_data_simple() diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/run.sh b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/run.sh new file mode 100644 index 000000000..450b717f2 --- /dev/null +++ b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/run.sh @@ -0,0 +1,52 @@ +#!/bin/bash + + +# 步骤1: 运行gen_data.py生成输入bin文件和golden标杆输出数据 +echo "正在生成输入数据和golden标杆数据..." +mkdir -p script/input +mkdir -p script/output +python3 script/gen_data.py +if [ $? -ne 0 ]; then + echo "生成数据失败,脚本终止。" + exit 1 +fi + +# 步骤2: 创建构建目录并进入 +mkdir -p build +cd build +if [ $? -ne 0 ]; then + echo "无法进入构建目录,脚本终止。" + exit 1 +fi + +# 步骤3: 使用CMake配置项目 +echo "正在配置CMake项目..." +cmake .. +if [ $? -ne 0 ]; then + echo "CMake配置失败,脚本终止。" + exit 1 +fi + +# 步骤4: 编译代码 +echo "正在编译代码..." +make +if [ $? -ne 0 ]; then + echo "编译失败,脚本终止。" + exit 1 +fi + +mv test_model ../ +cd .. + +# 步骤5: 运行可执行文件生成实际输出文件 +echo "正在运行可执行文件生成实际输出..." +./test_model +if [ $? -ne 0 ]; then + echo "运行可执行文件失败,脚本终止。" + exit 1 +fi + +# 步骤6: 调用verify_result.py进行golden标杆数据和实际输出数据的比对 +echo "正在验证结果..." +python3 script/verify_result.py script/output/output_0.bin script/output/golden0.bin + diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/verify_result.py b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/verify_result.py new file mode 100644 index 000000000..a0798d50a --- /dev/null +++ b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/verify_result.py @@ -0,0 +1,25 @@ +import os +import sys +import numpy + +loss = 1e-3 # 容忍偏差,一般fp16要求绝对误差和相对误差均不超过千分之一 +minimum = 10e-10 + +def verify_result(real_result, golden): + real_result = numpy.fromfile(real_result, dtype=numpy.float32) # 从bin文件读取实际运算结果 + golden = numpy.fromfile(golden, dtype=numpy.float32) # 从bin文件读取预期运算结果 + print("=" * 50, real_result[:5], golden[:5], "=" * 50, sep='\n', end='\n', file=sys.stderr) + result = numpy.abs(real_result - golden) # 计算运算结果和预期结果偏差 + deno = numpy.maximum(numpy.abs(real_result), numpy.abs(golden)) # 获取最大值并组成新数组 + result_atol = numpy.less_equal(result, loss) # 计算绝对误差 + result_rtol = numpy.less_equal(result / numpy.add(deno, minimum), loss) # 计算相对误差 + if not result_rtol.all() and not result_atol.all(): + if numpy.sum(result_rtol == False) > real_result.size * loss and numpy.sum(result_atol == False) > real_result.size * loss: # 误差超出预期时返回打印错误,返回对比失败 + print("[ERROR] result error") + return False + print("test Operation success!") + return True + + +if __name__ == '__main__': + verify_result(sys.argv[1],sys.argv[2]) diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/src/main.cpp b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/src/main.cpp new file mode 100644 index 000000000..0bd444165 --- /dev/null +++ b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/src/main.cpp @@ -0,0 +1,195 @@ +/* + * Copyright (c) Huawei Technologies Co., Ltd. 2023. All rights reserved. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +#include "main.h" + +bool SetInputData(std::vector &inputData){ + char *xPath = "./script/input/input0.bin"; + InputData inputX; + inputX.data = ReadBinFile(xPath,inputX.size); + inputData.push_back(inputX); + return true; +} + +bool SetOperationInputDesc(atb::SVector &intensorDescs){ + atb::TensorDesc xDesc; + xDesc.dtype = ACL_FLOAT; + xDesc.format = ACL_FORMAT_ND; + xDesc.shape.dimNum = 4; + xDesc.shape.dims[0] = 3; + xDesc.shape.dims[1] = 4; + xDesc.shape.dims[2] = 133; + xDesc.shape.dims[3] = 4095; + intensorDescs.at(0) = xDesc; +} + + + +static void SetCurrentDevice() +{ + const int deviceId = 0; + std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; + int ret = aclrtSetDevice(deviceId); + if (ret != 0) { + std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; + return; + } + std::cout << "[INFO]: aclrtSetDevice success" << std::endl; +} + + +static void FreeTensor(atb::Tensor &tensor) +{ + if (tensor.deviceData) { + int ret = aclrtFree(tensor.deviceData); + if (ret != 0) { + std::cout << "[ERROR]: aclrtFree fail" << std::endl; + } + tensor.deviceData = nullptr; + tensor.dataSize = 0; + } + if (tensor.hostData) { + int ret = aclrtFreeHost(tensor.hostData); + if (ret != 0) { + std::cout << "[ERROR]: aclrtFreeHost fail, ret = " << ret << std::endl; + } + tensor.hostData = nullptr; + tensor.dataSize = 0; + } +} + +static void FreeTensors(atb::SVector &inTensors) +{ + for (size_t i = 0; i < inTensors.size(); ++i) { + FreeTensor(inTensors.at(i)); + } +} +bool SaveMemoryToBinFile(void* memoryAddress, size_t memorySize, size_t i) { + // 创建 output 目录(如果不存在) + std::filesystem::create_directories("output"); + + // 生成文件名 + std::string filename = "script/output/output_" + std::to_string(i) + ".bin"; + + // 打开文件以二进制写入模式 + std::ofstream file(filename, std::ios::binary); + if (!file) { + std::cerr << "无法打开文件: " << filename << std::endl; + return false; + } + + // 写入数据 + file.write(static_cast(memoryAddress), memorySize); + if (!file) { + std::cerr << "写入文件时出错: " << filename << std::endl; + file.close(); + return false; + } + + // 关闭文件 + file.close(); + std::cout << "数据已成功保存到: " << filename << std::endl; + return true; +} + +int main(int argc, const char *argv[]) +{ + const int deviceId = 0; + std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; + int ret = aclrtSetDevice(deviceId); + if (ret != 0) { + std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; + return 1; + } + std::cout << "[INFO]: aclrtSetDevice success" << std::endl; + atb::Context *context = nullptr; + ret = atb::CreateContext(&context); + void *stream = nullptr; + ret = aclrtCreateStream(&stream); + if (ret != 0) { + std::cout << "[ERROR]: AsdRtStreamCreate fail, ret:" << ret << std::endl; + return 1; + } + context->SetExecuteStream(stream); + + std::vector input; + SetInputData(input); + + EyeAttrParam eyeAttrParam; + eyeAttrParam.num_rows = 133; + eyeAttrParam.num_columns = 4095; + + std::vector batchShape = {3,4}; + eyeAttrParam.batch_shape = aclCreateIntArray(batchShape.data(),batchShape.size()); + eyeAttrParam.dtype = 0; + EyeOperation *op = new EyeOperation("Eye",eyeAttrParam); + std::cout << "[INFO]: complete CreateOp!" << std::endl; + + if(input.size() != op->GetInputNum()) std::cout << "[ERROR]: Operation actual input num is not equal to GetInputNum()"; + + atb::SVector intensorDescs; + atb::SVector outtensorDescs; + intensorDescs.resize(op->GetInputNum()); + outtensorDescs.resize(op->GetOutputNum()); + SetOperationInputDesc(intensorDescs); + atb::Status st = op->InferShape(intensorDescs,outtensorDescs); + if (st != 0) { + std::cout << "[ERROR]: Operation InferShape fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation InferShape success" << std::endl; + atb::VariantPack variantPack; + variantPack.inTensors.resize(op->GetInputNum()); + for(size_t i=0;iGetInputNum();i++){ + variantPack.inTensors.at(i).desc = intensorDescs.at(i); + variantPack.inTensors.at(i).hostData = input[i].data; + variantPack.inTensors.at(i).dataSize = input[i].size; + CheckAcl(aclrtMalloc(&variantPack.inTensors.at(i).deviceData, input[i].size, ACL_MEM_MALLOC_HUGE_FIRST)); + CheckAcl(aclrtMemcpy(variantPack.inTensors.at(i).deviceData, input[i].size, input[i].data, input[i].size, ACL_MEMCPY_HOST_TO_DEVICE)); + } + std::cout << "[INFO]: Operation Input prepare sucess" << std::endl; + + uint64_t workspaceSize = 0; + st = op->Setup(variantPack, workspaceSize, context); + if (st != 0) { + std::cout << "[ERROR]: Operation setup fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation setup success" << std::endl; + void *workspace = nullptr; + if (workspaceSize > 0) { + aclrtMalloc(&workspace, workspaceSize, ACL_MEM_MALLOC_HUGE_FIRST); + } + + std::cout << "[INFO]: Operation execute start" << std::endl; + st = op->Execute(variantPack, (uint8_t*)workspace, workspaceSize, context); + if (st != 0) { + std::cout << "[ERROR]: Operation execute fail" << std::endl; + return -1; + } + std::cout << "[INFO]: Operation execute success" << std::endl; + + CheckAcl(aclrtMemcpy(variantPack.inTensors.at(0).hostData, variantPack.inTensors.at(0).dataSize, variantPack.inTensors.at(0).deviceData, + variantPack.inTensors.at(0).dataSize, ACL_MEMCPY_DEVICE_TO_HOST)); + SaveMemoryToBinFile(variantPack.inTensors.at(0).hostData,variantPack.inTensors.at(0).dataSize,0); + + + FreeTensors(variantPack.inTensors); + st = atb::DestroyContext(context); + CheckAcl(aclrtDestroyStream(stream)); + CheckAcl(aclrtResetDevice(0)); + CheckAcl(aclFinalize()); + return atb::ErrorType::NO_ERROR; +} diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/src/main.h b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/src/main.h new file mode 100644 index 000000000..015c779c3 --- /dev/null +++ b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/src/main.h @@ -0,0 +1,55 @@ +#include +#include +#include +#include +#include +#include +#include +#include "securec.h" +#include "atb/atb_infer.h" +#include "aclnn_eye_operation.h" +#include + +struct InputData{ + void* data; + uint64_t size; +}; +aclError CheckAcl(aclError ret) +{ + if (ret != ACL_ERROR_NONE) { + std::cerr << __FILE__ << ":" << __LINE__ << " aclError:" << ret << std::endl; + } + return ret; +} +void* ReadBinFile(const char* filename, size_t& size) { + std::ifstream file(filename, std::ios::binary | std::ios::ate); + if (!file) { + std::cerr << "无法打开文件: " << filename << std::endl; + return nullptr; + } + + // 获取文件大小 + size = file.tellg(); + file.seekg(0, std::ios::beg); + + // 分配内存 + void* buffer; + int ret = aclrtMallocHost(&buffer,size); + if (!buffer) { + std::cerr << "内存分配失败" << std::endl; + file.close(); + return nullptr; + } + + // 读取文件内容到内存 + file.read(static_cast(buffer), size); + if (!file) { + std::cerr << "读取文件失败" << std::endl; + delete[] static_cast(buffer); + file.close(); + return nullptr; + } + + file.close(); + return buffer; +} \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/readme.md b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/readme.md new file mode 100644 index 000000000..e0a38627c --- /dev/null +++ b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/readme.md @@ -0,0 +1,70 @@ +## 概述 + +本样例基于AscendC自定义[Add](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)算子,开发了ATB插件并进行了插件调用测试. + +## 项目结构介绍 +``` + +├── AddAscendC //Add AscendC自定义算子工程 + +├── AddOperationATBPlugin //AddOperation ATB插件代码 + +├── AddOperationTest //AddOperation 测试代码 +``` +## 样例运行 + +### Add AscendC自定义算子部署 + +参照cann-ops仓[add_custom算子](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)" **算子包编译部署** "章节 + +### AddOperation ATB插件部署 + +- 运行编译脚本完成部署(脚本会生成静态库.a文件,同时将头文件拷贝到/usr/include,.a文件拷贝到/usr/local/lib下) + + ``` + cd AddOperationATBPlugin + bash build.sh + ``` + +### AddOperation测试 + +- 运行脚本完成算子测试 + + ```shell + cd AddOperationTest + bash run.sh + ``` + +## AddOperation算子介绍 + +### 功能 + +实现两个输入张量相加 + +### 定义 + +``` +struct AddParam { +}; +``` + +### 参数列表 + +该算子参数为空 + +### 输入 + +| **参数** | **维度** | **数据类型** | **格式** | 描述 | +| -------- | -------------------------- | --------------------- | -------- | ---------- | +| x | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | +| y | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | + +### 输出 + +| **参数** | **维度** | **数据类型** | **格式** | 描述 | +| -------- | -------------------------- | --------------------- | -------- | ---------------------------------------- | +| output | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输出tensor。数据类型和shape与x保持一致。 | + +### 规格约束 + +暂无 \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_eye/EyeCustom/README.md b/operator/ascendc/4_best_practices/25_eye/EyeCustom/README.md new file mode 100644 index 000000000..25c5ffc7e --- /dev/null +++ b/operator/ascendc/4_best_practices/25_eye/EyeCustom/README.md @@ -0,0 +1,3 @@ +### 概述 + +具体算子工程,编译部署请见[AddCustom](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom) \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_eye/README.md b/operator/ascendc/4_best_practices/25_eye/README.md new file mode 100644 index 000000000..16780e8fa --- /dev/null +++ b/operator/ascendc/4_best_practices/25_eye/README.md @@ -0,0 +1,49 @@ +### 概述 + +本样例介基于AddCustom算子工程,介绍了单算子工程以及ATB插件调用方法 + +### 目录结构介绍 + +``` +├── 24_add_custom // 使用框架调用的方式调用AllGatherMatmul自定义算子 +│ ├── ATBInvocation // 通过ATB调用的方式调用 +│ ├── AddCustom // AddCustom算子工程,引用cann-ops仓的AddCustom工程 +``` + +### 算子描述 + +`AddCustom`算子返回两个数据相加的结果。 + +### 算子规格描述 + +| 算子类型(OpType) | AddCustom | | | | +| ---------------- | ---------- | -------- | --------- | ------ | +| 算子输入 | name | shape | data type | format | +| x | 8 * 2048 | float16 | ND | | +| y | 8 * 2048 | float16 | ND | | +| 算子输出 | z | 8 * 2048 | float16 | ND | +| 核函数名 | add_custom | | | | + +### 支持的产品型号 + +本样例支持如下产品型号: + +- Atlas 训练系列产品 +- Atlas 推理系列产品 +- Atlas A2训练系列产品 +- Atlas 800I A2推理产品 +- Atlas 200I/500 A2推理产品 + +### 算子编译部署 + +- [AddCustom编译部署(需要下载cann-ops仓)](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom) + +### 算子ATB插件调用 + +- [AddOperation编译调用](https://gitee.com/youmoxiao/samples/tree/atb_op/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation) + +## 更新说明 + +| 时间 | 更新事项 | +| -------- | -------- | +| 2025/4/1 | 新增样例 | \ No newline at end of file -- Gitee From 9fc66732b2401aa73286b97ba2e439664ce83556 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E5=88=98=E5=96=9C=E5=BC=BA?= Date: Tue, 1 Apr 2025 12:41:30 +0000 Subject: [PATCH 16/26] update --- .../AddOperationATBPlugin/CMakeLists.txt | 2 +- .../AddOperationTest/script/gen_data.py | 10 ++-------- .../AddOperationTest/script/verify_result.py | 7 ++----- .../ATBInvocation/AddOperationTest/src/main.cpp | 12 ++++++------ 4 files changed, 11 insertions(+), 20 deletions(-) diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/CMakeLists.txt b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/CMakeLists.txt index c89578fcd..eb8a058bc 100644 --- a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/CMakeLists.txt +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationATBPlugin/CMakeLists.txt @@ -18,4 +18,4 @@ link_directories( ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 ) add_library(atb_plugin_operations STATIC ${atb_plugin_operations_SRCS}) -target_link_libraries(atb_plugin_operations PRIVATE ${ACL_SHARED_LIBS} ${ACL_SHARED_LIBS} ascendcl) \ No newline at end of file +target_link_libraries(atb_plugin_operations PRIVATE ${ACL_SHARED_LIBS} ascendcl) \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/gen_data.py b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/gen_data.py index 83ecafbce..b85e30bc7 100644 --- a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/gen_data.py +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/gen_data.py @@ -1,20 +1,14 @@ import torch import os def gen_golden_data_simple(): - dtype = "float" - input1 = torch.randn(133, 4095, dtype=torch.float) - input2 = torch.randn(133, 4095, dtype=torch.float) + input1 = torch.randn(108, 4096, dtype=torch.float16) + input2 = torch.randn(108, 4096, dtype=torch.float16) golden = input1 + input2 - os.system("mkdir -p input") - os.system("mkdir -p output") input1.numpy().tofile('./script/input/input0.bin') input2.numpy().tofile('./script/input/input1.bin') golden.numpy().tofile("./script/output/golden0.bin") - - with open("./output/meta", "w") as fp: - print(dtype, file=fp) if __name__ == "__main__": gen_golden_data_simple() diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/verify_result.py b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/verify_result.py index 629845e13..d64d11014 100644 --- a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/verify_result.py +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/script/verify_result.py @@ -6,11 +6,8 @@ loss = 1e-3 # 容忍偏差,一般fp16要求绝对误差和相对误差均不 minimum = 10e-10 def verify_result(real_result, golden): - with open("output/meta", "r") as fp: - dtype_str = fp.readline().strip() - dtype = eval(dtype_str) - real_result = numpy.fromfile(real_result, dtype=dtype) # 从bin文件读取实际运算结果 - golden = numpy.fromfile(golden, dtype=dtype) # 从bin文件读取预期运算结果 + real_result = numpy.fromfile(real_result, dtype=numpy.float16) # 从bin文件读取实际运算结果 + golden = numpy.fromfile(golden, dtype=numpy.float16) # 从bin文件读取预期运算结果 print("=" * 50, real_result[:5], golden[:5], "=" * 50, sep='\n', end='\n', file=sys.stderr) result = numpy.abs(real_result - golden) # 计算运算结果和预期结果偏差 deno = numpy.maximum(numpy.abs(real_result), numpy.abs(golden)) # 获取最大值并组成新数组 diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp index 258c0cb3c..2fca70549 100644 --- a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp @@ -29,18 +29,18 @@ bool SetInputData(std::vector &inputData){ bool SetOperationInputDesc(atb::SVector &intensorDescs){ atb::TensorDesc xDesc; - xDesc.dtype = ACL_FLOAT; + xDesc.dtype = ACL_FLOAT16; xDesc.format = ACL_FORMAT_ND; xDesc.shape.dimNum = 2; - xDesc.shape.dims[0] = 133; - xDesc.shape.dims[1] = 4095; + xDesc.shape.dims[0] = 108; + xDesc.shape.dims[1] = 4096; atb::TensorDesc yDesc; - yDesc.dtype = ACL_FLOAT; + yDesc.dtype = ACL_FLOAT16; yDesc.format = ACL_FORMAT_ND; yDesc.shape.dimNum = 2; - yDesc.shape.dims[0] = 133; - yDesc.shape.dims[1] = 4095; + yDesc.shape.dims[0] = 108; + yDesc.shape.dims[1] = 4096; intensorDescs.at(0) = xDesc; intensorDescs.at(1) = yDesc; -- Gitee From b114fe8759eecc086d0ff2fb1bc301a3ccae9431 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Tue, 1 Apr 2025 12:43:06 +0000 Subject: [PATCH 17/26] =?UTF-8?q?=E5=88=A0=E9=99=A4=E6=96=87=E4=BB=B6=20op?= =?UTF-8?q?erator/ascendc/4=5Fbest=5Fpractices/25=5FEye=5Fcustom?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .../EyeOperationATBPlugin/CMakeLists.txt | 21 -- .../aclnn_eye_operation.cpp | 133 ------------ .../aclnn_eye_operation.h | 60 ------ .../EyeOperationATBPlugin/build.sh | 33 --- .../EyeOperationTest/CMakeLists.txt | 40 ---- .../EyeOperationTest/script/gen_data.py | 16 -- .../EyeOperationTest/script/run.sh | 52 ----- .../EyeOperationTest/script/verify_result.py | 25 --- .../EyeOperationTest/src/main.cpp | 195 ------------------ .../ATBInvocation/EyeOperationTest/src/main.h | 55 ----- .../25_Eye_custom/ATBInvocation/readme.md | 70 ------- .../25_Eye_custom/EyeCustom/README.md | 3 - .../4_best_practices/25_Eye_custom/README.md | 49 ----- 13 files changed, 752 deletions(-) delete mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt delete mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp delete mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h delete mode 100755 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/build.sh delete mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/CMakeLists.txt delete mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/gen_data.py delete mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/run.sh delete mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/verify_result.py delete mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.cpp delete mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.h delete mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/readme.md delete mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/EyeCustom/README.md delete mode 100644 operator/ascendc/4_best_practices/25_Eye_custom/README.md diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt deleted file mode 100644 index c89578fcd..000000000 --- a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright 2024 Tencent Inc. All rights reserved. -# -# ============================================================================== -cmake_minimum_required(VERSION 3.14) -set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") -file(GLOB_RECURSE atb_plugin_operations_SRCS "./*.cpp") -set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") -message(STATUS "atb_plugin_operations_SRCS: ${atb_plugin_operations_SRCS}") -include_directories( - $ENV{ATB_HOME_PATH}/include - $ENV{ASCEND_HOME_PATH}/include - ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 - ${CMAKE_CURRENT_SOURCE_DIR} - ) -link_directories( - $ENV{ATB_HOME_PATH}/lib - $ENV{ASCEND_HOME_PATH}/lib64 - ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 - ) -add_library(atb_plugin_operations STATIC ${atb_plugin_operations_SRCS}) -target_link_libraries(atb_plugin_operations PRIVATE ${ACL_SHARED_LIBS} ${ACL_SHARED_LIBS} ascendcl) \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp deleted file mode 100644 index a4f156c4b..000000000 --- a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp +++ /dev/null @@ -1,133 +0,0 @@ -#include "aclnn_eye_operation.h" -#include "aclnn_eye.h" - -EyeOperation::EyeOperation(const std::string &name, EyeAttrParam param){ - attrParam = param; - opName_ = name; -} - -atb::SVector GetCopyTensorStride(atb::Dims &tensorDims) -{ - atb::SVector tmpStrides(tensorDims.dimNum, 1); - if (tensorDims.dimNum > 8) { // 8: tensor最大维度数量 - printf("tensor's dimNum is larger than 8, GetCopyTensorStride failed."); - return tmpStrides; - } - for (int64_t i = static_cast(tensorDims.dimNum) - 2; i >= 0; i--) { - tmpStrides[i] = (tensorDims.dims[i + 1] * tmpStrides[i + 1]); - } - return tmpStrides; -} - -std::shared_ptr EyeOperation::CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx) -{ - auto aclnnTensor = std::make_shared(); - aclnnTensor->tensorIdx = static_cast(tensorIdx); - aclnnTensor->needUpdateTensorDataPtr = true; - aclnnTensor->atbTensor = atbTensor; - aclnnTensor->strides = GetCopyTensorStride(atbTensor.desc.shape); - - // 创建Aclnn tensor - aclnnTensor->tensor = aclCreateTensor(atbTensor.desc.shape.dims, - atbTensor.desc.shape.dimNum, - atbTensor.desc.dtype, - aclnnTensor->strides.data(), - 0, - atbTensor.desc.format, - atbTensor.desc.shape.dims, - atbTensor.desc.shape.dimNum, - atbTensor.deviceData); - return aclnnTensor; -} - -atb::Status EyeOperation::UpdateAclnnVariantPack(const atb::VariantPack &variantPack) -{ - // 更新inTensor的device地址 - for (size_t i = 0; i < aclInTensors_.size(); ++i) { - int ret = -1; - if (!aclInTensors_[i]->needUpdateTensorDataPtr) { - continue; - } - aclInTensors_[i]->atbTensor = variantPack.inTensors.at(i); - ret = aclSetInputTensorAddr(aclExecutor_, - aclInTensors_[i]->tensorIdx, - aclInTensors_[i]->tensor, - aclInTensors_[i]->atbTensor.deviceData); - - if (ret != 0) { - printf("set input fail"); - return atb::ERROR_CANN_ERROR; - } - } - - // 更新outTensor的device地址 - for (size_t i = 0; i < aclOutTensors_.size(); ++i) { - int ret = -1; - if (!aclOutTensors_[i]->needUpdateTensorDataPtr) { - continue; - } - aclOutTensors_[i]->atbTensor = variantPack.outTensors.at(i); - ret = aclSetOutputTensorAddr(aclExecutor_, - aclOutTensors_[i]->tensorIdx, - aclOutTensors_[i]->tensor, - aclOutTensors_[i]->atbTensor.deviceData); - - if (ret != 0) { - printf("set output fail"); - return atb::ERROR_CANN_ERROR; - } - } - return atb::NO_ERROR; -} - -atb::Status EyeOperation::Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) { - - aclInTensors_.resize(GetInputNum()); - for (size_t i = 0; i < aclInTensors_.size(); ++i) { - auto aclnnTensor = CreateAclnnTensor(variantPack.inTensors.at(i), i); - if (aclnnTensor->tensor == nullptr) { - printf("creat input tensor %ld fail", i); - return atb::ERROR_INTERNAL_ERROR; - } - aclInTensors_[i] = aclnnTensor; - } - - auto ret = aclnnEyeGetWorkspaceSize(aclInTensors_.at(0)->tensor, - attrParam.num_rows, - attrParam.num_columns, - attrParam.batch_shape, - attrParam.dtype, - &workspaceSize_, - &aclExecutor_); - - workspaceSize = workspaceSize_; - return ret; - -} - -atb::Status EyeOperation::Execute(const atb::VariantPack &variantPack, uint8_t *workspace, uint64_t workspaceSize, atb::Context *context) { - - - - aclrtStream stream = context->GetExecuteStream(); - if (!stream) { - printf("get stream fail"); - return atb::ERROR_INVALID_PARAM; - } - // 更新数据传入的地址 - int ret = UpdateAclnnVariantPack(variantPack); - if (ret != 0) { - printf("UpdateAclnnVariantPack fail"); - return atb::ERROR_CANN_ERROR; - } - ret = aclnnEye(workspace, workspaceSize_, aclExecutor_, stream); - - return ret; -} - -atb::Status EyeOperation::InferShape( - const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const -{ - outTensorDesc.at(0) = inTensorDesc.at(0); - return atb::NO_ERROR; -} diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h deleted file mode 100644 index b09697135..000000000 --- a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h +++ /dev/null @@ -1,60 +0,0 @@ -#include -#include -#include -#include -#include -#include "atb/infer_op_params.h" -#include - -struct EyeAttrParam -{ - uint64_t num_rows; - uint64_t num_columns = 0; - std::vector batchShape = {1}; - aclIntArray* batch_shape = aclCreateIntArray(batchShape.data(),batchShape.size()); - uint64_t dtype = 0; -}; - -struct AclnnTensor -{ -public: - atb::Tensor atbTensor; // - aclTensor *tensor = nullptr; - int tensorIdx = -1; // aclTensor在aclExecutor中的index - bool needUpdateTensorDataPtr = false; - atb::SVector strides = {}; -}; - -class EyeOperation: public atb::Operation{ -public: - EyeOperation(const std::string &name, EyeAttrParam param); - atb::Status Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) override; - atb::Status Execute(const atb::VariantPack &variantPack, uint8_t *workspace, - uint64_t workspaceSize, atb::Context *context) override; - atb::Status InferShape( - const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const; - std::shared_ptr CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx); - atb::Status UpdateAclnnVariantPack(const atb::VariantPack &variantPack); - - uint32_t GetInputNum() const - { - return 1; // 算子入参个数 - } - - uint32_t GetOutputNum() const - { - return 1; // 算子出参个数 - } - std::string GetName() const - { - return opName_; - } - - aclOpExecutor *aclExecutor_ = nullptr; - EyeAttrParam attrParam; - std::string opName_; - uint64_t workspaceSize_; - - atb::SVector> aclInTensors_; - atb::SVector> aclOutTensors_; -}; \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/build.sh b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/build.sh deleted file mode 100755 index ba28cc143..000000000 --- a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationATBPlugin/build.sh +++ /dev/null @@ -1,33 +0,0 @@ -#!/bin/bash - -# 定义构建目录 -BUILD_DIR="build" - -# 创建构建目录 -mkdir -p "$BUILD_DIR" -cd "$BUILD_DIR" - -# 运行 CMake 配置和编译 -cmake .. -make - -# 查找生成的 .a 文件 -A_FILE=$(find . -name "*.a" -type f) - -# 检查是否找到了 .a 文件 -if [ -z "$A_FILE" ]; then - echo "未找到 .a 文件,编译可能失败。" - exit 1 -fi - -# 复制头文件到 /usr/include -HEADER_FILES=$(find .. -name "*.h" -type f) -for header in $HEADER_FILES; do - cp "$header" /usr/include/ -done - -# 复制 .a 文件到 /usr/local/lib -cp "$A_FILE" /usr/local/lib/ - -echo "构建完成,头文件和 .a 文件已复制到目标目录。" - diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/CMakeLists.txt b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/CMakeLists.txt deleted file mode 100644 index b9c2cec46..000000000 --- a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/CMakeLists.txt +++ /dev/null @@ -1,40 +0,0 @@ -cmake_minimum_required(VERSION 3.16) -project("test_model") -option(USE_CXX11_ABI "USE_CXX11_ABI" OFF) -set(CMAKE_BUILD_TYPE Debug) -if(USE_CXX11_ABI) - set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=1") -else() - set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") -endif() - -set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -std=c++17") - -set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") - -include_directories( - $ENV{ATB_HOME_PATH}/include - $ENV{ASCEND_HOME_PATH}/include - ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 - ${CMAKE_CURRENT_SOURCE_DIR} - ./src/ - ) - -message(".h path = ${CUST_PKG_PATH}/include") -link_directories( - $ENV{ATB_HOME_PATH}/lib - $ENV{ASCEND_HOME_PATH}/lib64 - ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 - ) - - -link_libraries(atb_plugin_operations) -add_executable(test_model ./src/main.cpp) - -# 列出所有的头文件目录 -target_link_libraries(test_model PRIVATE atb ascendcl cust_opapi opapi nnopbase pthread) # 添加自定义算子so及适配so - - - - - diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/gen_data.py b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/gen_data.py deleted file mode 100644 index b0a8ba36f..000000000 --- a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/gen_data.py +++ /dev/null @@ -1,16 +0,0 @@ -import torch -import os -def gen_golden_data_simple(): - dtype = "float32" - input1 = torch.zeros(3 , 4, 133, 4095, dtype=torch.float) - golden = torch.eye(133,4095, dtype=torch.float) - golden = golden.unsqueeze(0).unsqueeze(0) - golden = golden.repeat(3,4,1,1) - input1.numpy().tofile('./script/input/input0.bin') - golden.numpy().tofile("./script/output/golden0.bin") - - with open("./script/output/meta", "w") as fp: - print(dtype, file=fp) - -if __name__ == "__main__": - gen_golden_data_simple() diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/run.sh b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/run.sh deleted file mode 100644 index 450b717f2..000000000 --- a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/run.sh +++ /dev/null @@ -1,52 +0,0 @@ -#!/bin/bash - - -# 步骤1: 运行gen_data.py生成输入bin文件和golden标杆输出数据 -echo "正在生成输入数据和golden标杆数据..." -mkdir -p script/input -mkdir -p script/output -python3 script/gen_data.py -if [ $? -ne 0 ]; then - echo "生成数据失败,脚本终止。" - exit 1 -fi - -# 步骤2: 创建构建目录并进入 -mkdir -p build -cd build -if [ $? -ne 0 ]; then - echo "无法进入构建目录,脚本终止。" - exit 1 -fi - -# 步骤3: 使用CMake配置项目 -echo "正在配置CMake项目..." -cmake .. -if [ $? -ne 0 ]; then - echo "CMake配置失败,脚本终止。" - exit 1 -fi - -# 步骤4: 编译代码 -echo "正在编译代码..." -make -if [ $? -ne 0 ]; then - echo "编译失败,脚本终止。" - exit 1 -fi - -mv test_model ../ -cd .. - -# 步骤5: 运行可执行文件生成实际输出文件 -echo "正在运行可执行文件生成实际输出..." -./test_model -if [ $? -ne 0 ]; then - echo "运行可执行文件失败,脚本终止。" - exit 1 -fi - -# 步骤6: 调用verify_result.py进行golden标杆数据和实际输出数据的比对 -echo "正在验证结果..." -python3 script/verify_result.py script/output/output_0.bin script/output/golden0.bin - diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/verify_result.py b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/verify_result.py deleted file mode 100644 index a0798d50a..000000000 --- a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/script/verify_result.py +++ /dev/null @@ -1,25 +0,0 @@ -import os -import sys -import numpy - -loss = 1e-3 # 容忍偏差,一般fp16要求绝对误差和相对误差均不超过千分之一 -minimum = 10e-10 - -def verify_result(real_result, golden): - real_result = numpy.fromfile(real_result, dtype=numpy.float32) # 从bin文件读取实际运算结果 - golden = numpy.fromfile(golden, dtype=numpy.float32) # 从bin文件读取预期运算结果 - print("=" * 50, real_result[:5], golden[:5], "=" * 50, sep='\n', end='\n', file=sys.stderr) - result = numpy.abs(real_result - golden) # 计算运算结果和预期结果偏差 - deno = numpy.maximum(numpy.abs(real_result), numpy.abs(golden)) # 获取最大值并组成新数组 - result_atol = numpy.less_equal(result, loss) # 计算绝对误差 - result_rtol = numpy.less_equal(result / numpy.add(deno, minimum), loss) # 计算相对误差 - if not result_rtol.all() and not result_atol.all(): - if numpy.sum(result_rtol == False) > real_result.size * loss and numpy.sum(result_atol == False) > real_result.size * loss: # 误差超出预期时返回打印错误,返回对比失败 - print("[ERROR] result error") - return False - print("test Operation success!") - return True - - -if __name__ == '__main__': - verify_result(sys.argv[1],sys.argv[2]) diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.cpp b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.cpp deleted file mode 100644 index 0bd444165..000000000 --- a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.cpp +++ /dev/null @@ -1,195 +0,0 @@ -/* - * Copyright (c) Huawei Technologies Co., Ltd. 2023. All rights reserved. - * - * Licensed under the Apache License, Version 2.0 (the "License"); - * you may not use this file except in compliance with the License. - * You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ -#include "main.h" - -bool SetInputData(std::vector &inputData){ - char *xPath = "./script/input/input0.bin"; - InputData inputX; - inputX.data = ReadBinFile(xPath,inputX.size); - inputData.push_back(inputX); - return true; -} - -bool SetOperationInputDesc(atb::SVector &intensorDescs){ - atb::TensorDesc xDesc; - xDesc.dtype = ACL_FLOAT; - xDesc.format = ACL_FORMAT_ND; - xDesc.shape.dimNum = 4; - xDesc.shape.dims[0] = 3; - xDesc.shape.dims[1] = 4; - xDesc.shape.dims[2] = 133; - xDesc.shape.dims[3] = 4095; - intensorDescs.at(0) = xDesc; -} - - - -static void SetCurrentDevice() -{ - const int deviceId = 0; - std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; - int ret = aclrtSetDevice(deviceId); - if (ret != 0) { - std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; - return; - } - std::cout << "[INFO]: aclrtSetDevice success" << std::endl; -} - - -static void FreeTensor(atb::Tensor &tensor) -{ - if (tensor.deviceData) { - int ret = aclrtFree(tensor.deviceData); - if (ret != 0) { - std::cout << "[ERROR]: aclrtFree fail" << std::endl; - } - tensor.deviceData = nullptr; - tensor.dataSize = 0; - } - if (tensor.hostData) { - int ret = aclrtFreeHost(tensor.hostData); - if (ret != 0) { - std::cout << "[ERROR]: aclrtFreeHost fail, ret = " << ret << std::endl; - } - tensor.hostData = nullptr; - tensor.dataSize = 0; - } -} - -static void FreeTensors(atb::SVector &inTensors) -{ - for (size_t i = 0; i < inTensors.size(); ++i) { - FreeTensor(inTensors.at(i)); - } -} -bool SaveMemoryToBinFile(void* memoryAddress, size_t memorySize, size_t i) { - // 创建 output 目录(如果不存在) - std::filesystem::create_directories("output"); - - // 生成文件名 - std::string filename = "script/output/output_" + std::to_string(i) + ".bin"; - - // 打开文件以二进制写入模式 - std::ofstream file(filename, std::ios::binary); - if (!file) { - std::cerr << "无法打开文件: " << filename << std::endl; - return false; - } - - // 写入数据 - file.write(static_cast(memoryAddress), memorySize); - if (!file) { - std::cerr << "写入文件时出错: " << filename << std::endl; - file.close(); - return false; - } - - // 关闭文件 - file.close(); - std::cout << "数据已成功保存到: " << filename << std::endl; - return true; -} - -int main(int argc, const char *argv[]) -{ - const int deviceId = 0; - std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; - int ret = aclrtSetDevice(deviceId); - if (ret != 0) { - std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; - return 1; - } - std::cout << "[INFO]: aclrtSetDevice success" << std::endl; - atb::Context *context = nullptr; - ret = atb::CreateContext(&context); - void *stream = nullptr; - ret = aclrtCreateStream(&stream); - if (ret != 0) { - std::cout << "[ERROR]: AsdRtStreamCreate fail, ret:" << ret << std::endl; - return 1; - } - context->SetExecuteStream(stream); - - std::vector input; - SetInputData(input); - - EyeAttrParam eyeAttrParam; - eyeAttrParam.num_rows = 133; - eyeAttrParam.num_columns = 4095; - - std::vector batchShape = {3,4}; - eyeAttrParam.batch_shape = aclCreateIntArray(batchShape.data(),batchShape.size()); - eyeAttrParam.dtype = 0; - EyeOperation *op = new EyeOperation("Eye",eyeAttrParam); - std::cout << "[INFO]: complete CreateOp!" << std::endl; - - if(input.size() != op->GetInputNum()) std::cout << "[ERROR]: Operation actual input num is not equal to GetInputNum()"; - - atb::SVector intensorDescs; - atb::SVector outtensorDescs; - intensorDescs.resize(op->GetInputNum()); - outtensorDescs.resize(op->GetOutputNum()); - SetOperationInputDesc(intensorDescs); - atb::Status st = op->InferShape(intensorDescs,outtensorDescs); - if (st != 0) { - std::cout << "[ERROR]: Operation InferShape fail" << std::endl; - return -1; - } - std::cout << "[INFO]: Operation InferShape success" << std::endl; - atb::VariantPack variantPack; - variantPack.inTensors.resize(op->GetInputNum()); - for(size_t i=0;iGetInputNum();i++){ - variantPack.inTensors.at(i).desc = intensorDescs.at(i); - variantPack.inTensors.at(i).hostData = input[i].data; - variantPack.inTensors.at(i).dataSize = input[i].size; - CheckAcl(aclrtMalloc(&variantPack.inTensors.at(i).deviceData, input[i].size, ACL_MEM_MALLOC_HUGE_FIRST)); - CheckAcl(aclrtMemcpy(variantPack.inTensors.at(i).deviceData, input[i].size, input[i].data, input[i].size, ACL_MEMCPY_HOST_TO_DEVICE)); - } - std::cout << "[INFO]: Operation Input prepare sucess" << std::endl; - - uint64_t workspaceSize = 0; - st = op->Setup(variantPack, workspaceSize, context); - if (st != 0) { - std::cout << "[ERROR]: Operation setup fail" << std::endl; - return -1; - } - std::cout << "[INFO]: Operation setup success" << std::endl; - void *workspace = nullptr; - if (workspaceSize > 0) { - aclrtMalloc(&workspace, workspaceSize, ACL_MEM_MALLOC_HUGE_FIRST); - } - - std::cout << "[INFO]: Operation execute start" << std::endl; - st = op->Execute(variantPack, (uint8_t*)workspace, workspaceSize, context); - if (st != 0) { - std::cout << "[ERROR]: Operation execute fail" << std::endl; - return -1; - } - std::cout << "[INFO]: Operation execute success" << std::endl; - - CheckAcl(aclrtMemcpy(variantPack.inTensors.at(0).hostData, variantPack.inTensors.at(0).dataSize, variantPack.inTensors.at(0).deviceData, - variantPack.inTensors.at(0).dataSize, ACL_MEMCPY_DEVICE_TO_HOST)); - SaveMemoryToBinFile(variantPack.inTensors.at(0).hostData,variantPack.inTensors.at(0).dataSize,0); - - - FreeTensors(variantPack.inTensors); - st = atb::DestroyContext(context); - CheckAcl(aclrtDestroyStream(stream)); - CheckAcl(aclrtResetDevice(0)); - CheckAcl(aclFinalize()); - return atb::ErrorType::NO_ERROR; -} diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.h b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.h deleted file mode 100644 index 015c779c3..000000000 --- a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/EyeOperationTest/src/main.h +++ /dev/null @@ -1,55 +0,0 @@ -#include -#include -#include -#include -#include -#include -#include -#include "securec.h" -#include "atb/atb_infer.h" -#include "aclnn_eye_operation.h" -#include - -struct InputData{ - void* data; - uint64_t size; -}; -aclError CheckAcl(aclError ret) -{ - if (ret != ACL_ERROR_NONE) { - std::cerr << __FILE__ << ":" << __LINE__ << " aclError:" << ret << std::endl; - } - return ret; -} -void* ReadBinFile(const char* filename, size_t& size) { - std::ifstream file(filename, std::ios::binary | std::ios::ate); - if (!file) { - std::cerr << "无法打开文件: " << filename << std::endl; - return nullptr; - } - - // 获取文件大小 - size = file.tellg(); - file.seekg(0, std::ios::beg); - - // 分配内存 - void* buffer; - int ret = aclrtMallocHost(&buffer,size); - if (!buffer) { - std::cerr << "内存分配失败" << std::endl; - file.close(); - return nullptr; - } - - // 读取文件内容到内存 - file.read(static_cast(buffer), size); - if (!file) { - std::cerr << "读取文件失败" << std::endl; - delete[] static_cast(buffer); - file.close(); - return nullptr; - } - - file.close(); - return buffer; -} \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/readme.md b/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/readme.md deleted file mode 100644 index e0a38627c..000000000 --- a/operator/ascendc/4_best_practices/25_Eye_custom/ATBInvocation/readme.md +++ /dev/null @@ -1,70 +0,0 @@ -## 概述 - -本样例基于AscendC自定义[Add](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)算子,开发了ATB插件并进行了插件调用测试. - -## 项目结构介绍 -``` - -├── AddAscendC //Add AscendC自定义算子工程 - -├── AddOperationATBPlugin //AddOperation ATB插件代码 - -├── AddOperationTest //AddOperation 测试代码 -``` -## 样例运行 - -### Add AscendC自定义算子部署 - -参照cann-ops仓[add_custom算子](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)" **算子包编译部署** "章节 - -### AddOperation ATB插件部署 - -- 运行编译脚本完成部署(脚本会生成静态库.a文件,同时将头文件拷贝到/usr/include,.a文件拷贝到/usr/local/lib下) - - ``` - cd AddOperationATBPlugin - bash build.sh - ``` - -### AddOperation测试 - -- 运行脚本完成算子测试 - - ```shell - cd AddOperationTest - bash run.sh - ``` - -## AddOperation算子介绍 - -### 功能 - -实现两个输入张量相加 - -### 定义 - -``` -struct AddParam { -}; -``` - -### 参数列表 - -该算子参数为空 - -### 输入 - -| **参数** | **维度** | **数据类型** | **格式** | 描述 | -| -------- | -------------------------- | --------------------- | -------- | ---------- | -| x | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | -| y | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | - -### 输出 - -| **参数** | **维度** | **数据类型** | **格式** | 描述 | -| -------- | -------------------------- | --------------------- | -------- | ---------------------------------------- | -| output | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输出tensor。数据类型和shape与x保持一致。 | - -### 规格约束 - -暂无 \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/EyeCustom/README.md b/operator/ascendc/4_best_practices/25_Eye_custom/EyeCustom/README.md deleted file mode 100644 index 25c5ffc7e..000000000 --- a/operator/ascendc/4_best_practices/25_Eye_custom/EyeCustom/README.md +++ /dev/null @@ -1,3 +0,0 @@ -### 概述 - -具体算子工程,编译部署请见[AddCustom](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom) \ No newline at end of file diff --git a/operator/ascendc/4_best_practices/25_Eye_custom/README.md b/operator/ascendc/4_best_practices/25_Eye_custom/README.md deleted file mode 100644 index 16780e8fa..000000000 --- a/operator/ascendc/4_best_practices/25_Eye_custom/README.md +++ /dev/null @@ -1,49 +0,0 @@ -### 概述 - -本样例介基于AddCustom算子工程,介绍了单算子工程以及ATB插件调用方法 - -### 目录结构介绍 - -``` -├── 24_add_custom // 使用框架调用的方式调用AllGatherMatmul自定义算子 -│ ├── ATBInvocation // 通过ATB调用的方式调用 -│ ├── AddCustom // AddCustom算子工程,引用cann-ops仓的AddCustom工程 -``` - -### 算子描述 - -`AddCustom`算子返回两个数据相加的结果。 - -### 算子规格描述 - -| 算子类型(OpType) | AddCustom | | | | -| ---------------- | ---------- | -------- | --------- | ------ | -| 算子输入 | name | shape | data type | format | -| x | 8 * 2048 | float16 | ND | | -| y | 8 * 2048 | float16 | ND | | -| 算子输出 | z | 8 * 2048 | float16 | ND | -| 核函数名 | add_custom | | | | - -### 支持的产品型号 - -本样例支持如下产品型号: - -- Atlas 训练系列产品 -- Atlas 推理系列产品 -- Atlas A2训练系列产品 -- Atlas 800I A2推理产品 -- Atlas 200I/500 A2推理产品 - -### 算子编译部署 - -- [AddCustom编译部署(需要下载cann-ops仓)](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom) - -### 算子ATB插件调用 - -- [AddOperation编译调用](https://gitee.com/youmoxiao/samples/tree/atb_op/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation) - -## 更新说明 - -| 时间 | 更新事项 | -| -------- | -------- | -| 2025/4/1 | 新增样例 | \ No newline at end of file -- Gitee From d9e97ace980d0221ae21da40c3aadc1bb164cc75 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Tue, 1 Apr 2025 12:53:26 +0000 Subject: [PATCH 18/26] update operator/ascendc/4_best_practices/25_eye/README.md. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: 万祖涛 --- .../ascendc/4_best_practices/25_eye/README.md | 45 ++++++++++++++----- 1 file changed, 33 insertions(+), 12 deletions(-) diff --git a/operator/ascendc/4_best_practices/25_eye/README.md b/operator/ascendc/4_best_practices/25_eye/README.md index 16780e8fa..30ef01ac2 100644 --- a/operator/ascendc/4_best_practices/25_eye/README.md +++ b/operator/ascendc/4_best_practices/25_eye/README.md @@ -5,24 +5,45 @@ ### 目录结构介绍 ``` -├── 24_add_custom // 使用框架调用的方式调用AllGatherMatmul自定义算子 +├── 24_eye // 使用框架调用的方式调用AllGatherMatmul自定义算子 │ ├── ATBInvocation // 通过ATB调用的方式调用 -│ ├── AddCustom // AddCustom算子工程,引用cann-ops仓的AddCustom工程 +│ ├── Eye // AddCustom算子工程,引用cann-ops仓的AddCustom工程 ``` ### 算子描述 -`AddCustom`算子返回两个数据相加的结果。 +创建一个二维矩阵 m×nm×n ,对角元素全为1,其它元素都为0 ### 算子规格描述 -| 算子类型(OpType) | AddCustom | | | | -| ---------------- | ---------- | -------- | --------- | ------ | -| 算子输入 | name | shape | data type | format | -| x | 8 * 2048 | float16 | ND | | -| y | 8 * 2048 | float16 | ND | | -| 算子输出 | z | 8 * 2048 | float16 | ND | -| 核函数名 | add_custom | | | | + + + + + + + + + + + + + + + + + + + + + + + + + + +
算子类型(OpType)Eye
nameshapedata typeformat默认值
算子输入y-float32, float16, int32ND\
算子输出y-float32, float16, int32ND\
attr属性num_rows\int\\
num_columns\int\0
batch_shape\list_int\{1}
dtype\int\0
核函数名eye
+ ### 支持的产品型号 @@ -36,11 +57,11 @@ ### 算子编译部署 -- [AddCustom编译部署(需要下载cann-ops仓)](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom) +- [Eye编译部署(需要下载cann-ops仓)](https://gitee.com/ascend/cann-ops/tree/master/src/math/eye) ### 算子ATB插件调用 -- [AddOperation编译调用](https://gitee.com/youmoxiao/samples/tree/atb_op/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation) +- [EyeOperation编译调用](https://gitee.com/youmoxiao/samples/tree/atb_op/operator/ascendc/4_best_practices/25_eye/ATBInvocation) ## 更新说明 -- Gitee From b4bf4edaaf028d2ef0dfa37b04e1282bf52cc1fa Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Tue, 1 Apr 2025 13:01:40 +0000 Subject: [PATCH 19/26] update operator/ascendc/4_best_practices/25_eye/ATBInvocation/readme.md. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: 万祖涛 --- .../25_eye/ATBInvocation/readme.md | 55 +++++++++++-------- 1 file changed, 33 insertions(+), 22 deletions(-) diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/readme.md b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/readme.md index e0a38627c..8a9d273f5 100644 --- a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/readme.md +++ b/operator/ascendc/4_best_practices/25_eye/ATBInvocation/readme.md @@ -1,41 +1,40 @@ ## 概述 -本样例基于AscendC自定义[Add](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)算子,开发了ATB插件并进行了插件调用测试. +本样例基于AscendC自定义[Eye](https://gitee.com/ascend/cann-ops/tree/master/src/math/eye)算子,开发了ATB插件并进行了插件调用测试. ## 项目结构介绍 -``` - -├── AddAscendC //Add AscendC自定义算子工程 -├── AddOperationATBPlugin //AddOperation ATB插件代码 +``` +├── EyeOperationATBPlugin //EyeOperation ATB插件代码 -├── AddOperationTest //AddOperation 测试代码 +├── EyeOperationTest //EyeOperation 测试代码 ``` + ## 样例运行 -### Add AscendC自定义算子部署 +### Eye AscendC自定义算子部署 -参照cann-ops仓[add_custom算子](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)" **算子包编译部署** "章节 +参照cann-ops仓[eye算子](https://gitee.com/ascend/cann-ops/tree/master/src/math/eye)" **算子包编译部署** "章节 -### AddOperation ATB插件部署 +### EyeOperation ATB插件部署 - 运行编译脚本完成部署(脚本会生成静态库.a文件,同时将头文件拷贝到/usr/include,.a文件拷贝到/usr/local/lib下) ``` - cd AddOperationATBPlugin + cd EyeOperationATBPlugin bash build.sh ``` -### AddOperation测试 +### EyeOperation测试 - 运行脚本完成算子测试 ```shell - cd AddOperationTest + cd EyeOperationTest bash run.sh ``` -## AddOperation算子介绍 +## EyeOperation算子介绍 ### 功能 @@ -44,26 +43,38 @@ ### 定义 ``` -struct AddParam { +struct EyeAttrParam +{ + uint64_t num_rows; + uint64_t num_columns = 0; + std::vector batchShape = {1}; + aclIntArray* batch_shape = aclCreateIntArray(batchShape.data(),batchShape.size()); + uint64_t dtype = 0; }; ``` ### 参数列表 -该算子参数为空 +| **成员名称** | 类型 | 默认值 | 取值范围 | **描述** | 是否必选 | +| ------------ | ------------ | ------ | -------- | ------------------------- | -------- | +| num_rows | uint64_t | / | / | 生成的矩阵的行数 | 是 | +| num_columns | uint64_t | 0 | / | 生成的矩阵的列数 | 是 | +| batch_shape | aclIntArray* | {1} | - | | 是 | +| dtype | uint64_t | 0 | 0,1 | 0表示float32,1表示float16 | 是 | + + ### 输入 -| **参数** | **维度** | **数据类型** | **格式** | 描述 | -| -------- | -------------------------- | --------------------- | -------- | ---------- | -| x | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | -| y | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | +| **参数** | **维度** | **数据类型** | **格式** | 描述 | +| -------- | -------------------------- | --------------- | -------- | ---------------------------------------- | +| y | [dim_0,dim_1,...,dim_n] | float16/float32 | ND | 输出tensor。数据类型和shape与x保持一致。 | ### 输出 -| **参数** | **维度** | **数据类型** | **格式** | 描述 | -| -------- | -------------------------- | --------------------- | -------- | ---------------------------------------- | -| output | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输出tensor。数据类型和shape与x保持一致。 | +| **参数** | **维度** | **数据类型** | **格式** | 描述 | +| -------- | -------------------------- | --------------- | -------- | ---------------------------------------- | +| y | [dim_0,dim_1,...,dim_n] | float16/float32 | ND | 输出tensor。数据类型和shape与x保持一致。 | ### 规格约束 -- Gitee From 87610adb072c1f8ae685e95ad1abdda926df5094 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Tue, 1 Apr 2025 13:11:57 +0000 Subject: [PATCH 20/26] update operator/ascendc/4_best_practices/25_eye/EyeCustom/README.md. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: 万祖涛 --- operator/ascendc/4_best_practices/25_eye/EyeCustom/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/operator/ascendc/4_best_practices/25_eye/EyeCustom/README.md b/operator/ascendc/4_best_practices/25_eye/EyeCustom/README.md index 25c5ffc7e..140123bcb 100644 --- a/operator/ascendc/4_best_practices/25_eye/EyeCustom/README.md +++ b/operator/ascendc/4_best_practices/25_eye/EyeCustom/README.md @@ -1,3 +1,3 @@ ### 概述 -具体算子工程,编译部署请见[AddCustom](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom) \ No newline at end of file +具体算子工程,编译部署请见[Eye](https://gitee.com/ascend/cann-ops/tree/master/src/math/eye) \ No newline at end of file -- Gitee From 00e4e31256a2ef6d58508375476a10d165015018 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Tue, 1 Apr 2025 13:22:33 +0000 Subject: [PATCH 21/26] update operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: 万祖涛 --- .../4_best_practices/24_add_custom/ATBInvocation/readme.md | 2 -- 1 file changed, 2 deletions(-) diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md index e0a38627c..c2c4d74a2 100644 --- a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md @@ -5,8 +5,6 @@ ## 项目结构介绍 ``` -├── AddAscendC //Add AscendC自定义算子工程 - ├── AddOperationATBPlugin //AddOperation ATB插件代码 ├── AddOperationTest //AddOperation 测试代码 -- Gitee From a439eb228f48e8614d3c29714134dd44135d3ed6 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E5=88=98=E5=96=9C=E5=BC=BA?= Date: Wed, 2 Apr 2025 02:48:31 +0000 Subject: [PATCH 22/26] update --- .../ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt | 0 .../ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp | 0 .../ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h | 0 .../ATBInvocation/EyeOperationATBPlugin/build.sh | 0 .../ATBInvocation/EyeOperationTest/CMakeLists.txt | 0 .../ATBInvocation/EyeOperationTest/script/gen_data.py | 0 .../ATBInvocation/EyeOperationTest/script/run.sh | 0 .../ATBInvocation/EyeOperationTest/script/verify_result.py | 0 .../ATBInvocation/EyeOperationTest/src/main.cpp | 0 .../{25_eye => 26_eye}/ATBInvocation/EyeOperationTest/src/main.h | 0 .../4_best_practices/{25_eye => 26_eye}/ATBInvocation/readme.md | 0 .../4_best_practices/{25_eye => 26_eye}/EyeCustom/README.md | 0 operator/ascendc/4_best_practices/{25_eye => 26_eye}/README.md | 0 13 files changed, 0 insertions(+), 0 deletions(-) rename operator/ascendc/4_best_practices/{25_eye => 26_eye}/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt (100%) rename operator/ascendc/4_best_practices/{25_eye => 26_eye}/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp (100%) rename operator/ascendc/4_best_practices/{25_eye => 26_eye}/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h (100%) rename operator/ascendc/4_best_practices/{25_eye => 26_eye}/ATBInvocation/EyeOperationATBPlugin/build.sh (100%) rename operator/ascendc/4_best_practices/{25_eye => 26_eye}/ATBInvocation/EyeOperationTest/CMakeLists.txt (100%) rename operator/ascendc/4_best_practices/{25_eye => 26_eye}/ATBInvocation/EyeOperationTest/script/gen_data.py (100%) rename operator/ascendc/4_best_practices/{25_eye => 26_eye}/ATBInvocation/EyeOperationTest/script/run.sh (100%) rename operator/ascendc/4_best_practices/{25_eye => 26_eye}/ATBInvocation/EyeOperationTest/script/verify_result.py (100%) rename operator/ascendc/4_best_practices/{25_eye => 26_eye}/ATBInvocation/EyeOperationTest/src/main.cpp (100%) rename operator/ascendc/4_best_practices/{25_eye => 26_eye}/ATBInvocation/EyeOperationTest/src/main.h (100%) rename operator/ascendc/4_best_practices/{25_eye => 26_eye}/ATBInvocation/readme.md (100%) rename operator/ascendc/4_best_practices/{25_eye => 26_eye}/EyeCustom/README.md (100%) rename operator/ascendc/4_best_practices/{25_eye => 26_eye}/README.md (100%) diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt b/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt similarity index 100% rename from operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt rename to operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationATBPlugin/CMakeLists.txt diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp b/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp similarity index 100% rename from operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp rename to operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.cpp diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h b/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h similarity index 100% rename from operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h rename to operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationATBPlugin/aclnn_eye_operation.h diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/build.sh b/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationATBPlugin/build.sh similarity index 100% rename from operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationATBPlugin/build.sh rename to operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationATBPlugin/build.sh diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/CMakeLists.txt b/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/CMakeLists.txt similarity index 100% rename from operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/CMakeLists.txt rename to operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/CMakeLists.txt diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/gen_data.py b/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/script/gen_data.py similarity index 100% rename from operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/gen_data.py rename to operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/script/gen_data.py diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/run.sh b/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/script/run.sh similarity index 100% rename from operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/run.sh rename to operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/script/run.sh diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/verify_result.py b/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/script/verify_result.py similarity index 100% rename from operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/script/verify_result.py rename to operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/script/verify_result.py diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/src/main.cpp b/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/src/main.cpp similarity index 100% rename from operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/src/main.cpp rename to operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/src/main.cpp diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/src/main.h b/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/src/main.h similarity index 100% rename from operator/ascendc/4_best_practices/25_eye/ATBInvocation/EyeOperationTest/src/main.h rename to operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/src/main.h diff --git a/operator/ascendc/4_best_practices/25_eye/ATBInvocation/readme.md b/operator/ascendc/4_best_practices/26_eye/ATBInvocation/readme.md similarity index 100% rename from operator/ascendc/4_best_practices/25_eye/ATBInvocation/readme.md rename to operator/ascendc/4_best_practices/26_eye/ATBInvocation/readme.md diff --git a/operator/ascendc/4_best_practices/25_eye/EyeCustom/README.md b/operator/ascendc/4_best_practices/26_eye/EyeCustom/README.md similarity index 100% rename from operator/ascendc/4_best_practices/25_eye/EyeCustom/README.md rename to operator/ascendc/4_best_practices/26_eye/EyeCustom/README.md diff --git a/operator/ascendc/4_best_practices/25_eye/README.md b/operator/ascendc/4_best_practices/26_eye/README.md similarity index 100% rename from operator/ascendc/4_best_practices/25_eye/README.md rename to operator/ascendc/4_best_practices/26_eye/README.md -- Gitee From 2d598d98d7517e0acd99cb0d6320e6f5bf702f66 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Wed, 2 Apr 2025 02:50:24 +0000 Subject: [PATCH 23/26] =?UTF-8?q?=E5=88=A0=E9=99=A4=E6=96=87=E4=BB=B6=20at?= =?UTF-8?q?b=5Fplugin?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .../AddOperationATBPlugin/CMakeLists.txt | 21 -- .../aclnn_add_operation.cpp | 142 ------------ .../aclnn_add_operation.h | 56 ----- .../AddOperationATBPlugin/build.sh | 33 --- .../AddOperationTest/CMakeLists.txt | 40 ---- .../AddOperationTest/script/gen_data.py | 20 -- .../AddOperationTest/script/run.sh | 52 ----- .../AddOperationTest/script/verify_result.py | 28 --- .../AddOperationTest/src/main.cpp | 217 ------------------ .../AtbAddSample/AddOperationTest/src/main.h | 55 ----- atb_plugin/AtbAddSample/readme.md | 70 ------ atb_plugin/README.md | 6 - 12 files changed, 740 deletions(-) delete mode 100644 atb_plugin/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt delete mode 100644 atb_plugin/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp delete mode 100644 atb_plugin/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h delete mode 100755 atb_plugin/AtbAddSample/AddOperationATBPlugin/build.sh delete mode 100644 atb_plugin/AtbAddSample/AddOperationTest/CMakeLists.txt delete mode 100644 atb_plugin/AtbAddSample/AddOperationTest/script/gen_data.py delete mode 100644 atb_plugin/AtbAddSample/AddOperationTest/script/run.sh delete mode 100644 atb_plugin/AtbAddSample/AddOperationTest/script/verify_result.py delete mode 100644 atb_plugin/AtbAddSample/AddOperationTest/src/main.cpp delete mode 100644 atb_plugin/AtbAddSample/AddOperationTest/src/main.h delete mode 100644 atb_plugin/AtbAddSample/readme.md delete mode 100644 atb_plugin/README.md diff --git a/atb_plugin/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt b/atb_plugin/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt deleted file mode 100644 index c89578fcd..000000000 --- a/atb_plugin/AtbAddSample/AddOperationATBPlugin/CMakeLists.txt +++ /dev/null @@ -1,21 +0,0 @@ -# Copyright 2024 Tencent Inc. All rights reserved. -# -# ============================================================================== -cmake_minimum_required(VERSION 3.14) -set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") -file(GLOB_RECURSE atb_plugin_operations_SRCS "./*.cpp") -set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") -message(STATUS "atb_plugin_operations_SRCS: ${atb_plugin_operations_SRCS}") -include_directories( - $ENV{ATB_HOME_PATH}/include - $ENV{ASCEND_HOME_PATH}/include - ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 - ${CMAKE_CURRENT_SOURCE_DIR} - ) -link_directories( - $ENV{ATB_HOME_PATH}/lib - $ENV{ASCEND_HOME_PATH}/lib64 - ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 - ) -add_library(atb_plugin_operations STATIC ${atb_plugin_operations_SRCS}) -target_link_libraries(atb_plugin_operations PRIVATE ${ACL_SHARED_LIBS} ${ACL_SHARED_LIBS} ascendcl) \ No newline at end of file diff --git a/atb_plugin/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp b/atb_plugin/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp deleted file mode 100644 index 468b84aff..000000000 --- a/atb_plugin/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.cpp +++ /dev/null @@ -1,142 +0,0 @@ -#include "aclnn_add_operation.h" -#include "aclnn_add_custom.h" - -AddOperation::AddOperation(const std::string &name, AddAttrParam param){ - attrParam = param; - opName_ = name; -} - -atb::SVector GetCopyTensorStride(atb::Dims &tensorDims) -{ - atb::SVector tmpStrides(tensorDims.dimNum, 1); - if (tensorDims.dimNum > 8) { // 8: tensor最大维度数量 - printf("tensor's dimNum is larger than 8, GetCopyTensorStride failed."); - return tmpStrides; - } - for (int64_t i = static_cast(tensorDims.dimNum) - 2; i >= 0; i--) { - tmpStrides[i] = (tensorDims.dims[i + 1] * tmpStrides[i + 1]); - } - return tmpStrides; -} - -std::shared_ptr AddOperation::CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx) -{ - auto aclnnTensor = std::make_shared(); - aclnnTensor->tensorIdx = static_cast(tensorIdx); - aclnnTensor->needUpdateTensorDataPtr = true; - aclnnTensor->atbTensor = atbTensor; - aclnnTensor->strides = GetCopyTensorStride(atbTensor.desc.shape); - - // 创建Aclnn tensor - aclnnTensor->tensor = aclCreateTensor(atbTensor.desc.shape.dims, - atbTensor.desc.shape.dimNum, - atbTensor.desc.dtype, - aclnnTensor->strides.data(), - 0, - atbTensor.desc.format, - atbTensor.desc.shape.dims, - atbTensor.desc.shape.dimNum, - atbTensor.deviceData); - return aclnnTensor; -} - -atb::Status AddOperation::UpdateAclnnVariantPack(const atb::VariantPack &variantPack) -{ - // 更新inTensor的device地址 - for (size_t i = 0; i < aclInTensors_.size(); ++i) { - int ret = -1; - if (!aclInTensors_[i]->needUpdateTensorDataPtr) { - continue; - } - aclInTensors_[i]->atbTensor = variantPack.inTensors.at(i); - ret = aclSetInputTensorAddr(aclExecutor_, - aclInTensors_[i]->tensorIdx, - aclInTensors_[i]->tensor, - aclInTensors_[i]->atbTensor.deviceData); - - if (ret != 0) { - printf("set input fail"); - return atb::ERROR_CANN_ERROR; - } - } - - // 更新outTensor的device地址 - for (size_t i = 0; i < aclOutTensors_.size(); ++i) { - int ret = -1; - if (!aclOutTensors_[i]->needUpdateTensorDataPtr) { - continue; - } - aclOutTensors_[i]->atbTensor = variantPack.outTensors.at(i); - ret = aclSetOutputTensorAddr(aclExecutor_, - aclOutTensors_[i]->tensorIdx, - aclOutTensors_[i]->tensor, - aclOutTensors_[i]->atbTensor.deviceData); - - if (ret != 0) { - printf("set output fail"); - return atb::ERROR_CANN_ERROR; - } - } - return atb::NO_ERROR; -} - -atb::Status AddOperation::Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) { - - aclInTensors_.resize(GetInputNum()); - for (size_t i = 0; i < aclInTensors_.size(); ++i) { - auto aclnnTensor = CreateAclnnTensor(variantPack.inTensors.at(i), i); - if (aclnnTensor->tensor == nullptr) { - printf("creat input tensor %ld fail", i); - return atb::ERROR_INTERNAL_ERROR; - } - aclInTensors_[i] = aclnnTensor; - } - - aclOutTensors_.resize(GetOutputNum()); - for (size_t i = 0; i < aclOutTensors_.size(); ++i) { - auto aclnnTensor = CreateAclnnTensor(variantPack.outTensors.at(i), i); - if (aclnnTensor->tensor == nullptr) { - printf("creat output tensor %ld fail", i); - return atb::ERROR_INTERNAL_ERROR; - } - aclOutTensors_[i] = aclnnTensor; - } - - - auto ret = aclnnAddCustomGetWorkspaceSize(aclInTensors_.at(0)->tensor, - aclInTensors_.at(1)->tensor, - aclOutTensors_.at(0)->tensor, - &workspaceSize_, - &aclExecutor_); - - workspaceSize = workspaceSize_; - return ret; - -} - -atb::Status AddOperation::Execute(const atb::VariantPack &variantPack, uint8_t *workspace, uint64_t workspaceSize, atb::Context *context) { - - - - aclrtStream stream = context->GetExecuteStream(); - if (!stream) { - printf("get stream fail"); - return atb::ERROR_INVALID_PARAM; - } - // 更新数据传入的地址 - int ret = UpdateAclnnVariantPack(variantPack); - if (ret != 0) { - printf("UpdateAclnnVariantPack fail"); - return atb::ERROR_CANN_ERROR; - } - ret = aclnnAddCustom(workspace, workspaceSize_, aclExecutor_, stream); - - return ret; -} - -atb::Status AddOperation::InferShape( - const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const -{ - outTensorDesc.at(0) = inTensorDesc.at(0); - return atb::NO_ERROR; -} diff --git a/atb_plugin/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h b/atb_plugin/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h deleted file mode 100644 index 8670088d3..000000000 --- a/atb_plugin/AtbAddSample/AddOperationATBPlugin/aclnn_add_operation.h +++ /dev/null @@ -1,56 +0,0 @@ -#include -#include -#include -#include -#include -#include "atb/infer_op_params.h" - - -struct AddAttrParam -{ - // add没属性,此处空 -}; - -struct AclnnTensor -{ -public: - atb::Tensor atbTensor; // - aclTensor *tensor = nullptr; - int tensorIdx = -1; // aclTensor在aclExecutor中的index - bool needUpdateTensorDataPtr = false; - atb::SVector strides = {}; -}; - -class AddOperation: public atb::Operation{ -public: - AddOperation(const std::string &name, AddAttrParam param); - atb::Status Setup(const atb::VariantPack &variantPack, uint64_t &workspaceSize, atb::Context *context) override; - atb::Status Execute(const atb::VariantPack &variantPack, uint8_t *workspace, - uint64_t workspaceSize, atb::Context *context) override; - atb::Status InferShape( - const atb::SVector &inTensorDesc, atb::SVector &outTensorDesc) const; - std::shared_ptr CreateAclnnTensor(atb::Tensor atbTensor, size_t tensorIdx); - atb::Status UpdateAclnnVariantPack(const atb::VariantPack &variantPack); - - uint32_t GetInputNum() const - { - return 2; // 算子入参个数 - } - - uint32_t GetOutputNum() const - { - return 1; // 算子出参个数 - } - std::string GetName() const - { - return opName_; - } - - aclOpExecutor *aclExecutor_ = nullptr; - AddAttrParam attrParam; - std::string opName_; - uint64_t workspaceSize_; - - atb::SVector> aclInTensors_; - atb::SVector> aclOutTensors_; -}; \ No newline at end of file diff --git a/atb_plugin/AtbAddSample/AddOperationATBPlugin/build.sh b/atb_plugin/AtbAddSample/AddOperationATBPlugin/build.sh deleted file mode 100755 index ba28cc143..000000000 --- a/atb_plugin/AtbAddSample/AddOperationATBPlugin/build.sh +++ /dev/null @@ -1,33 +0,0 @@ -#!/bin/bash - -# 定义构建目录 -BUILD_DIR="build" - -# 创建构建目录 -mkdir -p "$BUILD_DIR" -cd "$BUILD_DIR" - -# 运行 CMake 配置和编译 -cmake .. -make - -# 查找生成的 .a 文件 -A_FILE=$(find . -name "*.a" -type f) - -# 检查是否找到了 .a 文件 -if [ -z "$A_FILE" ]; then - echo "未找到 .a 文件,编译可能失败。" - exit 1 -fi - -# 复制头文件到 /usr/include -HEADER_FILES=$(find .. -name "*.h" -type f) -for header in $HEADER_FILES; do - cp "$header" /usr/include/ -done - -# 复制 .a 文件到 /usr/local/lib -cp "$A_FILE" /usr/local/lib/ - -echo "构建完成,头文件和 .a 文件已复制到目标目录。" - diff --git a/atb_plugin/AtbAddSample/AddOperationTest/CMakeLists.txt b/atb_plugin/AtbAddSample/AddOperationTest/CMakeLists.txt deleted file mode 100644 index b9c2cec46..000000000 --- a/atb_plugin/AtbAddSample/AddOperationTest/CMakeLists.txt +++ /dev/null @@ -1,40 +0,0 @@ -cmake_minimum_required(VERSION 3.16) -project("test_model") -option(USE_CXX11_ABI "USE_CXX11_ABI" OFF) -set(CMAKE_BUILD_TYPE Debug) -if(USE_CXX11_ABI) - set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=1") -else() - set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -D_GLIBCXX_USE_CXX11_ABI=0") -endif() - -set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -std=c++17") - -set(CUST_PKG_PATH "$ENV{ASCEND_HOME_PATH}/opp/vendors/customize/op_api") - -include_directories( - $ENV{ATB_HOME_PATH}/include - $ENV{ASCEND_HOME_PATH}/include - ${CUST_PKG_PATH}/include # 添加自定义算子头文件路径 - ${CMAKE_CURRENT_SOURCE_DIR} - ./src/ - ) - -message(".h path = ${CUST_PKG_PATH}/include") -link_directories( - $ENV{ATB_HOME_PATH}/lib - $ENV{ASCEND_HOME_PATH}/lib64 - ${CUST_PKG_PATH}/lib # 添加自定义算子so路径 - ) - - -link_libraries(atb_plugin_operations) -add_executable(test_model ./src/main.cpp) - -# 列出所有的头文件目录 -target_link_libraries(test_model PRIVATE atb ascendcl cust_opapi opapi nnopbase pthread) # 添加自定义算子so及适配so - - - - - diff --git a/atb_plugin/AtbAddSample/AddOperationTest/script/gen_data.py b/atb_plugin/AtbAddSample/AddOperationTest/script/gen_data.py deleted file mode 100644 index 83ecafbce..000000000 --- a/atb_plugin/AtbAddSample/AddOperationTest/script/gen_data.py +++ /dev/null @@ -1,20 +0,0 @@ -import torch -import os -def gen_golden_data_simple(): - dtype = "float" - input1 = torch.randn(133, 4095, dtype=torch.float) - input2 = torch.randn(133, 4095, dtype=torch.float) - - - golden = input1 + input2 - os.system("mkdir -p input") - os.system("mkdir -p output") - input1.numpy().tofile('./script/input/input0.bin') - input2.numpy().tofile('./script/input/input1.bin') - golden.numpy().tofile("./script/output/golden0.bin") - - with open("./output/meta", "w") as fp: - print(dtype, file=fp) - -if __name__ == "__main__": - gen_golden_data_simple() diff --git a/atb_plugin/AtbAddSample/AddOperationTest/script/run.sh b/atb_plugin/AtbAddSample/AddOperationTest/script/run.sh deleted file mode 100644 index 450b717f2..000000000 --- a/atb_plugin/AtbAddSample/AddOperationTest/script/run.sh +++ /dev/null @@ -1,52 +0,0 @@ -#!/bin/bash - - -# 步骤1: 运行gen_data.py生成输入bin文件和golden标杆输出数据 -echo "正在生成输入数据和golden标杆数据..." -mkdir -p script/input -mkdir -p script/output -python3 script/gen_data.py -if [ $? -ne 0 ]; then - echo "生成数据失败,脚本终止。" - exit 1 -fi - -# 步骤2: 创建构建目录并进入 -mkdir -p build -cd build -if [ $? -ne 0 ]; then - echo "无法进入构建目录,脚本终止。" - exit 1 -fi - -# 步骤3: 使用CMake配置项目 -echo "正在配置CMake项目..." -cmake .. -if [ $? -ne 0 ]; then - echo "CMake配置失败,脚本终止。" - exit 1 -fi - -# 步骤4: 编译代码 -echo "正在编译代码..." -make -if [ $? -ne 0 ]; then - echo "编译失败,脚本终止。" - exit 1 -fi - -mv test_model ../ -cd .. - -# 步骤5: 运行可执行文件生成实际输出文件 -echo "正在运行可执行文件生成实际输出..." -./test_model -if [ $? -ne 0 ]; then - echo "运行可执行文件失败,脚本终止。" - exit 1 -fi - -# 步骤6: 调用verify_result.py进行golden标杆数据和实际输出数据的比对 -echo "正在验证结果..." -python3 script/verify_result.py script/output/output_0.bin script/output/golden0.bin - diff --git a/atb_plugin/AtbAddSample/AddOperationTest/script/verify_result.py b/atb_plugin/AtbAddSample/AddOperationTest/script/verify_result.py deleted file mode 100644 index 629845e13..000000000 --- a/atb_plugin/AtbAddSample/AddOperationTest/script/verify_result.py +++ /dev/null @@ -1,28 +0,0 @@ -import os -import sys -import numpy - -loss = 1e-3 # 容忍偏差,一般fp16要求绝对误差和相对误差均不超过千分之一 -minimum = 10e-10 - -def verify_result(real_result, golden): - with open("output/meta", "r") as fp: - dtype_str = fp.readline().strip() - dtype = eval(dtype_str) - real_result = numpy.fromfile(real_result, dtype=dtype) # 从bin文件读取实际运算结果 - golden = numpy.fromfile(golden, dtype=dtype) # 从bin文件读取预期运算结果 - print("=" * 50, real_result[:5], golden[:5], "=" * 50, sep='\n', end='\n', file=sys.stderr) - result = numpy.abs(real_result - golden) # 计算运算结果和预期结果偏差 - deno = numpy.maximum(numpy.abs(real_result), numpy.abs(golden)) # 获取最大值并组成新数组 - result_atol = numpy.less_equal(result, loss) # 计算绝对误差 - result_rtol = numpy.less_equal(result / numpy.add(deno, minimum), loss) # 计算相对误差 - if not result_rtol.all() and not result_atol.all(): - if numpy.sum(result_rtol == False) > real_result.size * loss and numpy.sum(result_atol == False) > real_result.size * loss: # 误差超出预期时返回打印错误,返回对比失败 - print("[ERROR] result error") - return False - print("test Operation success!") - return True - - -if __name__ == '__main__': - verify_result(sys.argv[1],sys.argv[2]) diff --git a/atb_plugin/AtbAddSample/AddOperationTest/src/main.cpp b/atb_plugin/AtbAddSample/AddOperationTest/src/main.cpp deleted file mode 100644 index 258c0cb3c..000000000 --- a/atb_plugin/AtbAddSample/AddOperationTest/src/main.cpp +++ /dev/null @@ -1,217 +0,0 @@ -/* - * Copyright (c) Huawei Technologies Co., Ltd. 2023. All rights reserved. - * - * Licensed under the Apache License, Version 2.0 (the "License"); - * you may not use this file except in compliance with the License. - * You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ -#include "main.h" - -bool SetInputData(std::vector &inputData){ - char *xPath = "./script/input/input0.bin"; - char *yPath = "./script/input/input1.bin"; - InputData inputX; - InputData inputY; - inputX.data = ReadBinFile(xPath,inputX.size); - inputY.data = ReadBinFile(yPath,inputY.size); - inputData.push_back(inputX); - inputData.push_back(inputY); - return true; -} - -bool SetOperationInputDesc(atb::SVector &intensorDescs){ - atb::TensorDesc xDesc; - xDesc.dtype = ACL_FLOAT; - xDesc.format = ACL_FORMAT_ND; - xDesc.shape.dimNum = 2; - xDesc.shape.dims[0] = 133; - xDesc.shape.dims[1] = 4095; - - atb::TensorDesc yDesc; - yDesc.dtype = ACL_FLOAT; - yDesc.format = ACL_FORMAT_ND; - yDesc.shape.dimNum = 2; - yDesc.shape.dims[0] = 133; - yDesc.shape.dims[1] = 4095; - - intensorDescs.at(0) = xDesc; - intensorDescs.at(1) = yDesc; -} - - - -static void SetCurrentDevice() -{ - const int deviceId = 0; - std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; - int ret = aclrtSetDevice(deviceId); - if (ret != 0) { - std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; - return; - } - std::cout << "[INFO]: aclrtSetDevice success" << std::endl; -} - - -static void FreeTensor(atb::Tensor &tensor) -{ - if (tensor.deviceData) { - int ret = aclrtFree(tensor.deviceData); - if (ret != 0) { - std::cout << "[ERROR]: aclrtFree fail" << std::endl; - } - tensor.deviceData = nullptr; - tensor.dataSize = 0; - } - if (tensor.hostData) { - int ret = aclrtFreeHost(tensor.hostData); - if (ret != 0) { - std::cout << "[ERROR]: aclrtFreeHost fail, ret = " << ret << std::endl; - } - tensor.hostData = nullptr; - tensor.dataSize = 0; - } -} - -static void FreeTensors(atb::SVector &inTensors, atb::SVector &outTensors) -{ - for (size_t i = 0; i < inTensors.size(); ++i) { - FreeTensor(inTensors.at(i)); - } - for (size_t i = 0; i < outTensors.size(); ++i) { - FreeTensor(outTensors.at(i)); - } -} -bool SaveMemoryToBinFile(void* memoryAddress, size_t memorySize, size_t i) { - // 创建 output 目录(如果不存在) - std::filesystem::create_directories("output"); - - // 生成文件名 - std::string filename = "script/output/output_" + std::to_string(i) + ".bin"; - - // 打开文件以二进制写入模式 - std::ofstream file(filename, std::ios::binary); - if (!file) { - std::cerr << "无法打开文件: " << filename << std::endl; - return false; - } - - // 写入数据 - file.write(static_cast(memoryAddress), memorySize); - if (!file) { - std::cerr << "写入文件时出错: " << filename << std::endl; - file.close(); - return false; - } - - // 关闭文件 - file.close(); - std::cout << "数据已成功保存到: " << filename << std::endl; - return true; -} - -int main(int argc, const char *argv[]) -{ - const int deviceId = 0; - std::cout << "[INFO]: aclrtSetDevice " << deviceId << std::endl; - int ret = aclrtSetDevice(deviceId); - if (ret != 0) { - std::cout << "[ERROR]: aclrtSetDevice fail, error:" << ret << std::endl; - return 1; - } - std::cout << "[INFO]: aclrtSetDevice success" << std::endl; - atb::Context *context = nullptr; - ret = atb::CreateContext(&context); - void *stream = nullptr; - ret = aclrtCreateStream(&stream); - if (ret != 0) { - std::cout << "[ERROR]: AsdRtStreamCreate fail, ret:" << ret << std::endl; - return 1; - } - context->SetExecuteStream(stream); - - std::vector input; - SetInputData(input); - - AddAttrParam addAttrParam; - AddOperation *op = new AddOperation("Add",addAttrParam); - std::cout << "[INFO]: complete CreateOp!" << std::endl; - - if(input.size() != op->GetInputNum()) std::cout << "[ERROR]: Operation actual input num is not equal to GetInputNum()"; - - atb::SVector intensorDescs; - atb::SVector outtensorDescs; - intensorDescs.resize(op->GetInputNum()); - outtensorDescs.resize(op->GetOutputNum()); - SetOperationInputDesc(intensorDescs); - atb::Status st = op->InferShape(intensorDescs,outtensorDescs); - if (st != 0) { - std::cout << "[ERROR]: Operation InferShape fail" << std::endl; - return -1; - } - std::cout << "[INFO]: Operation InferShape success" << std::endl; - atb::VariantPack variantPack; - variantPack.inTensors.resize(op->GetInputNum()); - variantPack.outTensors.resize(op->GetOutputNum()); - for(size_t i=0;iGetInputNum();i++){ - variantPack.inTensors.at(i).desc = intensorDescs.at(i); - variantPack.inTensors.at(i).hostData = input[i].data; - variantPack.inTensors.at(i).dataSize = input[i].size; - CheckAcl(aclrtMalloc(&variantPack.inTensors.at(i).deviceData, input[i].size, ACL_MEM_MALLOC_HUGE_FIRST)); - CheckAcl(aclrtMemcpy(variantPack.inTensors.at(i).deviceData, input[i].size, input[i].data, input[i].size, ACL_MEMCPY_HOST_TO_DEVICE)); - } - std::cout << "[INFO]: Operation Input prepare sucess" << std::endl; - for(size_t i=0;iGetOutputNum();i++){ - int64_t *dims = new int64_t[outtensorDescs.at(i).shape.dimNum]; - for(size_t j=0;jSetup(variantPack, workspaceSize, context); - if (st != 0) { - std::cout << "[ERROR]: Operation setup fail" << std::endl; - return -1; - } - std::cout << "[INFO]: Operation setup success" << std::endl; - void *workspace = nullptr; - if (workspaceSize > 0) { - aclrtMalloc(&workspace, workspaceSize, ACL_MEM_MALLOC_HUGE_FIRST); - } - - std::cout << "[INFO]: Operation execute start" << std::endl; - st = op->Execute(variantPack, (uint8_t*)workspace, workspaceSize, context); - if (st != 0) { - std::cout << "[ERROR]: Operation execute fail" << std::endl; - return -1; - } - std::cout << "[INFO]: Operation execute success" << std::endl; - for(size_t i = 0; i < op->GetOutputNum(); i++){ - CheckAcl(aclrtMemcpy(variantPack.outTensors.at(i).hostData, variantPack.outTensors.at(i).dataSize, variantPack.outTensors.at(0).deviceData, - variantPack.outTensors.at(i).dataSize, ACL_MEMCPY_DEVICE_TO_HOST)); - SaveMemoryToBinFile(variantPack.outTensors.at(i).hostData,variantPack.outTensors.at(i).dataSize,i); - } - - FreeTensors(variantPack.inTensors, variantPack.outTensors); - st = atb::DestroyContext(context); - CheckAcl(aclrtDestroyStream(stream)); - CheckAcl(aclrtResetDevice(0)); - CheckAcl(aclFinalize()); - return atb::ErrorType::NO_ERROR; -} diff --git a/atb_plugin/AtbAddSample/AddOperationTest/src/main.h b/atb_plugin/AtbAddSample/AddOperationTest/src/main.h deleted file mode 100644 index 44a94e9ec..000000000 --- a/atb_plugin/AtbAddSample/AddOperationTest/src/main.h +++ /dev/null @@ -1,55 +0,0 @@ -#include -#include -#include -#include -#include -#include -#include -#include "securec.h" -#include "atb/atb_infer.h" -#include "aclnn_add_operation.h" -#include - -struct InputData{ - void* data; - uint64_t size; -}; -aclError CheckAcl(aclError ret) -{ - if (ret != ACL_ERROR_NONE) { - std::cerr << __FILE__ << ":" << __LINE__ << " aclError:" << ret << std::endl; - } - return ret; -} -void* ReadBinFile(const char* filename, size_t& size) { - std::ifstream file(filename, std::ios::binary | std::ios::ate); - if (!file) { - std::cerr << "无法打开文件: " << filename << std::endl; - return nullptr; - } - - // 获取文件大小 - size = file.tellg(); - file.seekg(0, std::ios::beg); - - // 分配内存 - void* buffer; - int ret = aclrtMallocHost(&buffer,size); - if (!buffer) { - std::cerr << "内存分配失败" << std::endl; - file.close(); - return nullptr; - } - - // 读取文件内容到内存 - file.read(static_cast(buffer), size); - if (!file) { - std::cerr << "读取文件失败" << std::endl; - delete[] static_cast(buffer); - file.close(); - return nullptr; - } - - file.close(); - return buffer; -} \ No newline at end of file diff --git a/atb_plugin/AtbAddSample/readme.md b/atb_plugin/AtbAddSample/readme.md deleted file mode 100644 index e0a38627c..000000000 --- a/atb_plugin/AtbAddSample/readme.md +++ /dev/null @@ -1,70 +0,0 @@ -## 概述 - -本样例基于AscendC自定义[Add](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)算子,开发了ATB插件并进行了插件调用测试. - -## 项目结构介绍 -``` - -├── AddAscendC //Add AscendC自定义算子工程 - -├── AddOperationATBPlugin //AddOperation ATB插件代码 - -├── AddOperationTest //AddOperation 测试代码 -``` -## 样例运行 - -### Add AscendC自定义算子部署 - -参照cann-ops仓[add_custom算子](https://gitee.com/ascend/cann-ops/tree/master/src/math/add_custom)" **算子包编译部署** "章节 - -### AddOperation ATB插件部署 - -- 运行编译脚本完成部署(脚本会生成静态库.a文件,同时将头文件拷贝到/usr/include,.a文件拷贝到/usr/local/lib下) - - ``` - cd AddOperationATBPlugin - bash build.sh - ``` - -### AddOperation测试 - -- 运行脚本完成算子测试 - - ```shell - cd AddOperationTest - bash run.sh - ``` - -## AddOperation算子介绍 - -### 功能 - -实现两个输入张量相加 - -### 定义 - -``` -struct AddParam { -}; -``` - -### 参数列表 - -该算子参数为空 - -### 输入 - -| **参数** | **维度** | **数据类型** | **格式** | 描述 | -| -------- | -------------------------- | --------------------- | -------- | ---------- | -| x | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | -| y | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输入tensor | - -### 输出 - -| **参数** | **维度** | **数据类型** | **格式** | 描述 | -| -------- | -------------------------- | --------------------- | -------- | ---------------------------------------- | -| output | [dim_0,dim_1,...,dim_n] | float/half/int8/int32 | ND | 输出tensor。数据类型和shape与x保持一致。 | - -### 规格约束 - -暂无 \ No newline at end of file diff --git a/atb_plugin/README.md b/atb_plugin/README.md deleted file mode 100644 index 6155a664d..000000000 --- a/atb_plugin/README.md +++ /dev/null @@ -1,6 +0,0 @@ - -| 目录名称 | 功能描述 | 运行环境 | 支持CANN版本 | -|------------------------------------------------------------|------------------------------------------| -- |---------------------| -| [AtbAddSample](./AtbAddSample) | Atb调用无属性算子样例 |Atlas A2训练系列产品| 8.0.0.alpha003及以上 | -| [AtbReduceSum](./AtbReduceSum) | Atb调用有属性算子样例 | Atlas A2训练系列产品 | 8.0.0.alpha003及以上 | - -- Gitee From b9d1b9312b60640c92319c17b56c6bd006ac0d7c Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Wed, 2 Apr 2025 08:13:27 +0000 Subject: [PATCH 24/26] update operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: 万祖涛 --- .../4_best_practices/24_add_custom/ATBInvocation/readme.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md index c2c4d74a2..714c8c791 100644 --- a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/readme.md @@ -30,7 +30,7 @@ ```shell cd AddOperationTest - bash run.sh + bash script/run.sh ``` ## AddOperation算子介绍 -- Gitee From b252fc9e68e1234309f326ea28dfe0c57330a2ff Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Wed, 2 Apr 2025 16:21:09 +0000 Subject: [PATCH 25/26] update operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: 万祖涛 --- .../24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp | 1 + 1 file changed, 1 insertion(+) diff --git a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp index 2fca70549..2500dd95f 100644 --- a/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp +++ b/operator/ascendc/4_best_practices/24_add_custom/ATBInvocation/AddOperationTest/src/main.cpp @@ -201,6 +201,7 @@ int main(int argc, const char *argv[]) std::cout << "[ERROR]: Operation execute fail" << std::endl; return -1; } + ret = aclrtSynchronizeStream(stream); std::cout << "[INFO]: Operation execute success" << std::endl; for(size_t i = 0; i < op->GetOutputNum(); i++){ CheckAcl(aclrtMemcpy(variantPack.outTensors.at(i).hostData, variantPack.outTensors.at(i).dataSize, variantPack.outTensors.at(0).deviceData, -- Gitee From dfb258e5cab7022a983e5aec35360c3acecaad0e Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E4=B8=87=E7=A5=96=E6=B6=9B?= Date: Wed, 2 Apr 2025 16:21:46 +0000 Subject: [PATCH 26/26] update operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/src/main.cpp. MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: 万祖涛 --- .../26_eye/ATBInvocation/EyeOperationTest/src/main.cpp | 1 + 1 file changed, 1 insertion(+) diff --git a/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/src/main.cpp b/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/src/main.cpp index 0bd444165..9075d3260 100644 --- a/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/src/main.cpp +++ b/operator/ascendc/4_best_practices/26_eye/ATBInvocation/EyeOperationTest/src/main.cpp @@ -179,6 +179,7 @@ int main(int argc, const char *argv[]) std::cout << "[ERROR]: Operation execute fail" << std::endl; return -1; } + ret = aclrtSynchronizeStream(stream); std::cout << "[INFO]: Operation execute success" << std::endl; CheckAcl(aclrtMemcpy(variantPack.inTensors.at(0).hostData, variantPack.inTensors.at(0).dataSize, variantPack.inTensors.at(0).deviceData, -- Gitee