From bc0b19b740e8c954a3e445fbada03c7c39933f14 Mon Sep 17 00:00:00 2001 From: wangcaoyu Date: Mon, 4 Dec 2023 20:22:58 +0800 Subject: [PATCH] =?UTF-8?q?=E6=B5=8B=E8=AF=95=E7=94=A8=E4=BE=8B=E8=B0=83?= =?UTF-8?q?=E8=AF=95?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: wangcaoyu --- build-tools/capi_parser/.gitignore | 4 +- .../capi_parser/src/typedef/diff/diff.py | 30 +++++- .../capi_parser/test/expect/diff/.gitkeep | 0 .../test/expect/diff/ut_diff_api_add_001.txt | 11 +++ .../expect/diff/ut_diff_api_delete_001.txt | 11 +++ .../test/expect/diff/ut_diff_dir_add_001.txt | 29 ++++++ .../expect/diff/ut_diff_dir_delete_001.txt | 29 ++++++ .../diff/ut_diff_doc_addtogroup_add_001.txt | 11 +++ .../ut_diff_doc_addtogroup_change_001.txt | 11 +++ .../ut_diff_doc_addtogroup_delete_001.txt | 11 +++ .../expect/diff/ut_diff_doc_brief_add_001.txt | 11 +++ .../diff/ut_diff_doc_brief_change_001.txt | 11 +++ .../diff/ut_diff_doc_brief_delete_001.txt | 11 +++ .../diff/ut_diff_doc_deprecated_add_001.txt | 11 +++ .../ut_diff_doc_deprecated_change_001.txt | 11 +++ .../ut_diff_doc_deprecated_change_002.txt | 11 +++ .../ut_diff_doc_deprecated_delete_001.txt | 11 +++ .../expect/diff/ut_diff_doc_file_add_001.txt | 11 +++ .../diff/ut_diff_doc_file_change_001.txt | 11 +++ .../diff/ut_diff_doc_file_delete_001.txt | 11 +++ .../diff/ut_diff_doc_library_add_001.txt | 11 +++ .../diff/ut_diff_doc_library_change_001.txt | 11 +++ .../diff/ut_diff_doc_library_delete_001.txt | 11 +++ .../diff/ut_diff_doc_permission_add_001.txt | 11 +++ .../diff/ut_diff_doc_permission_add_002.txt | 11 +++ .../diff/ut_diff_doc_permission_add_003.txt | 11 +++ .../ut_diff_doc_permission_change_001.txt | 11 +++ .../ut_diff_doc_permission_delete_001.txt | 11 +++ .../ut_diff_doc_permission_delete_002.txt | 11 +++ .../ut_diff_doc_permission_delete_003.txt | 11 +++ .../expect/diff/ut_diff_doc_since_add_001.txt | 11 +++ .../diff/ut_diff_doc_since_change_001.txt | 11 +++ .../diff/ut_diff_doc_since_delete_001.txt | 11 +++ .../diff/ut_diff_doc_syscap_add_001.txt | 11 +++ .../diff/ut_diff_doc_syscap_change_001.txt | 11 +++ .../diff/ut_diff_doc_syscap_delete_001.txt | 11 +++ .../test/expect/diff/ut_diff_file_add_001.txt | 1 + .../test/expect/diff/ut_diff_file_add_002.txt | 1 + .../expect/diff/ut_diff_file_delete_001.txt | 1 + .../expect/diff/ut_diff_file_delete_002.txt | 1 + .../diff/ut_diff_node_enum_member_add_001.txt | 11 +++ ..._diff_node_enum_member_name_change_001.txt | 11 +++ .../ut_diff_node_enum_member_reduce_001.txt | 11 +++ ...diff_node_enum_member_value_change_001.txt | 11 +++ .../ut_diff_node_function_param_add_001.txt | 11 +++ ...ut_diff_node_function_param_reduce_001.txt | 11 +++ ...ff_node_function_param_type_change_001.txt | 20 ++++ .../diff/ut_diff_node_function_return_001.txt | 11 +++ .../diff/ut_diff_node_mar_define_add_001.txt | 11 +++ .../diff/ut_diff_node_mar_define_name_001.txt | 38 ++++++++ .../ut_diff_node_mar_define_reduce_001.txt | 11 +++ .../diff/ut_diff_node_mar_define_text_001.txt | 11 +++ .../ut_diff_node_struct_member_add_001.txt | 92 +++++++++++++++++++ ...iff_node_struct_member_name_change_001.txt | 11 +++ .../ut_diff_node_struct_member_reduce_001.txt | 11 +++ ...iff_node_struct_member_type_change_001.txt | 11 +++ ...diff_node_typedef_name_change_001 copy.txt | 20 ++++ .../ut_diff_node_union_member_add_001.txt | 11 +++ ...diff_node_union_member_name_change_001.txt | 11 +++ .../ut_diff_node_union_member_reduce_001.txt | 29 ++++++ ...diff_node_union_member_type_change_001.txt | 11 +++ .../capi_parser/test/output/diff/.gitkeep | 0 .../capi_parser/test/testCase/run_main.py | 34 ++++++- .../ut_diff_api_add_001/ut_diff_case_001.h | 73 +++++++++++++++ .../ut_diff_api_delete_001/ut_diff_case_001.h | 62 +++++++++++++ .../file1/ut_diff_case_001.h | 62 +++++++++++++ .../file2/ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_dir_add_001/ut_diff_case_001.h | 62 +++++++++++++ .../file1/ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_dir_delete_001/ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 61 ++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 58 ++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 61 ++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 61 ++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 61 ++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 61 ++++++++++++ .../ut_diff_file_add_001/ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_file_add_002/ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_file_add_002/ut_diff_case_002.h | 62 +++++++++++++ .../diff/new/ut_diff_file_delete_001/.gitkeep | 0 .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_node_case1.h | 61 ++++++++++++ .../ut_diff_node_case1.h | 61 ++++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 61 ++++++++++++ .../ut_diff_node_case1.h | 54 +++++++++++ .../ut_diff_node_case1.h | 54 +++++++++++ .../ut_diff_node_case1.h | 54 +++++++++++ .../ut_diff_node_case1.h | 54 +++++++++++ .../ut_diff_node_case1.h | 56 +++++++++++ .../ut_diff_node_case1.h | 56 +++++++++++ .../ut_diff_node_case1.h | 55 +++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 56 +++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 59 ++++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 56 +++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_api_add_001/ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_api_delete_001/ut_diff_case_001.h | 73 +++++++++++++++ .../file1/ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_dir_add_001/ut_diff_case_001.h | 62 +++++++++++++ .../file1/ut_diff_case_001.h | 62 +++++++++++++ .../file2/ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_dir_delete_001/ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 61 ++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 58 ++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 61 ++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 61 ++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 63 +++++++++++++ .../ut_diff_case_001.h | 61 ++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 61 ++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut/diff/old/ut_diff_file_add_001/.gitkeep | 0 .../ut_diff_file_add_002/ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_001.h | 62 +++++++++++++ .../ut_diff_case_002.h | 62 +++++++++++++ .../ut_diff_node_case1.h | 59 ++++++++++++ .../ut_diff_node_case1.h | 61 ++++++++++++ .../ut_diff_node_case1.h | 59 ++++++++++++ .../ut_diff_node_case1.h | 61 ++++++++++++ .../ut_diff_node_case1.h | 54 +++++++++++ .../ut_diff_node_case1.h | 54 +++++++++++ .../ut_diff_node_case1.h | 54 +++++++++++ .../ut_diff_node_case1.h | 54 +++++++++++ .../ut_diff_node_case1.h | 56 +++++++++++ .../ut_diff_node_case1.h | 56 +++++++++++ .../ut_diff_node_case1.h | 56 +++++++++++ .../ut_diff_node_case1.h | 56 +++++++++++ .../ut_diff_node_case1.h | 56 +++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 59 ++++++++++++ .../ut_diff_node_case1.h | 56 +++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ .../ut_diff_node_case1.h | 57 ++++++++++++ 190 files changed, 8379 insertions(+), 6 deletions(-) create mode 100644 build-tools/capi_parser/test/expect/diff/.gitkeep create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_api_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_api_delete_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_dir_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_dir_delete_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_addtogroup_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_addtogroup_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_addtogroup_delete_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_brief_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_brief_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_brief_delete_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_change_002.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_delete_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_file_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_file_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_file_delete_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_library_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_library_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_library_delete_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_add_002.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_add_003.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_delete_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_delete_002.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_delete_003.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_since_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_since_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_since_delete_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_syscap_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_syscap_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_doc_syscap_delete_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_file_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_file_add_002.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_file_delete_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_file_delete_002.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_name_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_reduce_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_value_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_function_param_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_function_param_reduce_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_function_param_type_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_function_return_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_name_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_reduce_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_text_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_name_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_reduce_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_type_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_typedef_name_change_001 copy.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_add_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_name_change_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_reduce_001.txt create mode 100644 build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_type_change_001.txt create mode 100644 build-tools/capi_parser/test/output/diff/.gitkeep create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_api_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_api_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_add_001/file1/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_add_001/file2/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_delete_001/file1/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_addtogroup_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_addtogroup_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_addtogroup_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_brief_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_brief_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_brief_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_change_002/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_file_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_file_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_file_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_library_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_library_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_library_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_add_002/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_add_003/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_delete_002/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_delete_003/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_since_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_since_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_since_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_syscap_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_syscap_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_syscap_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_file_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_file_add_002/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_file_add_002/ut_diff_case_002.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_file_delete_001/.gitkeep create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_file_delete_002/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_add_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_name_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_reduce_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_value_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_param_add_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_param_reduce_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_param_type_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_return_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_add_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_name_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_reduce_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_text_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_add_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_name_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_reduce_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_type_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_typedef_name_change_001 copy/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_add_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_name_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_reduce_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_type_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_api_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_api_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_add_001/file1/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_delete_001/file1/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_delete_001/file2/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_addtogroup_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_addtogroup_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_addtogroup_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_brief_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_brief_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_brief_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_change_002/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_file_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_file_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_file_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_library_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_library_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_library_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_add_002/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_add_003/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_delete_002/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_delete_003/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_since_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_since_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_since_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_syscap_add_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_syscap_change_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_syscap_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_file_add_001/.gitkeep create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_file_add_002/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_file_delete_001/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_file_delete_002/ut_diff_case_001.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_file_delete_002/ut_diff_case_002.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_add_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_name_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_reduce_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_value_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_param_add_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_param_reduce_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_param_type_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_return_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_add_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_name_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_reduce_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_text_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_struct_member_name_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_struct_member_reduce_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_struct_member_type_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_name_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_reduce_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_type_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_typedef_name_change_001 copy/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_add_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_name_change_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_reduce_001/ut_diff_node_case1.h create mode 100644 build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_type_change_001/ut_diff_node_case1.h diff --git a/build-tools/capi_parser/.gitignore b/build-tools/capi_parser/.gitignore index 05d361948..4ca112d82 100644 --- a/build-tools/capi_parser/.gitignore +++ b/build-tools/capi_parser/.gitignore @@ -1,2 +1,4 @@ /**/__pycache__/* -.idea/ \ No newline at end of file +.idea/ +test/ut/check/*.h.gch +test/output \ No newline at end of file diff --git a/build-tools/capi_parser/src/typedef/diff/diff.py b/build-tools/capi_parser/src/typedef/diff/diff.py index 8b94d8beb..fb784722f 100644 --- a/build-tools/capi_parser/src/typedef/diff/diff.py +++ b/build-tools/capi_parser/src/typedef/diff/diff.py @@ -106,7 +106,7 @@ class DiffType(enum.Enum): DOC_TAG_LIBRARY_A_TO_B = '变更library' DOC_TAG_PARAM_NA_TO_HAVE = '新增param标签' DOC_TAG_PARAM_HAVE_TO_NA = '删除param标签' - DOC_TAG_PARAM_NAME_A_TO_B = '修改param标签描述信息' + DOC_TAG_PARAM_NAME_A_TO_B = '修改param名称' DOC_TAG_PARAM_A_TO_B = '修改param标签描述信息' DOC_TAG_PERMISSION_NA_TO_HAVE = '权限从无到有' DOC_TAG_PERMISSION_HAVE_TO_NA = '权限从有到无' @@ -125,7 +125,33 @@ class DiffType(enum.Enum): DOC_TAG_RIGHT_BRACE_HAVE_TO_NA = '删除右括号' -compatible_list = [DiffType.FUNCTION_PARAM_NAME_CHANGE] +compatible_list = [ + DiffType.ADD_API, + DiffType.ADD_DOC, + DiffType.ADD_DOC_TAG, + DiffType.FUNCTION_PARAM_NAME_CHANGE, + DiffType.DOC_TAG_ADDTOGROUP_NA_TO_HAVE, + DiffType.DOC_TAG_ADDTOGROUP_HAVE_TO_NA, + DiffType.DOC_TAG_ADDTOGROUP_A_TO_B, + DiffType.DOC_TAG_BRIEF_NA_TO_HAVE, + DiffType.DOC_TAG_BRIEF_HAVE_TO_NA, + DiffType.DOC_TAG_BRIEF_A_TO_B, + DiffType.DOC_TAG_DEPRECATED_NA_TO_HAVE, + DiffType.DOC_TAG_DEPRECATED_HAVE_TO_NA, + DiffType.DOC_TAG_DEPRECATED_A_TO_B, + DiffType.DOC_TAG_FILE_NA_TO_HAVE, + DiffType.DOC_TAG_FILE_HAVE_TO_NA, + DiffType.DOC_TAG_FILE_A_TO_B, + DiffType.DOC_TAG_PARAM_A_TO_B, + DiffType.DOC_TAG_PERMISSION_HAVE_TO_NA, + DiffType.DOC_TAG_PERMISSION_RANGE_BIGGER, + DiffType.DOC_TAG_SINCE_NA_TO_HAVE, + DiffType.DOC_TAG_SINCE_A_TO_B, + DiffType.DOC_TAG_LEFT_BRACE_NA_TO_HAVE, + DiffType.DOC_TAG_LEFT_BRACE_HAVE_TO_NA, + DiffType.DOC_TAG_RIGHT_BRACE_NA_TO_HAVE, + DiffType.DOC_TAG_RIGHT_BRACE_HAVE_TO_NA, +] class DiffInfo: diff --git a/build-tools/capi_parser/test/expect/diff/.gitkeep b/build-tools/capi_parser/test/expect/diff/.gitkeep new file mode 100644 index 000000000..e69de29bb diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_api_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_api_add_001.txt new file mode 100644 index 000000000..0571009c4 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_api_add_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId2", + "api_type": "FUNCTION_DECL", + "diff_type": "ADD_API", + "diff_message": "add api", + "old_api_full_text": "", + "new_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId2(int *keyEvent, int *deviceId)", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_api_delete_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_api_delete_001.txt new file mode 100644 index 000000000..17e0506ae --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_api_delete_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId2", + "api_type": "FUNCTION_DECL", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId2(int *keyEvent, int *deviceId)", + "new_api_full_text": "", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_dir_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_dir_add_001.txt new file mode 100644 index 000000000..4f58c6be3 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_dir_add_001.txt @@ -0,0 +1,29 @@ +[ + { + "api_name": "NEURAL_NETWORK_RUNTIME_H", + "api_type": "MACRO_DEFINITION", + "diff_type": "ADD_API", + "diff_message": "add api", + "old_api_full_text": "", + "new_api_full_text": "NEURAL_NETWORK_RUNTIME_H", + "is_compatible": true + }, + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId", + "api_type": "FUNCTION_DECL", + "diff_type": "ADD_API", + "diff_message": "add api", + "old_api_full_text": "", + "new_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "is_compatible": true + }, + { + "api_name": "file2\\ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "ADD_API", + "diff_message": "add api", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_dir_delete_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_dir_delete_001.txt new file mode 100644 index 000000000..deb3629bf --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_dir_delete_001.txt @@ -0,0 +1,29 @@ +[ + { + "api_name": "NEURAL_NETWORK_RUNTIME_H", + "api_type": "MACRO_DEFINITION", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "NEURAL_NETWORK_RUNTIME_H", + "new_api_full_text": "", + "is_compatible": false + }, + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId", + "api_type": "FUNCTION_DECL", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "new_api_full_text": "", + "is_compatible": false + }, + { + "api_name": "file2\\ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_addtogroup_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_addtogroup_add_001.txt new file mode 100644 index 000000000..34095c3fe --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_addtogroup_add_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_ADDTOGROUP_NA_TO_HAVE", + "diff_message": "add addtogroup\u6807\u7b7e", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_addtogroup_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_addtogroup_change_001.txt new file mode 100644 index 000000000..bcc9bc517 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_addtogroup_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_ADDTOGROUP_A_TO_B", + "diff_message": "\u4fee\u6539addtogroup\u6807\u7b7e", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_addtogroup_delete_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_addtogroup_delete_001.txt new file mode 100644 index 000000000..476684e4a --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_addtogroup_delete_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_ADDTOGROUP_HAVE_TO_NA", + "diff_message": "delete addtogroup\u6807\u7b7e", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_brief_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_brief_add_001.txt new file mode 100644 index 000000000..cbae751f5 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_brief_add_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_BRIEF_NA_TO_HAVE", + "diff_message": "add brief\u6807\u7b7e", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_brief_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_brief_change_001.txt new file mode 100644 index 000000000..de6790b71 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_brief_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_BRIEF_A_TO_B", + "diff_message": "\u4fee\u6539brief\u6807\u7b7e", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_brief_delete_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_brief_delete_001.txt new file mode 100644 index 000000000..54f7491bf --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_brief_delete_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_BRIEF_HAVE_TO_NA", + "diff_message": "delete brief\u6807\u7b7e", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_add_001.txt new file mode 100644 index 000000000..ab876b344 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_add_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId", + "api_type": "FUNCTION_DECL", + "diff_type": "DOC_TAG_DEPRECATED_NA_TO_HAVE", + "diff_message": "\u63a5\u53e3\u53d8\u66f4\u4e3a\u5e9f\u5f03", + "old_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "new_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_change_001.txt new file mode 100644 index 000000000..192a415be --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId", + "api_type": "FUNCTION_DECL", + "diff_type": "DOC_TAG_DEPRECATED_A_TO_B", + "diff_message": "\u63a5\u53e3\u5e9f\u5f03\u7248\u672c\u53d1\u751f\u53d8\u5316", + "old_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "new_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_change_002.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_change_002.txt new file mode 100644 index 000000000..192a415be --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_change_002.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId", + "api_type": "FUNCTION_DECL", + "diff_type": "DOC_TAG_DEPRECATED_A_TO_B", + "diff_message": "\u63a5\u53e3\u5e9f\u5f03\u7248\u672c\u53d1\u751f\u53d8\u5316", + "old_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "new_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_delete_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_delete_001.txt new file mode 100644 index 000000000..3c9124ff2 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_deprecated_delete_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId", + "api_type": "FUNCTION_DECL", + "diff_type": "DOC_TAG_DEPRECATED_HAVE_TO_NA", + "diff_message": "\u5e9f\u5f03\u63a5\u53e3\u53d8\u66f4\u4e3a\u4e0d\u5e9f\u5f03", + "old_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "new_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_file_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_file_add_001.txt new file mode 100644 index 000000000..3529697ca --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_file_add_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_FILE_NA_TO_HAVE", + "diff_message": "add file\u6807\u7b7e", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_file_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_file_change_001.txt new file mode 100644 index 000000000..e43642de2 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_file_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_FILE_A_TO_B", + "diff_message": "\u4fee\u6539file\u6807\u7b7e", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_file_delete_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_file_delete_001.txt new file mode 100644 index 000000000..b808ceb9e --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_file_delete_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_FILE_HAVE_TO_NA", + "diff_message": "delete file\u6807\u7b7e", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_library_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_library_add_001.txt new file mode 100644 index 000000000..6aabc5340 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_library_add_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_LIBRARY_NA_TO_HAVE", + "diff_message": "add library", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_library_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_library_change_001.txt new file mode 100644 index 000000000..104abfa6e --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_library_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_LIBRARY_A_TO_B", + "diff_message": "\u53d8\u66f4library", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_library_delete_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_library_delete_001.txt new file mode 100644 index 000000000..91a6ea88a --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_library_delete_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_LIBRARY_HAVE_TO_NA", + "diff_message": "delete library", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_add_001.txt new file mode 100644 index 000000000..8048bd32a --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_add_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId", + "api_type": "FUNCTION_DECL", + "diff_type": "DOC_TAG_PERMISSION_NA_TO_HAVE", + "diff_message": "\u6743\u9650\u4ece\u65e0\u5230\u6709", + "old_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "new_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_add_002.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_add_002.txt new file mode 100644 index 000000000..acc1c32e8 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_add_002.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId", + "api_type": "FUNCTION_DECL", + "diff_type": "DOC_TAG_PERMISSION_RANGE_BIGGER", + "diff_message": "\u589e\u52a0or\u6216\u51cf\u5c11and\u6743\u9650", + "old_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "new_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_add_003.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_add_003.txt new file mode 100644 index 000000000..acc1c32e8 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_add_003.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId", + "api_type": "FUNCTION_DECL", + "diff_type": "DOC_TAG_PERMISSION_RANGE_BIGGER", + "diff_message": "\u589e\u52a0or\u6216\u51cf\u5c11and\u6743\u9650", + "old_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "new_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_change_001.txt new file mode 100644 index 000000000..f7634df16 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId", + "api_type": "FUNCTION_DECL", + "diff_type": "DOC_TAG_PERMISSION_RANGE_CHANGE", + "diff_message": "\u6743\u9650\u53d1\u9001\u6539\u53d8\u65e0\u6cd5\u5224\u65ad\u8303\u56f4\u53d8\u5316", + "old_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "new_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_delete_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_delete_001.txt new file mode 100644 index 000000000..69678035d --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_delete_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId", + "api_type": "FUNCTION_DECL", + "diff_type": "DOC_TAG_PERMISSION_HAVE_TO_NA", + "diff_message": "\u6743\u9650\u4ece\u6709\u5230\u65e0", + "old_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "new_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_delete_002.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_delete_002.txt new file mode 100644 index 000000000..26e3621f1 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_delete_002.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId", + "api_type": "FUNCTION_DECL", + "diff_type": "DOC_TAG_PERMISSION_RANGE_SMALLER", + "diff_message": "\u51cf\u5c11or\u6216\u589e\u52a0and\u6743\u9650", + "old_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "new_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_delete_003.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_delete_003.txt new file mode 100644 index 000000000..26e3621f1 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_permission_delete_003.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NativeXComponent_GetKeyEventDeviceId", + "api_type": "FUNCTION_DECL", + "diff_type": "DOC_TAG_PERMISSION_RANGE_SMALLER", + "diff_message": "\u51cf\u5c11or\u6216\u589e\u52a0and\u6743\u9650", + "old_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "new_api_full_text": "int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId)", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_since_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_since_add_001.txt new file mode 100644 index 000000000..a090da276 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_since_add_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_SINCE_NA_TO_HAVE", + "diff_message": "add since\u6807\u7b7e", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_since_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_since_change_001.txt new file mode 100644 index 000000000..93d5d2f9f --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_since_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_SINCE_A_TO_B", + "diff_message": "\u4fee\u6539since\u6807\u7b7e", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_since_delete_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_since_delete_001.txt new file mode 100644 index 000000000..4b1f6b814 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_since_delete_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_SINCE_HAVE_TO_NA", + "diff_message": "delete since\u6807\u7b7e", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_syscap_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_syscap_add_001.txt new file mode 100644 index 000000000..c1cf75828 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_syscap_add_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_SYSCAP_NA_TO_HAVE", + "diff_message": "\u4ece\u6ca1\u6709syscap\u5230\u6709syscap", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_syscap_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_syscap_change_001.txt new file mode 100644 index 000000000..b0344d9f8 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_syscap_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_SYSCAP_A_TO_B", + "diff_message": "syscap\u53d1\u751f\u6539\u53d8", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_doc_syscap_delete_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_syscap_delete_001.txt new file mode 100644 index 000000000..fd507f121 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_doc_syscap_delete_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ut_diff_case_001.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "DOC_TAG_SYSCAP_HAVE_TO_NA", + "diff_message": "\u4ece\u6709syscap\u5230\u6ca1\u6709syscap", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_file_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_file_add_001.txt new file mode 100644 index 000000000..0637a088a --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_file_add_001.txt @@ -0,0 +1 @@ +[] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_file_add_002.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_file_add_002.txt new file mode 100644 index 000000000..0637a088a --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_file_add_002.txt @@ -0,0 +1 @@ +[] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_file_delete_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_file_delete_001.txt new file mode 100644 index 000000000..0637a088a --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_file_delete_001.txt @@ -0,0 +1 @@ +[] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_file_delete_002.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_file_delete_002.txt new file mode 100644 index 000000000..0637a088a --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_file_delete_002.txt @@ -0,0 +1 @@ +[] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_add_001.txt new file mode 100644 index 000000000..995f5e5f0 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_add_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NN_FORMAT_NHWC", + "api_type": "ENUM_CONSTANT_DECL", + "diff_type": "ENUM_MEMBER_ADD", + "diff_message": "add \u679a\u4e3e\u6210\u5458", + "old_api_full_text": "enum Network {\n /** The tensor does not have a specific layout type (such as scalar or vector). */\n OH_NN_FORMAT_NONE = 0,\n /** The tensor arranges data in NCHW format.*/\n OH_NN_FORMAT_NCHW = 1\n};\n\n#i", + "new_api_full_text": "OH_NN_FORMAT_NHWC = 2", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_name_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_name_change_001.txt new file mode 100644 index 000000000..e74da6602 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_name_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NN_FORMAT", + "api_type": "ENUM_CONSTANT_DECL", + "diff_type": "ENUM_MEMBER_NAME_CHANGE", + "diff_message": "\u679a\u4e3e\u6210\u5458\u540d\u6539\u53d8", + "old_api_full_text": "OH_NN_FORMAT_NONE = 0", + "new_api_full_text": "OH_NN_FORMAT = 0", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_reduce_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_reduce_001.txt new file mode 100644 index 000000000..9893e9841 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_reduce_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "Network", + "api_type": "ENUM_DECL", + "diff_type": "ENUM_MEMBER_REDUCE", + "diff_message": "delete \u679a\u4e3e\u6210\u5458", + "old_api_full_text": "OH_NN_FORMAT_NCHW = 1", + "new_api_full_text": "enum Network {\n /** The tensor does not have a specific layout type (such as scalar or vector). */\n OH_NN_FORMAT_NONE = 0\n};\n\n", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_value_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_value_change_001.txt new file mode 100644 index 000000000..792bcd6fa --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_enum_member_value_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NN_FORMAT_NHWC", + "api_type": "ENUM_CONSTANT_DECL", + "diff_type": "ENUM_MEMBER_VALUE_CHANGE", + "diff_message": "\u679a\u4e3e\u6210\u5458\u503c\u6539\u53d8", + "old_api_full_text": "OH_NN_FORMAT_NHWC = 2", + "new_api_full_text": "OH_NN_FORMAT_NHWC = 5", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_function_param_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_function_param_add_001.txt new file mode 100644 index 000000000..e757570bc --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_function_param_add_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "outputShape", + "api_type": "PARM_DECL", + "diff_type": "FUNCTION_PARAM_ADD", + "diff_message": "add \u51fd\u6570\u53c2\u6570", + "old_api_full_text": "int OH_NNExecutor_GetOutputShape(int *executor, int *shapeLength)", + "new_api_full_text": "int *outputShape", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_function_param_reduce_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_function_param_reduce_001.txt new file mode 100644 index 000000000..732b7a4cd --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_function_param_reduce_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NNExecutor_GetOutputShape", + "api_type": "FUNCTION_DECL", + "diff_type": "FUNCTION_PARAM_REDUCE", + "diff_message": "delete \u51fd\u6570\u53c2\u6570", + "old_api_full_text": "int *shapeLength", + "new_api_full_text": "int OH_NNExecutor_GetOutputShape(int *executor)", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_function_param_type_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_function_param_type_change_001.txt new file mode 100644 index 000000000..a59f26ce9 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_function_param_type_change_001.txt @@ -0,0 +1,20 @@ +[ + { + "api_name": "executor", + "api_type": "PARM_DECL", + "diff_type": "FUNCTION_PARAM_TYPE_CHANGE", + "diff_message": "\u51fd\u6570\u53c2\u6570\u7c7b\u578b\u6539\u53d8", + "old_api_full_text": "int *executor", + "new_api_full_text": "char executor", + "is_compatible": false + }, + { + "api_name": "shapeLength", + "api_type": "PARM_DECL", + "diff_type": "FUNCTION_PARAM_TYPE_CHANGE", + "diff_message": "\u51fd\u6570\u53c2\u6570\u7c7b\u578b\u6539\u53d8", + "old_api_full_text": "int *shapeLength", + "new_api_full_text": "int shapeLength", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_function_return_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_function_return_001.txt new file mode 100644 index 000000000..067fa72c0 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_function_return_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "OH_NNExecutor_GetOutputShape", + "api_type": "FUNCTION_DECL", + "diff_type": "FUNCTION_RETURN_CHANGE", + "diff_message": "\u51fd\u6570\u8fd4\u56de\u7c7b\u578b\u6539\u53d8", + "old_api_full_text": "int OH_NNExecutor_GetOutputShape(int *executor, int *shapeLength)", + "new_api_full_text": "int *OH_NNExecutor_GetOutputShape(int *executor, int *shapeLength)", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_add_001.txt new file mode 100644 index 000000000..7c530d2ef --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_add_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "MAX(a,b)", + "api_type": "MACRO_DEFINITION", + "diff_type": "ADD_API", + "diff_message": "add api", + "old_api_full_text": "", + "new_api_full_text": "MAX(a,b) (a > b?a:b)", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_name_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_name_001.txt new file mode 100644 index 000000000..74f529fe9 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_name_001.txt @@ -0,0 +1,38 @@ +[ + { + "api_name": "NEURAL", + "api_type": "MACRO_DEFINITION", + "diff_type": "ADD_API", + "diff_message": "add api", + "old_api_full_text": "", + "new_api_full_text": "NEURAL", + "is_compatible": true + }, + { + "api_name": "KHRONOS_APICALL_AST", + "api_type": "MACRO_DEFINITION", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "KHRONOS_APICALL_AST __declspec(dllimport)", + "new_api_full_text": "", + "is_compatible": false + }, + { + "api_name": "NEURAL_AST", + "api_type": "MACRO_DEFINITION", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "NEURAL_AST", + "new_api_full_text": "", + "is_compatible": false + }, + { + "api_name": "KHRONOS_APICALL", + "api_type": "MACRO_DEFINITION", + "diff_type": "ADD_API", + "diff_message": "add api", + "old_api_full_text": "", + "new_api_full_text": "KHRONOS_APICALL __declspec(dllimport)", + "is_compatible": true + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_reduce_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_reduce_001.txt new file mode 100644 index 000000000..2377d5aff --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_reduce_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "KHRONOS_APICALL", + "api_type": "MACRO_DEFINITION", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "KHRONOS_APICALL __declspec(dllimport)", + "new_api_full_text": "", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_text_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_text_001.txt new file mode 100644 index 000000000..b071869c7 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_mar_define_text_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "NEURAL", + "api_type": "MACRO_DEFINITION", + "diff_type": "DEFINE_TEXT_CHANGE", + "diff_message": "\u5b8f\u6587\u672c\u6539\u53d8", + "old_api_full_text": "NEURAL 50", + "new_api_full_text": "NEURAL", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_add_001.txt new file mode 100644 index 000000000..21c0772ed --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_add_001.txt @@ -0,0 +1,92 @@ +[ + { + "api_name": "NEURAL_NETWORK_H", + "api_type": "MACRO_DEFINITION", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "NEURAL_NETWORK_H", + "new_api_full_text": "", + "is_compatible": false + }, + { + "api_name": "std", + "api_type": "STRUCT_DECL", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "struct std{\n char ch;\n int len;\n};\n\n", + "new_api_full_text": "", + "is_compatible": false + }, + { + "api_name": "ut_diff_node_struct_member_reduce_001\\ut_diff_node_case1.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": false + }, + { + "api_name": "NEURAL_NETWORK_H", + "api_type": "MACRO_DEFINITION", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "NEURAL_NETWORK_H", + "new_api_full_text": "", + "is_compatible": false + }, + { + "api_name": "std", + "api_type": "STRUCT_DECL", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "struct std{\n char ch;\n int len;\n};\n\n", + "new_api_full_text": "", + "is_compatible": false + }, + { + "api_name": "ut_diff_node_struct_member_name_change_001\\ut_diff_node_case1.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": false + }, + { + "api_name": "NEURAL_NETWORK_H", + "api_type": "MACRO_DEFINITION", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "NEURAL_NETWORK_H", + "new_api_full_text": "", + "is_compatible": false + }, + { + "api_name": "std", + "api_type": "STRUCT_DECL", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "struct std{\n char ch;\n int len;\n};\n\n", + "new_api_full_text": "", + "is_compatible": false + }, + { + "api_name": "ut_diff_node_struct_member_type_change_001\\ut_diff_node_case1.h", + "api_type": "TRANSLATION_UNIT", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "", + "new_api_full_text": "", + "is_compatible": false + }, + { + "api_name": "id", + "api_type": "FIELD_DECL", + "diff_type": "STRUCT_MEMBER_ADD", + "diff_message": "add \u7ed3\u6784\u4f53\u6210\u5458", + "old_api_full_text": "struct std{\n char ch;\n};\n", + "new_api_full_text": "int id", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_name_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_name_change_001.txt new file mode 100644 index 000000000..65a38efd2 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_name_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "character", + "api_type": "FIELD_DECL", + "diff_type": "STRUCT_MEMBER_NAME_CHANGE", + "diff_message": "\u7ed3\u6784\u4f53\u6210\u5458\u540d\u6539\u53d8", + "old_api_full_text": "char ch", + "new_api_full_text": "char character", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_reduce_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_reduce_001.txt new file mode 100644 index 000000000..d072b730c --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_reduce_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "std", + "api_type": "STRUCT_DECL", + "diff_type": "STRUCT_MEMBER_REDUCE", + "diff_message": "delete \u7ed3\u6784\u4f53\u6210\u5458", + "old_api_full_text": "int len", + "new_api_full_text": "struct std{\n char ch;\n};\n", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_type_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_type_change_001.txt new file mode 100644 index 000000000..bf1a249fd --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_struct_member_type_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "len", + "api_type": "FIELD_DECL", + "diff_type": "STRUCT_MEMBER_TYPE_CHANGE", + "diff_message": "\u7ed3\u6784\u4f53\u6210\u5458\u7c7b\u578b\u6539\u53d8", + "old_api_full_text": "int len", + "new_api_full_text": "char len", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_typedef_name_change_001 copy.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_typedef_name_change_001 copy.txt new file mode 100644 index 000000000..d05ef3a1a --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_typedef_name_change_001 copy.txt @@ -0,0 +1,20 @@ +[ + { + "api_name": "Network_New", + "api_type": "TYPEDEF_DECL", + "diff_type": "ADD_API", + "diff_message": "add api", + "old_api_full_text": "", + "new_api_full_text": "typedef enum {\n /** The tensor does not have a specific layout type (such as scalar or vector). */\n OH_NN_FORMAT_NONE = 0,\n /** The tensor arranges data in NCHW format.*/\n OH_NN_FORMAT_NCHW = 1\n} Network_New;\n\n#i", + "is_compatible": true + }, + { + "api_name": "Network", + "api_type": "TYPEDEF_DECL", + "diff_type": "REDUCE_API", + "diff_message": "delete api", + "old_api_full_text": "typedef enum {\n /** The tensor does not have a specific layout type (such as scalar or vector). */\n OH_NN_FORMAT_NONE = 0,\n /** The tensor arranges data in NCHW format.*/\n OH_NN_FORMAT_NCHW = 1\n} Network;\n\n#i", + "new_api_full_text": "", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_add_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_add_001.txt new file mode 100644 index 000000000..e67469333 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_add_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "na", + "api_type": "FIELD_DECL", + "diff_type": "UNION_MEMBER_ADD", + "diff_message": "add \u8054\u5408\u4f53\u6210\u5458", + "old_api_full_text": "union std{\n char ch;\n};\n", + "new_api_full_text": "char na", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_name_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_name_change_001.txt new file mode 100644 index 000000000..db7c6b6a2 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_name_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "key", + "api_type": "FIELD_DECL", + "diff_type": "UNION_MEMBER_NAME_CHANGE", + "diff_message": "\u8054\u5408\u4f53\u6210\u5458\u540d\u6539\u53d8", + "old_api_full_text": "int count", + "new_api_full_text": "int key", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_reduce_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_reduce_001.txt new file mode 100644 index 000000000..adb3de03b --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_reduce_001.txt @@ -0,0 +1,29 @@ +[ + { + "api_name": "count", + "api_type": "FIELD_DECL", + "diff_type": "UNION_MEMBER_TYPE_CHANGE", + "diff_message": "\u8054\u5408\u4f53\u6210\u5458\u7c7b\u578b\u6539\u53d8", + "old_api_full_text": "char ch", + "new_api_full_text": "int count", + "is_compatible": false + }, + { + "api_name": "count", + "api_type": "FIELD_DECL", + "diff_type": "UNION_MEMBER_NAME_CHANGE", + "diff_message": "\u8054\u5408\u4f53\u6210\u5458\u540d\u6539\u53d8", + "old_api_full_text": "char ch", + "new_api_full_text": "int count", + "is_compatible": false + }, + { + "api_name": "std", + "api_type": "UNION_DECL", + "diff_type": "UNION_MEMBER_REDUCE", + "diff_message": "delete \u8054\u5408\u4f53\u6210\u5458", + "old_api_full_text": "int count", + "new_api_full_text": "union std{\n int count;\n};\n", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_type_change_001.txt b/build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_type_change_001.txt new file mode 100644 index 000000000..a487872f9 --- /dev/null +++ b/build-tools/capi_parser/test/expect/diff/ut_diff_node_union_member_type_change_001.txt @@ -0,0 +1,11 @@ +[ + { + "api_name": "ch", + "api_type": "FIELD_DECL", + "diff_type": "UNION_MEMBER_TYPE_CHANGE", + "diff_message": "\u8054\u5408\u4f53\u6210\u5458\u7c7b\u578b\u6539\u53d8", + "old_api_full_text": "char ch", + "new_api_full_text": "unsigned int ch", + "is_compatible": false + } +] \ No newline at end of file diff --git a/build-tools/capi_parser/test/output/diff/.gitkeep b/build-tools/capi_parser/test/output/diff/.gitkeep new file mode 100644 index 000000000..e69de29bb diff --git a/build-tools/capi_parser/test/testCase/run_main.py b/build-tools/capi_parser/test/testCase/run_main.py index 9c0822758..ed624f937 100644 --- a/build-tools/capi_parser/test/testCase/run_main.py +++ b/build-tools/capi_parser/test/testCase/run_main.py @@ -19,7 +19,8 @@ import os import json import unittest sys.path.append(os.path.abspath(os.path.join(os.path.dirname(__file__), "../../src"))) -from coreImpl.check.check import get_check_result_list, write_in_txt +from coreImpl.check import check +from coreImpl.diff import diff_file class TestMethods(unittest.TestCase): @@ -31,13 +32,40 @@ class TestMethods(unittest.TestCase): for dirpath, dirnames, filenames in os.walk(test_case_path): for item in filenames: file_name = item.split('.')[0] - check_result = get_check_result_list([os.path.join(dirpath, item)]) - write_in_txt(check_result, os.path.join(output_path, "{}.txt".format(file_name))) + check_result = check.get_check_result_list([os.path.join(dirpath, item)]) + check.write_in_txt(check_result, os.path.join(output_path, "{}.txt".format(file_name))) with open(os.path.join(expect_path, "{}.txt".format(file_name))) as json_file: permission_file_content = json.load(json_file) result_json = json.dumps(permission_file_content, default=lambda obj: obj.__dict__, indent=4) self.assertEqual(result_json, "result_json", "{} case is error".format(os.path.join(dirpath, item))) + def test_diff(self): + test_path = os.path.abspath(os.path.join(os.path.dirname(__file__), "..\\")) + test_case_old_path = os.path.join(test_path, "ut\\diff\\old") + test_case_new_path = os.path.join(test_path, "ut\\diff\\new") + output_path = os.path.join(test_path, "output\\diff") + expect_path = os.path.join(test_path, "expect\\diff") + old_file_list = os.listdir(test_case_old_path) + for case_name in old_file_list: + output_file_name = os.path.join(output_path, "{}.txt".format(case_name)) + test_case_old_dir = os.path.join(test_case_old_path, case_name) + test_case_new_dir = os.path.join(test_case_new_path, case_name) + diff_result = diff_file.global_assignment(test_case_old_dir, test_case_new_dir) + diff_json = diff_file.result_to_json(diff_result) + diff_file.write_in_txt(diff_json, output_file_name) + expect_file_name = os.path.join(expect_path, "{}.txt".format(case_name)) + if not os.path.exists(expect_file_name): + continue + expect_file = open(expect_file_name) + expect_file_content = expect_file.read() + expect_file.close() + self.assertEqual(diff_json, expect_file_content, + f"\n{case_name} case is error.\n" + + f"\n{test_case_old_dir} is old case dir.\n" + + f"\n{test_case_new_dir} is new case dir.\n" + + f"{output_file_name} is out file.\n" + + f"{expect_file_name} is expect out file") + if __name__ == '__main__': unittest.main() diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_api_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_api_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9985af996 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_api_add_001/ut_diff_case_001.h @@ -0,0 +1,73 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId2(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_api_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_api_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_api_delete_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_add_001/file1/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_add_001/file1/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_add_001/file1/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_add_001/file2/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_add_001/file2/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_add_001/file2/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_add_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_delete_001/file1/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_delete_001/file1/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_delete_001/file1/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_dir_delete_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_addtogroup_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_addtogroup_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_addtogroup_add_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_addtogroup_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_addtogroup_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_addtogroup_change_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_addtogroup_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_addtogroup_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..ed6838add --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_addtogroup_delete_001/ut_diff_case_001.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_brief_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_brief_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_brief_add_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_brief_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_brief_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_brief_change_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_brief_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_brief_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..fe4761642 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_brief_delete_001/ut_diff_case_001.h @@ -0,0 +1,58 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..67e5912d9 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_add_001/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + * @deprecated since 10 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..fa8ad247f --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_change_001/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + * @deprecated since 11 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_change_002/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_change_002/ut_diff_case_001.h new file mode 100644 index 000000000..67e5912d9 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_change_002/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + * @deprecated since 10 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_deprecated_delete_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_file_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_file_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_file_add_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_file_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_file_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_file_change_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_file_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_file_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9beb2e9e6 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_file_delete_001/ut_diff_case_001.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_library_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_library_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_library_add_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_library_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_library_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_library_change_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_library_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_library_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..008e0fa59 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_library_delete_001/ut_diff_case_001.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..a117bfafe --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_add_001/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @permission ohos.permission.ANSWER_CALL + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_add_002/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_add_002/ut_diff_case_001.h new file mode 100644 index 000000000..db41a3b65 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_add_002/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @permission ohos.permission.ANSWER_CALL or ohos.permission.USE_BLUETOOTH + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_add_003/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_add_003/ut_diff_case_001.h new file mode 100644 index 000000000..a117bfafe --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_add_003/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @permission ohos.permission.ANSWER_CALL + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..47ed99e41 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_change_001/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @permission ohos.permission.USE_BLUETOOTH + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_delete_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_delete_002/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_delete_002/ut_diff_case_001.h new file mode 100644 index 000000000..a117bfafe --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_delete_002/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @permission ohos.permission.ANSWER_CALL + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_delete_003/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_delete_003/ut_diff_case_001.h new file mode 100644 index 000000000..d9385c6dc --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_permission_delete_003/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @permission ohos.permission.ANSWER_CALL and ohos.permission.USE_BLUETOOTH + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_since_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_since_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_since_add_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_since_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_since_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_since_change_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_since_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_since_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..0f9b91f48 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_since_delete_001/ut_diff_case_001.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_syscap_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_syscap_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_syscap_add_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_syscap_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_syscap_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_syscap_change_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_syscap_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_syscap_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..e90a40ca0 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_doc_syscap_delete_001/ut_diff_case_001.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_add_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_add_002/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_add_002/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_add_002/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_add_002/ut_diff_case_002.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_add_002/ut_diff_case_002.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_add_002/ut_diff_case_002.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_delete_001/.gitkeep b/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_delete_001/.gitkeep new file mode 100644 index 000000000..e69de29bb diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_delete_002/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_delete_002/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_file_delete_002/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_add_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_add_001/ut_diff_node_case1.h new file mode 100644 index 000000000..b63364cb8 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_add_001/ut_diff_node_case1.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +enum Network { + /** The tensor does not have a specific layout type (such as scalar or vector). */ + OH_NN_FORMAT_NONE = 0, + /** The tensor arranges data in NCHW format.*/ + OH_NN_FORMAT_NCHW = 1, + + OH_NN_FORMAT_NHWC = 2 +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_name_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_name_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..e145008d0 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_name_change_001/ut_diff_node_case1.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +enum Network { + /** The tensor does not have a specific layout type (such as scalar or vector). */ + OH_NN_FORMAT = 0, + /** The tensor arranges data in NCHW format.*/ + OH_NN_FORMAT_NCHW = 1, + /** The tensor arranges data in NHWC format.*/ + OH_NN_FORMAT_NHWC = 2 +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_reduce_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_reduce_001/ut_diff_node_case1.h new file mode 100644 index 000000000..e584559f9 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_reduce_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +enum Network { + /** The tensor does not have a specific layout type (such as scalar or vector). */ + OH_NN_FORMAT_NONE = 0 +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_value_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_value_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..9a2302789 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_enum_member_value_change_001/ut_diff_node_case1.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +enum Network { + /** The tensor does not have a specific layout type (such as scalar or vector). */ + OH_NN_FORMAT_NONE = 0, + /** The tensor arranges data in NCHW format.*/ + OH_NN_FORMAT_NCHW = 1, + /** The tensor arranges data in NHWC format.*/ + OH_NN_FORMAT_NHWC = 5 +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_param_add_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_param_add_001/ut_diff_node_case1.h new file mode 100644 index 000000000..59715f54a --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_param_add_001/ut_diff_node_case1.h @@ -0,0 +1,54 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +int OH_NNExecutor_GetOutputShape(int *executor, int *shapeLength, int *outputShape); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_param_reduce_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_param_reduce_001/ut_diff_node_case1.h new file mode 100644 index 000000000..bc0d56dc9 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_param_reduce_001/ut_diff_node_case1.h @@ -0,0 +1,54 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +int OH_NNExecutor_GetOutputShape(int *executor); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_param_type_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_param_type_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..a4675a6e3 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_param_type_change_001/ut_diff_node_case1.h @@ -0,0 +1,54 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +int OH_NNExecutor_GetOutputShape(char executor, int shapeLength); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_return_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_return_001/ut_diff_node_case1.h new file mode 100644 index 000000000..7b94278cc --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_function_return_001/ut_diff_node_case1.h @@ -0,0 +1,54 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +int *OH_NNExecutor_GetOutputShape(int *executor, int *shapeLength); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_add_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_add_001/ut_diff_node_case1.h new file mode 100644 index 000000000..478093ac2 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_add_001/ut_diff_node_case1.h @@ -0,0 +1,56 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +#define NEURAL +#define MAX(a,b) (a > b?a:b) +#define KHRONOS_APICALL __declspec(dllimport) + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_name_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_name_001/ut_diff_node_case1.h new file mode 100644 index 000000000..80a3597fc --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_name_001/ut_diff_node_case1.h @@ -0,0 +1,56 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +#define NEURAL + +#define KHRONOS_APICALL __declspec(dllimport) + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_reduce_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_reduce_001/ut_diff_node_case1.h new file mode 100644 index 000000000..632c0c0a4 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_reduce_001/ut_diff_node_case1.h @@ -0,0 +1,55 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +#define NEURAL + + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_text_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_text_001/ut_diff_node_case1.h new file mode 100644 index 000000000..b81992e5e --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_mar_define_text_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +#define NEURAL + + +#define KHRONOS_APICALL __declspec + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_add_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_add_001/ut_diff_node_case1.h new file mode 100644 index 000000000..d76320a36 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_add_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +struct std{ + char ch; + int id; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_name_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_name_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..dff0014a9 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_name_change_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +struct std{ + char character; + int len; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_reduce_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_reduce_001/ut_diff_node_case1.h new file mode 100644 index 000000000..14fe9fb54 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_reduce_001/ut_diff_node_case1.h @@ -0,0 +1,56 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +struct std{ + char ch; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_type_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_type_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..9da616fb7 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_struct_member_type_change_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +struct std{ + char ch; + char len; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_typedef_name_change_001 copy/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_typedef_name_change_001 copy/ut_diff_node_case1.h new file mode 100644 index 000000000..abac3a731 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_typedef_name_change_001 copy/ut_diff_node_case1.h @@ -0,0 +1,59 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +typedef enum { + /** The tensor does not have a specific layout type (such as scalar or vector). */ + OH_NN_FORMAT_NONE = 0, + /** The tensor arranges data in NCHW format.*/ + OH_NN_FORMAT_NCHW = 1 +} Network_New; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_add_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_add_001/ut_diff_node_case1.h new file mode 100644 index 000000000..5927da804 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_add_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +union std{ + char ch; + char na; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_name_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_name_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..f299ea8a2 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_name_change_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +union std{ + char ch; + int key; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_reduce_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_reduce_001/ut_diff_node_case1.h new file mode 100644 index 000000000..c0179ba54 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_reduce_001/ut_diff_node_case1.h @@ -0,0 +1,56 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +union std{ + int count; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_type_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_type_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..87bc2944d --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/new/ut_diff_node_union_member_type_change_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +union std{ + unsigned int ch; + int count; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_api_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_api_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_api_add_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_api_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_api_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9985af996 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_api_delete_001/ut_diff_case_001.h @@ -0,0 +1,73 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId2(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_add_001/file1/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_add_001/file1/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_add_001/file1/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_add_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_delete_001/file1/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_delete_001/file1/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_delete_001/file1/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_delete_001/file2/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_delete_001/file2/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_delete_001/file2/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_dir_delete_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_addtogroup_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_addtogroup_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..ed6838add --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_addtogroup_add_001/ut_diff_case_001.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_addtogroup_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_addtogroup_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..3e2dfeab2 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_addtogroup_change_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime1 + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_addtogroup_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_addtogroup_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_addtogroup_delete_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_brief_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_brief_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..fe4761642 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_brief_add_001/ut_diff_case_001.h @@ -0,0 +1,58 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_brief_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_brief_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..f7a65078d --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_brief_change_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines123123 the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_brief_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_brief_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_brief_delete_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_add_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..67e5912d9 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_change_001/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + * @deprecated since 10 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_change_002/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_change_002/ut_diff_case_001.h new file mode 100644 index 000000000..c8a47618b --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_change_002/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + * @deprecated 10 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..67e5912d9 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_deprecated_delete_001/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + * @deprecated since 10 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_file_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_file_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9beb2e9e6 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_file_add_001/ut_diff_case_001.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_file_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_file_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..95b4c350a --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_file_change_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime1.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_file_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_file_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_file_delete_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_library_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_library_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..008e0fa59 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_library_add_001/ut_diff_case_001.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_library_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_library_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..2cd98520a --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_library_change_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime1.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_library_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_library_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_library_delete_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_add_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_add_002/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_add_002/ut_diff_case_001.h new file mode 100644 index 000000000..a117bfafe --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_add_002/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @permission ohos.permission.ANSWER_CALL + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_add_003/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_add_003/ut_diff_case_001.h new file mode 100644 index 000000000..d9385c6dc --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_add_003/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @permission ohos.permission.ANSWER_CALL and ohos.permission.USE_BLUETOOTH + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..a117bfafe --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_change_001/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @permission ohos.permission.ANSWER_CALL + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..a117bfafe --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_delete_001/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @permission ohos.permission.ANSWER_CALL + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_delete_002/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_delete_002/ut_diff_case_001.h new file mode 100644 index 000000000..db41a3b65 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_delete_002/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @permission ohos.permission.ANSWER_CALL or ohos.permission.USE_BLUETOOTH + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_delete_003/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_delete_003/ut_diff_case_001.h new file mode 100644 index 000000000..a117bfafe --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_permission_delete_003/ut_diff_case_001.h @@ -0,0 +1,63 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @permission ohos.permission.ANSWER_CALL + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_since_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_since_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..0f9b91f48 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_since_add_001/ut_diff_case_001.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_since_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_since_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..fb7da4fa7 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_since_change_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 8 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_since_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_since_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_since_delete_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_syscap_add_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_syscap_add_001/ut_diff_case_001.h new file mode 100644 index 000000000..e90a40ca0 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_syscap_add_001/ut_diff_case_001.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_syscap_change_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_syscap_change_001/ut_diff_case_001.h new file mode 100644 index 000000000..b55f7ebf0 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_syscap_change_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData1 + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_syscap_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_syscap_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_doc_syscap_delete_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_add_001/.gitkeep b/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_add_001/.gitkeep new file mode 100644 index 000000000..e69de29bb diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_add_002/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_add_002/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_add_002/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_delete_001/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_delete_001/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_delete_001/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_delete_002/ut_diff_case_001.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_delete_002/ut_diff_case_001.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_delete_002/ut_diff_case_001.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_delete_002/ut_diff_case_002.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_delete_002/ut_diff_case_002.h new file mode 100644 index 000000000..9c8d137ce --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_file_delete_002/ut_diff_case_002.h @@ -0,0 +1,62 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs + * provided by Neural Network Runtime to construct and compile models + * and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + +#ifdef __cplusplus +extern "C" { +#endif + + /** + * @brief Obtains the deviceId of the key event. + * + * @param keyEvent keyEvent value + * @param deviceId deviceId value + * @return Returns the status code of the execution. + * @since 10 + * @version 1.0 + */ + int OH_NativeXComponent_GetKeyEventDeviceId(int *keyEvent, int *deviceId); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_add_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_add_001/ut_diff_node_case1.h new file mode 100644 index 000000000..b9567e15b --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_add_001/ut_diff_node_case1.h @@ -0,0 +1,59 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +enum Network { + /** The tensor does not have a specific layout type (such as scalar or vector). */ + OH_NN_FORMAT_NONE = 0, + /** The tensor arranges data in NCHW format.*/ + OH_NN_FORMAT_NCHW = 1 +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_name_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_name_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..d0349acd0 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_name_change_001/ut_diff_node_case1.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +enum Network { + /** The tensor does not have a specific layout type (such as scalar or vector). */ + OH_NN_FORMAT_NONE = 0, + /** The tensor arranges data in NCHW format.*/ + OH_NN_FORMAT_NCHW = 1, + /** The tensor arranges data in NHWC format.*/ + OH_NN_FORMAT_NHWC = 2 +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_reduce_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_reduce_001/ut_diff_node_case1.h new file mode 100644 index 000000000..b9567e15b --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_reduce_001/ut_diff_node_case1.h @@ -0,0 +1,59 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +enum Network { + /** The tensor does not have a specific layout type (such as scalar or vector). */ + OH_NN_FORMAT_NONE = 0, + /** The tensor arranges data in NCHW format.*/ + OH_NN_FORMAT_NCHW = 1 +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_value_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_value_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..d0349acd0 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_enum_member_value_change_001/ut_diff_node_case1.h @@ -0,0 +1,61 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +enum Network { + /** The tensor does not have a specific layout type (such as scalar or vector). */ + OH_NN_FORMAT_NONE = 0, + /** The tensor arranges data in NCHW format.*/ + OH_NN_FORMAT_NCHW = 1, + /** The tensor arranges data in NHWC format.*/ + OH_NN_FORMAT_NHWC = 2 +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_param_add_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_param_add_001/ut_diff_node_case1.h new file mode 100644 index 000000000..8579077e1 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_param_add_001/ut_diff_node_case1.h @@ -0,0 +1,54 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +int OH_NNExecutor_GetOutputShape(int *executor, int *shapeLength); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_param_reduce_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_param_reduce_001/ut_diff_node_case1.h new file mode 100644 index 000000000..8579077e1 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_param_reduce_001/ut_diff_node_case1.h @@ -0,0 +1,54 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +int OH_NNExecutor_GetOutputShape(int *executor, int *shapeLength); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_param_type_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_param_type_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..8579077e1 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_param_type_change_001/ut_diff_node_case1.h @@ -0,0 +1,54 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +int OH_NNExecutor_GetOutputShape(int *executor, int *shapeLength); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_return_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_return_001/ut_diff_node_case1.h new file mode 100644 index 000000000..8579077e1 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_function_return_001/ut_diff_node_case1.h @@ -0,0 +1,54 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +int OH_NNExecutor_GetOutputShape(int *executor, int *shapeLength); + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_add_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_add_001/ut_diff_node_case1.h new file mode 100644 index 000000000..80a3597fc --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_add_001/ut_diff_node_case1.h @@ -0,0 +1,56 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +#define NEURAL + +#define KHRONOS_APICALL __declspec(dllimport) + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_name_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_name_001/ut_diff_node_case1.h new file mode 100644 index 000000000..e93b209bf --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_name_001/ut_diff_node_case1.h @@ -0,0 +1,56 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +#define NEURAL_AST + +#define KHRONOS_APICALL_AST __declspec(dllimport) + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_reduce_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_reduce_001/ut_diff_node_case1.h new file mode 100644 index 000000000..80a3597fc --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_reduce_001/ut_diff_node_case1.h @@ -0,0 +1,56 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_RUNTIME_H +#define NEURAL_NETWORK_RUNTIME_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +#define NEURAL + +#define KHRONOS_APICALL __declspec(dllimport) + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_RUNTIME_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_text_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_text_001/ut_diff_node_case1.h new file mode 100644 index 000000000..6378c3b2b --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_mar_define_text_001/ut_diff_node_case1.h @@ -0,0 +1,56 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +#define NEURAL 50 + +#define KHRONOS_APICALL __declspec + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_case1.h new file mode 100644 index 000000000..14fe9fb54 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_case1.h @@ -0,0 +1,56 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +struct std{ + char ch; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_struct_member_name_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_struct_member_name_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..2f250f932 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_struct_member_name_change_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +struct std{ + char ch; + int len; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_struct_member_reduce_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_struct_member_reduce_001/ut_diff_node_case1.h new file mode 100644 index 000000000..2f250f932 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_struct_member_reduce_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +struct std{ + char ch; + int len; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_struct_member_type_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_struct_member_type_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..2f250f932 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_add_001/ut_diff_node_struct_member_type_change_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +struct std{ + char ch; + int len; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_name_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_name_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..2f250f932 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_name_change_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +struct std{ + char ch; + int len; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_reduce_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_reduce_001/ut_diff_node_case1.h new file mode 100644 index 000000000..2f250f932 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_reduce_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +struct std{ + char ch; + int len; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_type_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_type_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..2f250f932 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_struct_member_type_change_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +struct std{ + char ch; + int len; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_typedef_name_change_001 copy/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_typedef_name_change_001 copy/ut_diff_node_case1.h new file mode 100644 index 000000000..fd1c2047e --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_typedef_name_change_001 copy/ut_diff_node_case1.h @@ -0,0 +1,59 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +typedef enum { + /** The tensor does not have a specific layout type (such as scalar or vector). */ + OH_NN_FORMAT_NONE = 0, + /** The tensor arranges data in NCHW format.*/ + OH_NN_FORMAT_NCHW = 1 +} Network; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_add_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_add_001/ut_diff_node_case1.h new file mode 100644 index 000000000..67126fc79 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_add_001/ut_diff_node_case1.h @@ -0,0 +1,56 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +union std{ + char ch; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_name_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_name_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..f33cc2505 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_name_change_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +union std{ + char ch; + int count; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_reduce_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_reduce_001/ut_diff_node_case1.h new file mode 100644 index 000000000..f33cc2505 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_reduce_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +union std{ + char ch; + int count; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H diff --git a/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_type_change_001/ut_diff_node_case1.h b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_type_change_001/ut_diff_node_case1.h new file mode 100644 index 000000000..f33cc2505 --- /dev/null +++ b/build-tools/capi_parser/test/ut/diff/old/ut_diff_node_union_member_type_change_001/ut_diff_node_case1.h @@ -0,0 +1,57 @@ +/* + * Copyright (c) 2022 Huawei Device Co., Ltd. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +/** + * @addtogroup NeuralNeworkRuntime + * @{ + * + * @brief Provides APIs of Neural Network Runtime for accelerating the model inference. + * + * @since 9 + * @version 1.0 + */ + +/** + * @file neural_network_runtime.h + * + * @brief Defines the Neural Network Runtime APIs. The AI inference framework uses the Native APIs provided by Neural Network Runtime + * to construct and compile models and perform inference and computing on acceleration hardware. + * Note: Currently, the APIs of Neural Network Runtime do not support multi-thread calling. \n + * + * @syscap SystemCapability.Applications.CalendarData + * @library libneural_network_runtime.so + * @since 9 + * @version 1.0 + */ +#ifndef NEURAL_NETWORK_H +#define NEURAL_NETWORK_H + + + +#ifdef __cplusplus +extern "C" { +#endif + +union std{ + char ch; + int count; +}; + +#ifdef __cplusplus +} +#endif // __cplusplus + +/** @} */ +#endif // NEURAL_NETWORK_H -- Gitee