From 5fbd9d35a72ed2a75775daad24f747d6e0390dd4 Mon Sep 17 00:00:00 2001
From: =?UTF-8?q?=E5=AE=A6=E6=99=93=E7=8E=B2?= <3174348550@qq.com>
Date: Wed, 10 Dec 2025 10:21:43 +0800
Subject: [PATCH] modify links 2.7.2
---
.jenkins/check/config/filter_notebooklint.txt | 2 +
docs/lite/api/_custom/sphinx_builder_html | 2 +-
.../api/source_en/api_c/lite_c_example.rst | 2 +-
.../source_en/api_cpp/lite_cpp_example.rst | 6 +-
.../source_en/api_java/ascend_device_info.md | 2 +-
.../lite/api/source_en/api_java/class_list.md | 28 +-
docs/lite/api/source_en/api_java/graph.md | 2 +-
.../source_en/api_java/lite_java_example.rst | 6 +-
docs/lite/api/source_en/api_java/model.md | 2 +-
.../api_java/model_parallel_runner.md | 2 +-
docs/lite/api/source_en/api_java/mscontext.md | 10 +-
docs/lite/api/source_en/api_java/mstensor.md | 4 +-
.../api/source_en/api_java/runner_config.md | 2 +-
docs/lite/api/source_en/api_java/train_cfg.md | 2 +-
docs/lite/api/source_en/api_java/version.md | 2 +-
docs/lite/api/source_en/index.rst | 116 +-
docs/lite/api/source_zh_cn/api_c/context_c.md | 4 +-
.../api/source_zh_cn/api_c/data_type_c.md | 2 +-
docs/lite/api/source_zh_cn/api_c/format_c.md | 2 +-
.../api/source_zh_cn/api_c/lite_c_example.rst | 2 +-
docs/lite/api/source_zh_cn/api_c/model_c.md | 10 +-
docs/lite/api/source_zh_cn/api_c/status_c.md | 2 +-
docs/lite/api/source_zh_cn/api_c/tensor_c.md | 8 +-
docs/lite/api/source_zh_cn/api_c/types_c.md | 2 +-
.../source_zh_cn/api_cpp/lite_cpp_example.rst | 6 +-
.../api/source_zh_cn/api_cpp/mindspore.md | 46 +-
.../api_cpp/mindspore_converter.md | 16 +-
.../api_cpp/mindspore_datatype.md | 2 +-
.../source_zh_cn/api_cpp/mindspore_format.md | 2 +-
.../source_zh_cn/api_cpp/mindspore_kernel.md | 36 +-
.../api_cpp/mindspore_registry.md | 108 +-
.../api_cpp/mindspore_registry_opencl.md | 6 +-
.../api_java/ascend_device_info.md | 2 +-
.../api/source_zh_cn/api_java/class_list.md | 28 +-
docs/lite/api/source_zh_cn/api_java/graph.md | 2 +-
.../api_java/lite_java_example.rst | 6 +-
docs/lite/api/source_zh_cn/api_java/model.md | 2 +-
.../api_java/model_parallel_runner.md | 2 +-
.../api/source_zh_cn/api_java/mscontext.md | 10 +-
.../api/source_zh_cn/api_java/mstensor.md | 4 +-
.../source_zh_cn/api_java/runner_config.md | 2 +-
.../api/source_zh_cn/api_java/train_cfg.md | 2 +-
.../lite/api/source_zh_cn/api_java/version.md | 2 +-
docs/lite/api/source_zh_cn/index.rst | 116 +-
.../source_en/advanced/image_processing.md | 12 +-
docs/lite/docs/source_en/advanced/micro.md | 44 +-
.../docs/source_en/advanced/quantization.md | 10 +-
.../docs/source_en/advanced/third_party.rst | 4 +-
.../advanced/third_party/ascend_info.md | 24 +-
.../source_en/advanced/third_party/asic.rst | 4 +-
.../third_party/converter_register.md | 42 +-
.../advanced/third_party/delegate.md | 48 +-
.../advanced/third_party/npu_info.md | 16 +-
.../advanced/third_party/register.rst | 4 +-
.../advanced/third_party/register_kernel.md | 38 +-
.../advanced/third_party/tensorrt_info.md | 22 +-
docs/lite/docs/source_en/build/build.md | 12 +-
.../source_en/converter/converter_tool.md | 12 +-
docs/lite/docs/source_en/index.rst | 104 +-
.../source_en/infer/device_infer_example.rst | 4 +-
.../source_en/infer/image_segmentation.md | 16 +-
docs/lite/docs/source_en/infer/quick_start.md | 10 +-
.../docs/source_en/infer/quick_start_c.md | 24 +-
.../docs/source_en/infer/quick_start_cpp.md | 28 +-
.../docs/source_en/infer/quick_start_java.md | 20 +-
docs/lite/docs/source_en/infer/runtime_cpp.md | 130 +--
.../lite/docs/source_en/infer/runtime_java.md | 58 +-
docs/lite/docs/source_en/mindir/benchmark.rst | 4 +-
.../docs/source_en/mindir/benchmark_tool.md | 6 +-
docs/lite/docs/source_en/mindir/build.md | 4 +-
docs/lite/docs/source_en/mindir/converter.rst | 4 +-
.../docs/source_en/mindir/converter_python.md | 16 +-
.../docs/source_en/mindir/converter_tool.md | 6 +-
.../source_en/mindir/converter_tool_ascend.md | 14 +-
docs/lite/docs/source_en/mindir/runtime.rst | 4 +-
.../lite/docs/source_en/mindir/runtime_cpp.md | 58 +-
.../source_en/mindir/runtime_distributed.rst | 4 +-
.../mindir/runtime_distributed_cpp.md | 32 +-
.../runtime_distributed_multicard_python.md | 22 +-
.../mindir/runtime_distributed_python.md | 32 +-
.../docs/source_en/mindir/runtime_java.md | 48 +-
.../source_en/mindir/runtime_parallel.rst | 4 +-
.../source_en/mindir/runtime_parallel_cpp.md | 24 +-
.../source_en/mindir/runtime_parallel_java.md | 22 +-
.../mindir/runtime_parallel_python.md | 26 +-
.../docs/source_en/mindir/runtime_python.md | 30 +-
.../quick_start/one_hour_introduction.md | 50 +-
.../source_en/reference/architecture_lite.md | 2 +-
.../reference/environment_variable_support.md | 2 +-
docs/lite/docs/source_en/reference/faq.md | 30 +-
.../reference/image_classification_lite.md | 2 +-
.../reference/image_segmentation_lite.md | 2 +-
docs/lite/docs/source_en/reference/log.md | 2 +-
.../docs/source_en/reference/model_lite.rst | 4 +-
.../reference/object_detection_lite.md | 2 +-
.../reference/operator_list_codegen.md | 2 +-
.../source_en/reference/operator_list_lite.md | 2 +-
.../reference/operator_list_lite_for_caffe.md | 2 +-
.../reference/operator_list_lite_for_onnx.md | 2 +-
.../operator_list_lite_for_tensorflow.md | 4 +-
.../operator_list_lite_for_tflite.md | 2 +-
.../source_en/reference/operator_lite.rst | 4 +-
.../reference/scene_detection_lite.md | 2 +-
.../reference/style_transfer_lite.md | 2 +-
docs/lite/docs/source_en/tools/benchmark.rst | 4 +-
.../source_en/tools/benchmark_golden_data.md | 6 +-
.../docs/source_en/tools/benchmark_tool.md | 12 +-
.../source_en/tools/benchmark_train_tool.md | 8 +-
.../lite/docs/source_en/tools/cropper_tool.md | 12 +-
.../docs/source_en/tools/obfuscator_tool.md | 4 +-
docs/lite/docs/source_en/tools/visual_tool.md | 2 +-
.../docs/source_en/train/converter_train.md | 8 +-
.../source_en/train/device_train_example.rst | 4 +-
.../docs/source_en/train/runtime_train.rst | 4 +-
.../docs/source_en/train/runtime_train_cpp.md | 22 +-
.../source_en/train/runtime_train_java.md | 30 +-
docs/lite/docs/source_en/train/train_lenet.md | 20 +-
.../docs/source_en/train/train_lenet_java.md | 10 +-
docs/lite/docs/source_en/use/downloads.md | 4 +-
.../source_zh_cn/advanced/image_processing.md | 12 +-
docs/lite/docs/source_zh_cn/advanced/micro.md | 44 +-
.../source_zh_cn/advanced/quantization.md | 10 +-
.../source_zh_cn/advanced/third_party.rst | 4 +-
.../advanced/third_party/ascend_info.md | 24 +-
.../advanced/third_party/asic.rst | 4 +-
.../third_party/converter_register.md | 42 +-
.../advanced/third_party/delegate.md | 48 +-
.../advanced/third_party/npu_info.md | 14 +-
.../advanced/third_party/register.rst | 4 +-
.../advanced/third_party/register_kernel.md | 38 +-
.../advanced/third_party/tensorrt_info.md | 22 +-
docs/lite/docs/source_zh_cn/build/build.md | 12 +-
.../source_zh_cn/converter/converter_tool.md | 14 +-
docs/lite/docs/source_zh_cn/index.rst | 104 +-
.../infer/device_infer_example.rst | 4 +-
.../source_zh_cn/infer/image_segmentation.md | 16 +-
.../docs/source_zh_cn/infer/quick_start.md | 10 +-
.../docs/source_zh_cn/infer/quick_start_c.md | 24 +-
.../source_zh_cn/infer/quick_start_cpp.md | 30 +-
.../source_zh_cn/infer/quick_start_java.md | 20 +-
.../docs/source_zh_cn/infer/runtime_cpp.md | 130 +--
.../docs/source_zh_cn/infer/runtime_java.md | 58 +-
.../docs/source_zh_cn/mindir/benchmark.rst | 4 +-
.../source_zh_cn/mindir/benchmark_tool.md | 6 +-
docs/lite/docs/source_zh_cn/mindir/build.md | 4 +-
.../docs/source_zh_cn/mindir/converter.rst | 4 +-
.../source_zh_cn/mindir/converter_custom.md | 20 +-
.../source_zh_cn/mindir/converter_python.md | 16 +-
.../source_zh_cn/mindir/converter_tool.md | 6 +-
.../mindir/converter_tool_ascend.md | 14 +-
.../lite/docs/source_zh_cn/mindir/runtime.rst | 4 +-
.../docs/source_zh_cn/mindir/runtime_cpp.md | 60 +-
.../mindir/runtime_distributed.rst | 4 +-
.../mindir/runtime_distributed_cpp.md | 30 +-
.../runtime_distributed_multicard_python.md | 22 +-
.../mindir/runtime_distributed_python.md | 30 +-
.../docs/source_zh_cn/mindir/runtime_java.md | 48 +-
.../source_zh_cn/mindir/runtime_parallel.rst | 4 +-
.../mindir/runtime_parallel_cpp.md | 24 +-
.../mindir/runtime_parallel_java.md | 22 +-
.../mindir/runtime_parallel_python.md | 26 +-
.../source_zh_cn/mindir/runtime_python.md | 30 +-
.../quick_start/one_hour_introduction.md | 50 +-
.../reference/architecture_lite.md | 2 +-
.../reference/environment_variable_support.md | 2 +-
docs/lite/docs/source_zh_cn/reference/faq.md | 30 +-
.../reference/image_classification_lite.md | 2 +-
.../reference/image_segmentation_lite.md | 2 +-
docs/lite/docs/source_zh_cn/reference/log.md | 2 +-
.../source_zh_cn/reference/model_lite.rst | 4 +-
.../reference/object_detection_lite.md | 2 +-
.../reference/operator_list_codegen.md | 2 +-
.../reference/operator_list_lite.md | 2 +-
.../reference/operator_list_lite_for_caffe.md | 2 +-
.../reference/operator_list_lite_for_onnx.md | 2 +-
.../operator_list_lite_for_tensorflow.md | 4 +-
.../operator_list_lite_for_tflite.md | 2 +-
.../source_zh_cn/reference/operator_lite.rst | 4 +-
.../reference/scene_detection_lite.md | 2 +-
.../reference/style_transfer_lite.md | 2 +-
.../docs/source_zh_cn/tools/benchmark.rst | 4 +-
.../tools/benchmark_golden_data.md | 6 +-
.../docs/source_zh_cn/tools/benchmark_tool.md | 12 +-
.../tools/benchmark_train_tool.md | 8 +-
.../docs/source_zh_cn/tools/cropper_tool.md | 12 +-
.../source_zh_cn/tools/obfuscator_tool.md | 4 +-
.../docs/source_zh_cn/tools/visual_tool.md | 2 +-
.../source_zh_cn/train/converter_train.md | 8 +-
.../train/device_train_example.rst | 4 +-
.../docs/source_zh_cn/train/runtime_train.rst | 4 +-
.../source_zh_cn/train/runtime_train_cpp.md | 32 +-
.../source_zh_cn/train/runtime_train_java.md | 30 +-
.../docs/source_zh_cn/train/train_lenet.md | 20 +-
.../source_zh_cn/train/train_lenet_java.md | 10 +-
docs/lite/docs/source_zh_cn/use/downloads.md | 4 +-
.../accuracy_comparison.md | 4 +-
.../advanced_development/dev_migration.md | 4 +-
.../inference_precision_comparison.md | 4 +-
.../performance_optimization.md | 18 +-
.../precision_optimization.md | 8 +-
.../training_template_instruction.md | 6 +-
.../advanced_development/weight_transfer.md | 2 +-
.../yaml_config_inference.md | 2 +-
.../contribution/mindformers_contribution.md | 2 +-
.../contribution/modelers_contribution.md | 2 +-
.../docs/source_en/env_variables.md | 4 +-
.../source_en/example/distilled/distilled.md | 4 +-
.../docs/source_en/faq/feature_related.md | 2 +-
.../docs/source_en/faq/model_related.md | 2 +-
.../docs/source_en/feature/ckpt.md | 2 +-
.../docs/source_en/feature/configuration.md | 20 +-
.../docs/source_en/feature/dataset.md | 10 +-
.../source_en/feature/high_availability.md | 2 +-
.../feature/load_huggingface_config.md | 2 +-
.../docs/source_en/feature/logging.md | 4 +-
.../source_en/feature/memory_optimization.md | 6 +-
.../docs/source_en/feature/monitor.md | 2 +-
.../feature/other_training_features.md | 8 +-
.../source_en/feature/parallel_training.md | 14 +-
.../source_en/feature/pma_fused_checkpoint.md | 2 +-
.../docs/source_en/feature/quantization.md | 2 +-
.../docs/source_en/feature/resume_training.md | 2 +-
.../docs/source_en/feature/safetensors.md | 12 +-
.../skip_data_and_ckpt_health_monitor.md | 4 +-
.../docs/source_en/feature/start_tasks.md | 8 +-
.../docs/source_en/feature/tokenizer.md | 2 +-
.../feature/training_hyperparameters.md | 6 +-
.../docs/source_en/guide/deployment.md | 2 +-
.../docs/source_en/guide/evaluation.md | 2 +-
.../docs/source_en/guide/inference.md | 2 +-
.../docs/source_en/guide/pre_training.md | 2 +-
.../source_en/guide/supervised_fine_tuning.md | 6 +-
.../docs/source_en/installation.md | 2 +-
.../docs/source_en/introduction/models.md | 2 +-
.../docs/source_en/introduction/overview.md | 2 +-
.../accuracy_comparison.md | 4 +-
.../advanced_development/dev_migration.md | 2 +-
.../inference_precision_comparison.md | 4 +-
.../performance_optimization.md | 18 +-
.../precision_optimization.md | 8 +-
.../training_template_instruction.md | 6 +-
.../advanced_development/weight_transfer.md | 2 +-
.../yaml_config_inference.md | 2 +-
.../contribution/mindformers_contribution.md | 2 +-
.../contribution/modelers_contribution.md | 2 +-
.../docs/source_zh_cn/env_variables.md | 4 +-
.../convert_ckpt_to_megatron.md | 4 +-
.../example/distilled/distilled.md | 4 +-
.../docs/source_zh_cn/faq/feature_related.md | 2 +-
.../docs/source_zh_cn/faq/model_related.md | 2 +-
.../docs/source_zh_cn/feature/ckpt.md | 2 +-
.../source_zh_cn/feature/configuration.md | 20 +-
.../docs/source_zh_cn/feature/dataset.md | 10 +-
.../source_zh_cn/feature/high_availability.md | 2 +-
.../feature/load_huggingface_config.md | 2 +-
.../docs/source_zh_cn/feature/logging.md | 4 +-
.../feature/memory_optimization.md | 6 +-
.../docs/source_zh_cn/feature/monitor.md | 2 +-
.../feature/other_training_features.md | 8 +-
.../source_zh_cn/feature/parallel_training.md | 14 +-
.../feature/pma_fused_checkpoint.md | 2 +-
.../docs/source_zh_cn/feature/quantization.md | 2 +-
.../source_zh_cn/feature/resume_training.md | 2 +-
.../docs/source_zh_cn/feature/safetensors.md | 12 +-
.../skip_data_and_ckpt_health_monitor.md | 4 +-
.../docs/source_zh_cn/feature/start_tasks.md | 8 +-
.../docs/source_zh_cn/feature/tokenizer.md | 2 +-
.../feature/training_hyperparameters.md | 6 +-
.../docs/source_zh_cn/guide/deployment.md | 2 +-
.../docs/source_zh_cn/guide/evaluation.md | 2 +-
.../docs/source_zh_cn/guide/inference.md | 2 +-
.../docs/source_zh_cn/guide/llm_training.md | 2 +-
.../docs/source_zh_cn/guide/pre_training.md | 2 +-
.../guide/supervised_fine_tuning.md | 6 +-
.../docs/source_zh_cn/installation.md | 2 +-
.../docs/source_zh_cn/introduction/models.md | 2 +-
.../source_zh_cn/introduction/overview.md | 2 +-
.../api/api_cn/API_sample_and_requirements.md | 4 +-
.../source_en/api_python/bfloat16_support.md | 72 +-
.../api_python/dynamic_shape_func.md | 482 ++++----
.../source_en/api_python/dynamic_shape_nn.md | 114 +-
.../api_python/dynamic_shape_primitive.md | 428 +++----
.../source_en/api_python/env_var_list.rst | 26 +-
.../api_python/operator_list_parallel.md | 338 +++---
.../source_en/faq/data_processing.md | 22 +-
.../source_en/faq/distributed_parallel.md | 4 +-
.../mindspore/source_en/faq/feature_advice.md | 4 +-
.../source_en/faq/implement_problem.md | 16 +-
docs/mindspore/source_en/faq/inference.md | 2 +-
docs/mindspore/source_en/faq/installation.md | 2 +-
.../source_en/faq/network_compilation.md | 14 +-
.../source_en/faq/operators_compile.md | 22 +-
.../source_en/faq/performance_tuning.md | 2 +-
.../source_en/faq/precision_tuning.md | 2 +-
.../features/compile/compilation_guide.md | 14 +-
.../source_en/features/data_engine.md | 20 +-
docs/mindspore/source_en/features/overview.md | 6 +-
.../features/parallel/auto_parallel.md | 6 +-
.../features/parallel/data_parallel.md | 14 +-
.../features/parallel/operator_parallel.md | 18 +-
.../features/parallel/optimizer_parallel.md | 12 +-
.../features/parallel/pipeline_parallel.md | 24 +-
.../features/runtime/memory_manager.md | 14 +-
.../features/runtime/multilevel_pipeline.md | 6 +-
.../runtime/multistream_concurrency.md | 6 +-
.../note/api_mapping/pytorch_api_mapping.md | 1036 ++++++++---------
.../note/api_mapping/pytorch_diff/AGNEWS.md | 4 +-
.../pytorch_diff/AmazonReviewFull.md | 4 +-
.../pytorch_diff/AmazonReviewPolarity.md | 4 +-
.../api_mapping/pytorch_diff/AmplitudeToDB.md | 4 +-
.../note/api_mapping/pytorch_diff/CIFAR10.md | 4 +-
.../note/api_mapping/pytorch_diff/CIFAR100.md | 4 +-
.../api_mapping/pytorch_diff/CMUARCTIC.md | 4 +-
.../note/api_mapping/pytorch_diff/CelebA.md | 4 +-
.../api_mapping/pytorch_diff/Cityscapes.md | 4 +-
.../pytorch_diff/CoNLL2000Chunking.md | 4 +-
.../api_mapping/pytorch_diff/CocoDataset.md | 4 +-
.../note/api_mapping/pytorch_diff/DBpedia.md | 4 +-
.../api_mapping/pytorch_diff/DataLoader.md | 4 +-
.../pytorch_diff/DistributedSampler.md | 4 +-
.../pytorch_diff/FrequencyMasking.md | 4 +-
.../note/api_mapping/pytorch_diff/GTZAN.md | 4 +-
.../api_mapping/pytorch_diff/GriffinLim.md | 4 +-
.../note/api_mapping/pytorch_diff/IMDB.md | 4 +-
.../api_mapping/pytorch_diff/IWSLT2016.md | 4 +-
.../api_mapping/pytorch_diff/IWSLT2017.md | 4 +-
.../api_mapping/pytorch_diff/ImageFolder.md | 4 +-
.../pytorch_diff/InverseMelScale.md | 4 +-
.../note/api_mapping/pytorch_diff/LIBRITTS.md | 4 +-
.../note/api_mapping/pytorch_diff/LJSPEECH.md | 4 +-
.../note/api_mapping/pytorch_diff/Lookup.md | 4 +-
.../note/api_mapping/pytorch_diff/MNIST.md | 4 +-
.../note/api_mapping/pytorch_diff/MelScale.md | 4 +-
.../pytorch_diff/MelSpectrogram.md | 4 +-
.../note/api_mapping/pytorch_diff/Ngram.md | 4 +-
.../api_mapping/pytorch_diff/Normalize.md | 4 +-
.../api_mapping/pytorch_diff/PennTreebank.md | 4 +-
.../api_mapping/pytorch_diff/RandomAffine.md | 4 +-
.../pytorch_diff/RandomPerspective.md | 4 +-
.../pytorch_diff/RandomResizedCrop.md | 4 +-
.../pytorch_diff/RandomRotation.md | 4 +-
.../api_mapping/pytorch_diff/RandomSampler.md | 4 +-
.../api_mapping/pytorch_diff/RegexReplace.md | 4 +-
.../note/api_mapping/pytorch_diff/Resample.md | 4 +-
.../pytorch_diff/SPEECHCOMMANDS.md | 4 +-
.../note/api_mapping/pytorch_diff/SQuAD1.md | 4 +-
.../note/api_mapping/pytorch_diff/SQuAD2.md | 4 +-
.../SentencePieceTokenizer_Out_INT.md | 4 +-
.../SentencePieceTokenizer_Out_STRING.md | 4 +-
.../pytorch_diff/SequentialSampler.md | 4 +-
.../api_mapping/pytorch_diff/SogouNews.md | 4 +-
.../pytorch_diff/SpectralCentroid.md | 4 +-
.../api_mapping/pytorch_diff/Spectrogram.md | 4 +-
.../pytorch_diff/SubsetRandomSampler.md | 4 +-
.../note/api_mapping/pytorch_diff/TEDLIUM.md | 4 +-
.../api_mapping/pytorch_diff/TimeMasking.md | 4 +-
.../note/api_mapping/pytorch_diff/ToPIL.md | 4 +-
.../note/api_mapping/pytorch_diff/ToTensor.md | 4 +-
.../note/api_mapping/pytorch_diff/TypeCast.md | 4 +-
.../note/api_mapping/pytorch_diff/UDPOS.md | 4 +-
.../api_mapping/pytorch_diff/VOCDetection.md | 4 +-
.../pytorch_diff/VOCSegmentation.md | 4 +-
.../pytorch_diff/WeightedRandomSampler.md | 4 +-
.../pytorch_diff/WhitespaceTokenizer.md | 4 +-
.../api_mapping/pytorch_diff/WikiText103.md | 4 +-
.../api_mapping/pytorch_diff/WikiText2.md | 4 +-
.../note/api_mapping/pytorch_diff/YESNO.md | 4 +-
.../api_mapping/pytorch_diff/YahooAnswers.md | 4 +-
.../pytorch_diff/YelpReviewFull.md | 4 +-
.../pytorch_diff/YelpReviewPolarity.md | 4 +-
.../api_mapping/pytorch_diff/checkpoint.md | 4 +-
.../api_mapping/pytorch_diff/deform_conv2d.md | 4 +-
.../api_mapping/pytorch_diff/load_sp_model.md | 4 +-
.../note/api_mapping/pytorch_diff/nms.md | 4 +-
.../api_mapping/pytorch_diff/roi_align.md | 4 +-
.../api_python/bfloat16_support.md | 72 +-
.../api_python/dynamic_shape_func.md | 482 ++++----
.../api_python/dynamic_shape_nn.md | 114 +-
.../api_python/dynamic_shape_primitive.md | 428 +++----
.../source_zh_cn/api_python/env_var_list.rst | 28 +-
.../api_python/operator_list_parallel.md | 338 +++---
.../source_zh_cn/faq/data_processing.md | 20 +-
.../source_zh_cn/faq/distributed_parallel.md | 4 +-
.../source_zh_cn/faq/feature_advice.md | 4 +-
.../source_zh_cn/faq/implement_problem.md | 16 +-
docs/mindspore/source_zh_cn/faq/inference.md | 2 +-
.../source_zh_cn/faq/installation.md | 2 +-
.../source_zh_cn/faq/network_compilation.md | 14 +-
.../source_zh_cn/faq/operators_compile.md | 22 +-
.../source_zh_cn/faq/performance_tuning.md | 2 +-
.../source_zh_cn/faq/precision_tuning.md | 2 +-
docs/mindspore/source_zh_cn/features/amp.md | 4 +-
.../features/compile/compilation_guide.md | 14 +-
.../source_zh_cn/features/data_engine.md | 16 +-
docs/mindspore/source_zh_cn/features/mint.md | 6 +-
.../source_zh_cn/features/overview.md | 6 +-
.../features/parallel/auto_parallel.md | 6 +-
.../features/parallel/data_parallel.md | 14 +-
.../features/parallel/operator_parallel.md | 18 +-
.../features/parallel/optimizer_parallel.md | 8 +-
.../features/parallel/pipeline_parallel.md | 12 +-
.../features/runtime/memory_manager.md | 10 +-
.../features/runtime/multilevel_pipeline.md | 2 +-
.../runtime/multistream_concurrency.md | 2 +-
docs/mindspore/source_zh_cn/features/view.md | 2 +-
.../note/api_mapping/pytorch_api_mapping.md | 1036 ++++++++---------
.../note/api_mapping/pytorch_diff/AGNEWS.md | 4 +-
.../pytorch_diff/AmazonReviewFull.md | 4 +-
.../pytorch_diff/AmazonReviewPolarity.md | 4 +-
.../api_mapping/pytorch_diff/AmplitudeToDB.md | 4 +-
.../note/api_mapping/pytorch_diff/CIFAR10.md | 4 +-
.../note/api_mapping/pytorch_diff/CIFAR100.md | 4 +-
.../api_mapping/pytorch_diff/CMUARCTIC.md | 4 +-
.../note/api_mapping/pytorch_diff/CelebA.md | 4 +-
.../api_mapping/pytorch_diff/Cityscapes.md | 4 +-
.../pytorch_diff/CoNLL2000Chunking.md | 4 +-
.../api_mapping/pytorch_diff/CocoDataset.md | 4 +-
.../note/api_mapping/pytorch_diff/DBpedia.md | 4 +-
.../api_mapping/pytorch_diff/DataLoader.md | 4 +-
.../pytorch_diff/DistributedSampler.md | 4 +-
.../pytorch_diff/FrequencyMasking.md | 4 +-
.../note/api_mapping/pytorch_diff/GTZAN.md | 4 +-
.../api_mapping/pytorch_diff/GriffinLim.md | 4 +-
.../note/api_mapping/pytorch_diff/IMDB.md | 4 +-
.../api_mapping/pytorch_diff/IWSLT2016.md | 4 +-
.../api_mapping/pytorch_diff/IWSLT2017.md | 4 +-
.../api_mapping/pytorch_diff/ImageFolder.md | 4 +-
.../pytorch_diff/InverseMelScale.md | 4 +-
.../note/api_mapping/pytorch_diff/LIBRITTS.md | 4 +-
.../note/api_mapping/pytorch_diff/LJSPEECH.md | 4 +-
.../note/api_mapping/pytorch_diff/Lookup.md | 4 +-
.../note/api_mapping/pytorch_diff/MNIST.md | 4 +-
.../note/api_mapping/pytorch_diff/MelScale.md | 4 +-
.../pytorch_diff/MelSpectrogram.md | 4 +-
.../note/api_mapping/pytorch_diff/Ngram.md | 4 +-
.../api_mapping/pytorch_diff/Normalize.md | 4 +-
.../api_mapping/pytorch_diff/PennTreebank.md | 4 +-
.../api_mapping/pytorch_diff/RandomAffine.md | 4 +-
.../pytorch_diff/RandomPerspective.md | 4 +-
.../pytorch_diff/RandomResizedCrop.md | 4 +-
.../pytorch_diff/RandomRotation.md | 4 +-
.../api_mapping/pytorch_diff/RandomSampler.md | 4 +-
.../api_mapping/pytorch_diff/RegexReplace.md | 4 +-
.../note/api_mapping/pytorch_diff/Resample.md | 4 +-
.../pytorch_diff/SPEECHCOMMANDS.md | 4 +-
.../note/api_mapping/pytorch_diff/SQuAD1.md | 4 +-
.../note/api_mapping/pytorch_diff/SQuAD2.md | 4 +-
.../SentencePieceTokenizer_Out_INT.md | 4 +-
.../SentencePieceTokenizer_Out_STRING.md | 4 +-
.../pytorch_diff/SequentialSampler.md | 4 +-
.../api_mapping/pytorch_diff/SogouNews.md | 4 +-
.../pytorch_diff/SpectralCentroid.md | 4 +-
.../api_mapping/pytorch_diff/Spectrogram.md | 4 +-
.../pytorch_diff/SubsetRandomSampler.md | 4 +-
.../note/api_mapping/pytorch_diff/TEDLIUM.md | 4 +-
.../api_mapping/pytorch_diff/TimeMasking.md | 4 +-
.../note/api_mapping/pytorch_diff/ToPIL.md | 4 +-
.../note/api_mapping/pytorch_diff/ToTensor.md | 4 +-
.../note/api_mapping/pytorch_diff/TypeCast.md | 4 +-
.../note/api_mapping/pytorch_diff/UDPOS.md | 4 +-
.../api_mapping/pytorch_diff/VOCDetection.md | 4 +-
.../pytorch_diff/VOCSegmentation.md | 4 +-
.../pytorch_diff/WeightedRandomSampler.md | 4 +-
.../pytorch_diff/WhitespaceTokenizer.md | 4 +-
.../api_mapping/pytorch_diff/WikiText103.md | 4 +-
.../api_mapping/pytorch_diff/WikiText2.md | 4 +-
.../note/api_mapping/pytorch_diff/YESNO.md | 4 +-
.../api_mapping/pytorch_diff/YahooAnswers.md | 4 +-
.../pytorch_diff/YelpReviewFull.md | 4 +-
.../pytorch_diff/YelpReviewPolarity.md | 4 +-
.../api_mapping/pytorch_diff/checkpoint.md | 4 +-
.../api_mapping/pytorch_diff/deform_conv2d.md | 4 +-
.../api_mapping/pytorch_diff/load_sp_model.md | 4 +-
.../note/api_mapping/pytorch_diff/nms.md | 4 +-
.../api_mapping/pytorch_diff/roi_align.md | 4 +-
.../docs/source_zh_cn/feature/performance.md | 10 +-
.../docs/source_zh_cn/feature/precision.md | 2 +-
.../docs/source_zh_cn/guide/get_start.md | 2 +-
.../docs/source_zh_cn/guide/large_model.md | 2 +-
docs/mindstudio/docs/source_zh_cn/overview.md | 4 +-
.../version/mindstudio_insight.md | 2 +-
.../source_en/developer_guide/contributing.md | 2 +-
.../developer_guide/operations/custom_ops.md | 6 +-
.../docs/source_en/faqs/faqs.md | 2 +-
.../docs/source_en/general/security.md | 2 +-
.../installation/installation.md | 2 +-
.../quick_start/quick_start.md | 2 +-
.../deepseek_r1_671b_w8a8_dp4_tp4_ep4.md | 4 +-
.../qwen2.5_32b_multiNPU.md | 2 +-
.../qwen2.5_7b_singleNPU.md | 2 +-
.../source_en/release_notes/release_notes.md | 2 +-
.../environment_variables.md | 4 +-
.../supported_features/benchmark/benchmark.md | 2 +-
.../features_list/features_list.md | 2 +-
.../supported_features/parallel/parallel.md | 4 +-
.../supported_features/profiling/profiling.md | 6 +-
.../quantization/quantization.md | 2 +-
.../models_list/models_list.md | 2 +-
.../developer_guide/contributing.md | 2 +-
.../developer_guide/operations/custom_ops.md | 6 +-
.../docs/source_zh_cn/faqs/faqs.md | 2 +-
.../docs/source_zh_cn/general/security.md | 2 +-
.../installation/installation.md | 2 +-
.../quick_start/quick_start.md | 2 +-
.../deepseek_r1_671b_w8a8_dp4_tp4_ep4.md | 4 +-
.../qwen2.5_32b_multiNPU.md | 2 +-
.../qwen2.5_7b_singleNPU.md | 2 +-
.../release_notes/release_notes.md | 2 +-
.../environment_variables.md | 4 +-
.../supported_features/benchmark/benchmark.md | 2 +-
.../features_list/features_list.md | 2 +-
.../supported_features/parallel/parallel.md | 4 +-
.../supported_features/profiling/profiling.md | 6 +-
.../quantization/quantization.md | 2 +-
.../models_list/models_list.md | 2 +-
install/mindspore_ascend_install_conda.md | 2 +-
install/mindspore_ascend_install_conda_en.md | 2 +-
install/mindspore_ascend_install_docker.md | 2 +-
install/mindspore_ascend_install_docker_en.md | 2 +-
install/mindspore_ascend_install_pip.md | 2 +-
install/mindspore_ascend_install_pip_en.md | 2 +-
install/mindspore_ascend_install_source.md | 4 +-
install/mindspore_ascend_install_source_en.md | 26 +-
install/mindspore_cpu_install_conda.md | 2 +-
install/mindspore_cpu_install_conda_en.md | 2 +-
install/mindspore_cpu_install_pip.md | 2 +-
install/mindspore_cpu_install_pip_en.md | 2 +-
install/mindspore_cpu_install_source.md | 4 +-
install/mindspore_cpu_install_source_en.md | 4 +-
install/mindspore_cpu_mac_install_conda.md | 2 +-
install/mindspore_cpu_mac_install_conda_en.md | 2 +-
install/mindspore_cpu_mac_install_pip.md | 2 +-
install/mindspore_cpu_mac_install_pip_en.md | 2 +-
install/mindspore_cpu_mac_install_source.md | 4 +-
.../mindspore_cpu_mac_install_source_en.md | 4 +-
install/mindspore_cpu_win_install_conda.md | 2 +-
install/mindspore_cpu_win_install_conda_en.md | 2 +-
install/mindspore_cpu_win_install_pip.md | 2 +-
install/mindspore_cpu_win_install_pip_en.md | 2 +-
install/mindspore_cpu_win_install_source.md | 6 +-
.../mindspore_cpu_win_install_source_en.md | 6 +-
install/third_party/msys_software_install.md | 2 +-
.../third_party/msys_software_install_en.md | 2 +-
.../third_party/third_party_cpu_install.md | 4 +-
.../beginner/accelerate_with_static_graph.md | 12 +-
tutorials/source_en/beginner/autograd.md | 12 +-
tutorials/source_en/beginner/dataset.md | 26 +-
tutorials/source_en/beginner/introduction.md | 8 +-
tutorials/source_en/beginner/model.md | 20 +-
tutorials/source_en/beginner/quick_start.md | 22 +-
tutorials/source_en/beginner/save_load.md | 16 +-
tutorials/source_en/beginner/tensor.md | 20 +-
tutorials/source_en/beginner/train.md | 12 +-
tutorials/source_en/compile/fusion_pass.md | 2 +-
tutorials/source_en/compile/operators.md | 2 +-
.../compile/python_builtin_functions.md | 14 +-
tutorials/source_en/compile/statements.md | 4 +-
tutorials/source_en/compile/static_graph.md | 40 +-
.../static_graph_expert_programming.md | 12 +-
.../source_en/custom_program/hook_program.md | 14 +-
.../source_en/custom_program/op_custom.rst | 12 +-
.../operation/cpp_api_for_custom_ops.md | 6 +-
.../custom_program/operation/op_custom_adv.md | 4 +-
.../custom_program/operation/op_custom_aot.md | 10 +-
.../operation/op_custom_ascendc.md | 10 +-
.../operation/op_custom_prim.rst | 14 +-
.../operation/op_customopbuilder.md | 12 +-
.../operation/op_customopbuilder_aclnn.md | 6 +-
.../operation/op_customopbuilder_asdsip.md | 6 +-
.../operation/op_customopbuilder_atb.md | 6 +-
tutorials/source_en/cv/fcn8s.md | 14 +-
tutorials/source_en/cv/resnet50.md | 16 +-
tutorials/source_en/cv/ssd.md | 34 +-
tutorials/source_en/cv/transfer_learning.md | 4 +-
tutorials/source_en/cv/vit.md | 26 +-
tutorials/source_en/dataset/augment.md | 4 +-
tutorials/source_en/dataset/cache.md | 8 +-
.../source_en/dataset/dataset_autotune.md | 10 +-
.../source_en/dataset/dataset_offload.md | 2 +-
tutorials/source_en/dataset/eager.md | 18 +-
tutorials/source_en/dataset/optimize.ipynb | 38 +-
tutorials/source_en/dataset/overview.md | 52 +-
tutorials/source_en/dataset/python_objects.md | 2 +-
tutorials/source_en/dataset/record.ipynb | 12 +-
tutorials/source_en/dataset/sampler.md | 12 +-
tutorials/source_en/debug/dryrun.md | 6 +-
tutorials/source_en/debug/dump.md | 24 +-
tutorials/source_en/debug/error_analysis.rst | 24 +-
.../debug/error_analysis/cann_error_cases.md | 2 +-
.../error_analysis/error_scenario_analysis.md | 44 +-
.../debug/error_analysis/minddata_debug.md | 10 +-
.../source_en/debug/error_analysis/mindir.md | 2 +-
.../debug/error_analysis/mindrt_debug.md | 6 +-
tutorials/source_en/debug/profiler.md | 28 +-
tutorials/source_en/debug/pynative.md | 10 +-
tutorials/source_en/debug/sdc.md | 4 +-
tutorials/source_en/generative/cyclegan.md | 10 +-
tutorials/source_en/generative/dcgan.md | 12 +-
tutorials/source_en/generative/diffusion.md | 12 +-
tutorials/source_en/generative/gan.md | 12 +-
tutorials/source_en/generative/pix2pix.md | 10 +-
.../source_en/model_infer/introduction.md | 2 +-
.../model_infer/lite_infer/overview.md | 14 +-
.../ms_infer/ms_infer_model_infer.rst | 6 +-
.../ms_infer/ms_infer_model_serving_infer.md | 2 +-
.../ms_infer/ms_infer_network_develop.md | 6 +-
.../ms_infer/ms_infer_parallel_infer.md | 16 +-
.../ms_infer/ms_infer_quantization.md | 4 +-
.../model_migration/model_migration.md | 22 +-
tutorials/source_en/nlp/sentiment_analysis.md | 22 +-
tutorials/source_en/nlp/sequence_labeling.md | 2 +-
tutorials/source_en/orange_pi/dev_start.md | 16 +-
.../source_en/orange_pi/environment_setup.md | 34 +-
tutorials/source_en/orange_pi/model_infer.md | 12 +-
tutorials/source_en/orange_pi/overview.md | 8 +-
tutorials/source_en/parallel/comm_fusion.md | 12 +-
tutorials/source_en/parallel/data_parallel.md | 12 +-
tutorials/source_en/parallel/dataset_slice.md | 10 +-
.../source_en/parallel/distributed_case.rst | 4 +-
.../distributed_gradient_accumulation.md | 12 +-
.../source_en/parallel/dynamic_cluster.md | 14 +-
.../high_dimension_tensor_parallel.md | 16 +-
.../parallel/host_device_training.md | 18 +-
tutorials/source_en/parallel/mpirun.md | 4 +-
.../source_en/parallel/msrun_launcher.md | 20 +-
tutorials/source_en/parallel/multiple_copy.md | 14 +-
.../source_en/parallel/multiple_mixed.md | 6 +-
.../source_en/parallel/operator_parallel.md | 22 +-
.../source_en/parallel/optimize_technique.rst | 22 +-
.../source_en/parallel/optimizer_parallel.md | 8 +-
tutorials/source_en/parallel/overview.md | 32 +-
.../source_en/parallel/pipeline_parallel.md | 14 +-
tutorials/source_en/parallel/rank_table.md | 4 +-
tutorials/source_en/parallel/recompute.md | 16 +-
.../source_en/parallel/split_technique.md | 12 +-
.../source_en/parallel/startup_method.rst | 12 +-
.../source_en/parallel/strategy_select.md | 10 +-
.../train_availability/fault_recover.md | 6 +-
.../train_availability/graceful_exit.md | 6 +-
.../accelerate_with_static_graph.ipynb | 12 +-
.../source_zh_cn/beginner/autograd.ipynb | 12 +-
tutorials/source_zh_cn/beginner/dataset.ipynb | 28 +-
.../source_zh_cn/beginner/introduction.ipynb | 8 +-
tutorials/source_zh_cn/beginner/model.ipynb | 20 +-
.../source_zh_cn/beginner/quick_start.ipynb | 22 +-
.../source_zh_cn/beginner/save_load.ipynb | 20 +-
tutorials/source_zh_cn/beginner/tensor.ipynb | 22 +-
tutorials/source_zh_cn/beginner/train.ipynb | 12 +-
tutorials/source_zh_cn/compile/fusion_pass.md | 2 +-
tutorials/source_zh_cn/compile/operators.md | 2 +-
.../compile/python_builtin_functions.ipynb | 14 +-
.../source_zh_cn/compile/statements.ipynb | 4 +-
.../source_zh_cn/compile/static_graph.md | 38 +-
.../static_graph_expert_programming.ipynb | 16 +-
.../custom_program/hook_program.ipynb | 14 +-
.../source_zh_cn/custom_program/op_custom.rst | 10 +-
.../operation/cpp_api_for_custom_ops.md | 6 +-
.../operation/op_custom_adv.ipynb | 4 +-
.../custom_program/operation/op_custom_aot.md | 10 +-
.../operation/op_custom_ascendc.md | 10 +-
.../operation/op_custom_prim.ipynb | 12 +-
.../operation/op_customopbuilder.md | 12 +-
.../operation/op_customopbuilder_aclnn.md | 6 +-
.../operation/op_customopbuilder_asdsip.md | 6 +-
.../operation/op_customopbuilder_atb.md | 6 +-
tutorials/source_zh_cn/cv/fcn8s.ipynb | 14 +-
tutorials/source_zh_cn/cv/resnet50.ipynb | 16 +-
tutorials/source_zh_cn/cv/ssd.ipynb | 34 +-
.../source_zh_cn/cv/transfer_learning.ipynb | 6 +-
tutorials/source_zh_cn/cv/vit.ipynb | 26 +-
tutorials/source_zh_cn/dataset/augment.ipynb | 4 +-
tutorials/source_zh_cn/dataset/cache.ipynb | 12 +-
.../source_zh_cn/dataset/dataset_autotune.md | 10 +-
.../source_zh_cn/dataset/dataset_offload.md | 2 +-
tutorials/source_zh_cn/dataset/eager.ipynb | 20 +-
tutorials/source_zh_cn/dataset/optimize.ipynb | 38 +-
tutorials/source_zh_cn/dataset/overview.ipynb | 52 +-
.../source_zh_cn/dataset/python_objects.ipynb | 2 +-
tutorials/source_zh_cn/dataset/record.ipynb | 16 +-
tutorials/source_zh_cn/dataset/sampler.ipynb | 16 +-
tutorials/source_zh_cn/debug/dryrun.md | 6 +-
tutorials/source_zh_cn/debug/dump.md | 24 +-
.../source_zh_cn/debug/error_analysis.rst | 22 +-
.../debug/error_analysis/cann_error_cases.md | 2 +-
.../error_analysis/error_scenario_analysis.md | 38 +-
.../debug/error_analysis/minddata_debug.md | 10 +-
.../debug/error_analysis/mindir.md | 2 +-
.../debug/error_analysis/mindrt_debug.md | 6 +-
tutorials/source_zh_cn/debug/profiler.md | 28 +-
tutorials/source_zh_cn/debug/pynative.md | 10 +-
tutorials/source_zh_cn/debug/sdc.md | 4 +-
.../source_zh_cn/generative/cyclegan.ipynb | 10 +-
tutorials/source_zh_cn/generative/dcgan.ipynb | 12 +-
.../source_zh_cn/generative/diffusion.ipynb | 12 +-
tutorials/source_zh_cn/generative/gan.ipynb | 12 +-
.../source_zh_cn/generative/pix2pix.ipynb | 10 +-
.../source_zh_cn/model_infer/introduction.md | 2 +-
.../model_infer/lite_infer/overview.md | 14 +-
.../ms_infer/ms_infer_model_infer.rst | 6 +-
.../ms_infer/ms_infer_model_serving_infer.md | 2 +-
.../ms_infer/ms_infer_network_develop.md | 4 +-
.../ms_infer/ms_infer_parallel_infer.md | 4 +-
.../ms_infer/ms_infer_quantization.md | 4 +-
.../model_migration/model_migration.md | 20 +-
.../source_zh_cn/nlp/sentiment_analysis.ipynb | 22 +-
.../source_zh_cn/nlp/sequence_labeling.ipynb | 2 +-
.../source_zh_cn/orange_pi/dev_start.ipynb | 16 +-
.../orange_pi/environment_setup.md | 4 +-
.../source_zh_cn/orange_pi/model_infer.md | 6 +-
tutorials/source_zh_cn/orange_pi/overview.md | 8 +-
.../source_zh_cn/parallel/comm_fusion.md | 10 +-
.../source_zh_cn/parallel/data_parallel.md | 12 +-
.../source_zh_cn/parallel/dataset_slice.md | 10 +-
.../parallel/distributed_case.rst | 4 +-
.../distributed_gradient_accumulation.md | 12 +-
.../source_zh_cn/parallel/dynamic_cluster.md | 14 +-
.../high_dimension_tensor_parallel.md | 10 +-
.../parallel/host_device_training.md | 14 +-
tutorials/source_zh_cn/parallel/mpirun.md | 4 +-
.../source_zh_cn/parallel/msrun_launcher.md | 20 +-
.../source_zh_cn/parallel/multiple_copy.md | 12 +-
.../source_zh_cn/parallel/multiple_mixed.md | 6 +-
.../parallel/operator_parallel.md | 22 +-
.../parallel/optimize_technique.rst | 22 +-
.../parallel/optimizer_parallel.md | 8 +-
tutorials/source_zh_cn/parallel/overview.md | 32 +-
.../parallel/pipeline_parallel.md | 14 +-
tutorials/source_zh_cn/parallel/rank_table.md | 4 +-
tutorials/source_zh_cn/parallel/recompute.md | 12 +-
.../source_zh_cn/parallel/split_technique.md | 6 +-
.../source_zh_cn/parallel/startup_method.rst | 12 +-
.../source_zh_cn/parallel/strategy_select.md | 10 +-
.../train_availability/fault_recover.md | 6 +-
.../train_availability/graceful_exit.md | 6 +-
734 files changed, 6161 insertions(+), 6177 deletions(-)
create mode 100644 .jenkins/check/config/filter_notebooklint.txt
diff --git a/.jenkins/check/config/filter_notebooklint.txt b/.jenkins/check/config/filter_notebooklint.txt
new file mode 100644
index 0000000000..8d547b5b1b
--- /dev/null
+++ b/.jenkins/check/config/filter_notebooklint.txt
@@ -0,0 +1,2 @@
+"docs/tutorials/source_zh_cn"
+"docs/tutorials/source_en/dataset"
\ No newline at end of file
diff --git a/docs/lite/api/_custom/sphinx_builder_html b/docs/lite/api/_custom/sphinx_builder_html
index 3518c2e3c3..ca95bb1260 100644
--- a/docs/lite/api/_custom/sphinx_builder_html
+++ b/docs/lite/api/_custom/sphinx_builder_html
@@ -1116,7 +1116,7 @@ class StandaloneHTMLBuilder(Builder):
# Add links to the Python operator interface.
if "mindspore.ops." in output:
- output = re.sub(r'(mindspore\.ops\.\w+) ', r'\1 ', output, count=0)
+ output = re.sub(r'(mindspore\.ops\.\w+) ', r'\1 ', output, count=0)
except UnicodeError:
logger.warning(__("a Unicode error occurred when rendering the page %s. "
diff --git a/docs/lite/api/source_en/api_c/lite_c_example.rst b/docs/lite/api/source_en/api_c/lite_c_example.rst
index 877f7a048a..d7fb80b3e2 100644
--- a/docs/lite/api/source_en/api_c/lite_c_example.rst
+++ b/docs/lite/api/source_en/api_c/lite_c_example.rst
@@ -4,4 +4,4 @@ Example
.. toctree::
:maxdepth: 1
- Simple Demo↗
+ Simple Demo↗
diff --git a/docs/lite/api/source_en/api_cpp/lite_cpp_example.rst b/docs/lite/api/source_en/api_cpp/lite_cpp_example.rst
index f6ae1bd446..becd55cf71 100644
--- a/docs/lite/api/source_en/api_cpp/lite_cpp_example.rst
+++ b/docs/lite/api/source_en/api_cpp/lite_cpp_example.rst
@@ -4,6 +4,6 @@ Example
.. toctree::
:maxdepth: 1
- Simple Demo↗
- Android Application Development Based on JNI Interface↗
- High-level Usage↗
\ No newline at end of file
+ Simple Demo↗
+ Android Application Development Based on JNI Interface↗
+ High-level Usage↗
\ No newline at end of file
diff --git a/docs/lite/api/source_en/api_java/ascend_device_info.md b/docs/lite/api/source_en/api_java/ascend_device_info.md
index 3423232687..ddbe79e030 100644
--- a/docs/lite/api/source_en/api_java/ascend_device_info.md
+++ b/docs/lite/api/source_en/api_java/ascend_device_info.md
@@ -1,6 +1,6 @@
# AscendDeviceInfo
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_en/api_java/ascend_device_info.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_en/api_java/ascend_device_info.md)
```java
import com.mindspore.config.AscendDeviceInfo;
diff --git a/docs/lite/api/source_en/api_java/class_list.md b/docs/lite/api/source_en/api_java/class_list.md
index d6a0d1e1f8..9b471301f1 100644
--- a/docs/lite/api/source_en/api_java/class_list.md
+++ b/docs/lite/api/source_en/api_java/class_list.md
@@ -1,20 +1,20 @@
# Class List
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_en/api_java/class_list.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_en/api_java/class_list.md)
| Package | Class Name | Description | Supported At Cloud-side Inference | Supported At Device-side Inference |
| ------------------------- | ------------------------------------------------------------ | ------------------------------------------------------------ |--------|--------|
-| com.mindspore | [Model](https://www.mindspore.cn/lite/api/en/r2.7.1/api_java/model.html) | Model defines model in MindSpore for compiling and running compute graph. | √ | √ |
-| com.mindspore.config | [MSContext](https://www.mindspore.cn/lite/api/en/r2.7.1/api_java/mscontext.html) | MSContext is used to save the context during execution. | √ | √ |
-| com.mindspore | [MSTensor](https://www.mindspore.cn/lite/api/en/r2.7.1/api_java/mstensor.html) | MSTensor defines the tensor in MindSpore. | √ | √ |
-| com.mindspore | [ModelParallelRunner](https://www.mindspore.cn/lite/api/en/r2.7.1/api_java/model_parallel_runner.html) | Defines MindSpore Lite concurrent inference. | √ | ✕ |
-| com.mindspore.config | [RunnerConfig](https://www.mindspore.cn/lite/api/en/r2.7.1/api_java/runner_config.html) | RunnerConfig defines configuration parameters for concurrent inference. | √ | ✕ |
-| com.mindspore | [Graph](https://www.mindspore.cn/lite/api/en/r2.7.1/api_java/graph.html) | Graph defines the compute graph in MindSpore. | ✕ | √ |
-| com.mindspore.config | [CpuBindMode](https://www.mindspore.cn/lite/api/en/r2.7.1/api_java/mscontext.html#cpubindmode) | CpuBindMode defines the CPU binding mode. | √ | √ |
-| com.mindspore.config | [DeviceType](https://www.mindspore.cn/lite/api/en/r2.7.1/api_java/mscontext.html#devicetype) | DeviceType defines the back-end device type. | √ | √ |
-| com.mindspore.config | [DataType](https://www.mindspore.cn/lite/api/en/r2.7.1/api_java/mstensor.html#datatype) | DataType defines the supported data types. | √ | √ |
-| com.mindspore.config | [Version](https://www.mindspore.cn/lite/api/en/r2.7.1/api_java/version.html) | Version is used to obtain the version information of MindSpore. | ✕ | √ |
-| com.mindspore.config | [ModelType](https://www.mindspore.cn/lite/api/en/r2.7.1/api_java/model.html#modeltype) | ModelType defines the model file type. | √ | √ |
-| com.mindspore.config | [AscendDeviceInfo](https://www.mindspore.cn/lite/api/en/r2.7.1/api_java/ascend_device_info.html) | The AscendDeviceInfo class is used to configure MindSpore Lite Ascend device options. | √ | ✕ |
-| com.mindspore.config | [TrainCfg](https://www.mindspore.cn/lite/api/en/r2.7.1/api_java/train_cfg.html) | Configuration parameters used for model training on the device. | ✕ | √ |
+| com.mindspore | [Model](https://www.mindspore.cn/lite/api/en/r2.7.2/api_java/model.html) | Model defines model in MindSpore for compiling and running compute graph. | √ | √ |
+| com.mindspore.config | [MSContext](https://www.mindspore.cn/lite/api/en/r2.7.2/api_java/mscontext.html) | MSContext is used to save the context during execution. | √ | √ |
+| com.mindspore | [MSTensor](https://www.mindspore.cn/lite/api/en/r2.7.2/api_java/mstensor.html) | MSTensor defines the tensor in MindSpore. | √ | √ |
+| com.mindspore | [ModelParallelRunner](https://www.mindspore.cn/lite/api/en/r2.7.2/api_java/model_parallel_runner.html) | Defines MindSpore Lite concurrent inference. | √ | ✕ |
+| com.mindspore.config | [RunnerConfig](https://www.mindspore.cn/lite/api/en/r2.7.2/api_java/runner_config.html) | RunnerConfig defines configuration parameters for concurrent inference. | √ | ✕ |
+| com.mindspore | [Graph](https://www.mindspore.cn/lite/api/en/r2.7.2/api_java/graph.html) | Graph defines the compute graph in MindSpore. | ✕ | √ |
+| com.mindspore.config | [CpuBindMode](https://www.mindspore.cn/lite/api/en/r2.7.2/api_java/mscontext.html#cpubindmode) | CpuBindMode defines the CPU binding mode. | √ | √ |
+| com.mindspore.config | [DeviceType](https://www.mindspore.cn/lite/api/en/r2.7.2/api_java/mscontext.html#devicetype) | DeviceType defines the back-end device type. | √ | √ |
+| com.mindspore.config | [DataType](https://www.mindspore.cn/lite/api/en/r2.7.2/api_java/mstensor.html#datatype) | DataType defines the supported data types. | √ | √ |
+| com.mindspore.config | [Version](https://www.mindspore.cn/lite/api/en/r2.7.2/api_java/version.html) | Version is used to obtain the version information of MindSpore. | ✕ | √ |
+| com.mindspore.config | [ModelType](https://www.mindspore.cn/lite/api/en/r2.7.2/api_java/model.html#modeltype) | ModelType defines the model file type. | √ | √ |
+| com.mindspore.config | [AscendDeviceInfo](https://www.mindspore.cn/lite/api/en/r2.7.2/api_java/ascend_device_info.html) | The AscendDeviceInfo class is used to configure MindSpore Lite Ascend device options. | √ | ✕ |
+| com.mindspore.config | [TrainCfg](https://www.mindspore.cn/lite/api/en/r2.7.2/api_java/train_cfg.html) | Configuration parameters used for model training on the device. | ✕ | √ |
diff --git a/docs/lite/api/source_en/api_java/graph.md b/docs/lite/api/source_en/api_java/graph.md
index 0b2f7f37ff..5e9330d91f 100644
--- a/docs/lite/api/source_en/api_java/graph.md
+++ b/docs/lite/api/source_en/api_java/graph.md
@@ -1,6 +1,6 @@
# Graph
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_en/api_java/graph.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_en/api_java/graph.md)
```java
import com.mindspore.Graph;
diff --git a/docs/lite/api/source_en/api_java/lite_java_example.rst b/docs/lite/api/source_en/api_java/lite_java_example.rst
index 9699614ba8..2ab29dd0a1 100644
--- a/docs/lite/api/source_en/api_java/lite_java_example.rst
+++ b/docs/lite/api/source_en/api_java/lite_java_example.rst
@@ -4,6 +4,6 @@ Example
.. toctree::
:maxdepth: 1
- Simple Demo↗
- Android Application Development Based on Java Interface↗
- High-level Usage↗
\ No newline at end of file
+ Simple Demo↗
+ Android Application Development Based on Java Interface↗
+ High-level Usage↗
\ No newline at end of file
diff --git a/docs/lite/api/source_en/api_java/model.md b/docs/lite/api/source_en/api_java/model.md
index 6908478264..a66738ee85 100644
--- a/docs/lite/api/source_en/api_java/model.md
+++ b/docs/lite/api/source_en/api_java/model.md
@@ -1,6 +1,6 @@
# Model
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_en/api_java/model.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_en/api_java/model.md)
```java
import com.mindspore.model;
diff --git a/docs/lite/api/source_en/api_java/model_parallel_runner.md b/docs/lite/api/source_en/api_java/model_parallel_runner.md
index 5f0186749f..6c8a553437 100644
--- a/docs/lite/api/source_en/api_java/model_parallel_runner.md
+++ b/docs/lite/api/source_en/api_java/model_parallel_runner.md
@@ -1,6 +1,6 @@
# ModelParallelRunner
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_en/api_java/model_parallel_runner.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_en/api_java/model_parallel_runner.md)
```java
import com.mindspore.config.RunnerConfig;
diff --git a/docs/lite/api/source_en/api_java/mscontext.md b/docs/lite/api/source_en/api_java/mscontext.md
index 8370a865b8..81949c20f5 100644
--- a/docs/lite/api/source_en/api_java/mscontext.md
+++ b/docs/lite/api/source_en/api_java/mscontext.md
@@ -1,6 +1,6 @@
# MSContext
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_en/api_java/mscontext.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_en/api_java/mscontext.md)
```java
import com.mindspore.config.MSContext;
@@ -54,7 +54,7 @@ Initialize MSContext for cpu.
- Parameters
- `threadNum`: Thread number config for thread pool.
- - `cpuBindMode`: A **[CpuBindMode](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/java/src/main/java/com/mindspore/config/CpuBindMode.java)** **enum** variable.
+ - `cpuBindMode`: A **[CpuBindMode](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/java/src/main/java/com/mindspore/config/CpuBindMode.java)** **enum** variable.
- Returns
@@ -69,7 +69,7 @@ Initialize MSContext.
- Parameters
- `threadNum`: Thread number config for thread pool.
- - `cpuBindMode`: A **[CpuBindMode](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/java/src/main/java/com/mindspore/config/CpuBindMode.java)** **enum** variable.
+ - `cpuBindMode`: A **[CpuBindMode](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/java/src/main/java/com/mindspore/config/CpuBindMode.java)** **enum** variable.
- `isEnableParallel`: Whether to enable parallel in different device.
- Returns
@@ -86,7 +86,7 @@ Add device info for mscontext.
- Parameters
- - `deviceType`: A **[DeviceType](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/java/src/main/java/com/mindspore/config/DeviceType.java)** **enum** type.
+ - `deviceType`: A **[DeviceType](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/java/src/main/java/com/mindspore/config/DeviceType.java)** **enum** type.
- `isEnableFloat16`: Whether to enable fp16.
- Returns
@@ -101,7 +101,7 @@ Add device info for mscontext.
- Parameters
- - `deviceType`: A **[DeviceType](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/java/src/main/java/com/mindspore/config/DeviceType.java)** **enum** type.
+ - `deviceType`: A **[DeviceType](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/java/src/main/java/com/mindspore/config/DeviceType.java)** **enum** type.
- `isEnableFloat16`: is enable fp16.
- `npuFreq`: Npu frequency.
diff --git a/docs/lite/api/source_en/api_java/mstensor.md b/docs/lite/api/source_en/api_java/mstensor.md
index 7b06075e1d..3234d84bc7 100644
--- a/docs/lite/api/source_en/api_java/mstensor.md
+++ b/docs/lite/api/source_en/api_java/mstensor.md
@@ -1,6 +1,6 @@
# MSTensor
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_en/api_java/mstensor.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_en/api_java/mstensor.md)
```java
import com.mindspore.MSTensor;
@@ -86,7 +86,7 @@ Get the shape of the MindSpore MSTensor.
public int getDataType()
```
-DataType is defined in [com.mindspore.DataType](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/java/src/main/java/com/mindspore/config/DataType.java).
+DataType is defined in [com.mindspore.DataType](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/java/src/main/java/com/mindspore/config/DataType.java).
- Returns
diff --git a/docs/lite/api/source_en/api_java/runner_config.md b/docs/lite/api/source_en/api_java/runner_config.md
index f21c1f0986..c5c8f16b66 100644
--- a/docs/lite/api/source_en/api_java/runner_config.md
+++ b/docs/lite/api/source_en/api_java/runner_config.md
@@ -1,6 +1,6 @@
# RunnerConfig
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_en/api_java/runner_config.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_en/api_java/runner_config.md)
RunnerConfig defines the configuration parameters of MindSpore Lite concurrent inference.
diff --git a/docs/lite/api/source_en/api_java/train_cfg.md b/docs/lite/api/source_en/api_java/train_cfg.md
index 1c16a767e3..8980944bea 100644
--- a/docs/lite/api/source_en/api_java/train_cfg.md
+++ b/docs/lite/api/source_en/api_java/train_cfg.md
@@ -1,6 +1,6 @@
# TrainCfg
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_en/api_java/train_cfg.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_en/api_java/train_cfg.md)
```java
import com.mindspore.config.TrainCfg;
diff --git a/docs/lite/api/source_en/api_java/version.md b/docs/lite/api/source_en/api_java/version.md
index 32299b70f8..72b78c1f7c 100644
--- a/docs/lite/api/source_en/api_java/version.md
+++ b/docs/lite/api/source_en/api_java/version.md
@@ -1,6 +1,6 @@
# Version
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_en/api_java/version.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_en/api_java/version.md)
```java
import com.mindspore.config.Version;
diff --git a/docs/lite/api/source_en/index.rst b/docs/lite/api/source_en/index.rst
index cebb66ccb0..404b742f39 100644
--- a/docs/lite/api/source_en/index.rst
+++ b/docs/lite/api/source_en/index.rst
@@ -12,21 +12,21 @@ Summary of MindSpore Lite API support
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Class | Description | C++ API | Python API |
+=========================================================+===================================================================================================================================+==========================================================================================================================================================================================================================+============================================================================================================================================================================================================================================================================================================================================================================+
-| Context | Set the number of threads at runtime | void SetThreadNum(int32_t thread_num) | `Context.cpu.thread_num `__ |
+| Context | Set the number of threads at runtime | void SetThreadNum(int32_t thread_num) | `Context.cpu.thread_num `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Context | Get the current thread number setting | int32_t GetThreadNum() const | `Context.cpu.thread_num `__ |
+| Context | Get the current thread number setting | int32_t GetThreadNum() const | `Context.cpu.thread_num `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Context | Set the parallel number of operators at runtime | void SetInterOpParallelNum(int32_t parallel_num) | `Context.cpu.inter_op_parallel_num `__ |
+| Context | Set the parallel number of operators at runtime | void SetInterOpParallelNum(int32_t parallel_num) | `Context.cpu.inter_op_parallel_num `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Context | Get the current operators parallel number setting | int32_t GetInterOpParallelNum() const | `Context.cpu.inter_op_parallel_num `__ |
+| Context | Get the current operators parallel number setting | int32_t GetInterOpParallelNum() const | `Context.cpu.inter_op_parallel_num `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Context | Set the thread affinity to CPU cores | void SetThreadAffinity(int mode) | `Context.cpu.thread_affinity_mode `__ |
+| Context | Set the thread affinity to CPU cores | void SetThreadAffinity(int mode) | `Context.cpu.thread_affinity_mode `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Context | Get the thread affinity of CPU cores | int GetThreadAffinityMode() const | `Context.cpu.thread_affinity_mode `__ |
+| Context | Get the thread affinity of CPU cores | int GetThreadAffinityMode() const | `Context.cpu.thread_affinity_mode `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Context | Set the thread lists to CPU cores | void SetThreadAffinity(const std::vector &core_list) | `Context.cpu.thread_affinity_core_list `__ |
+| Context | Set the thread lists to CPU cores | void SetThreadAffinity(const std::vector &core_list) | `Context.cpu.thread_affinity_core_list `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Context | Get the thread lists of CPU cores | std::vector GetThreadAffinityCoreList() const | `Context.cpu.thread_affinity_core_list `__ |
+| Context | Get the thread lists of CPU cores | std::vector GetThreadAffinityCoreList() const | `Context.cpu.thread_affinity_core_list `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Context | Set the status whether to perform model inference or training in parallel | void SetEnableParallel(bool is_parallel) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
@@ -44,7 +44,7 @@ Summary of MindSpore Lite API support
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Context | Get the mode of the model run | bool GetMultiModalHW() const | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Context | Get a mutable reference of DeviceInfoContext vector in this context | std::vector> &MutableDeviceInfo() | Wrapped in `Context.target `__ |
+| Context | Get a mutable reference of DeviceInfoContext vector in this context | std::vector> &MutableDeviceInfo() | Wrapped in `Context.target `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| DeviceInfoContext | Get the type of this DeviceInfoContext | enum DeviceType GetDeviceType() const | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
@@ -62,29 +62,29 @@ Summary of MindSpore Lite API support
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| DeviceInfoContext | obtain memory allocator | std::shared_ptr GetAllocator() const | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| CPUDeviceInfo | Get the type of this DeviceInfoContext | enum DeviceType GetDeviceType() const | `context.cpu `__ |
+| CPUDeviceInfo | Get the type of this DeviceInfoContext | enum DeviceType GetDeviceType() const | `context.cpu `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| CPUDeviceInfo | Set enables to perform the float16 inference | void SetEnableFP16(bool is_fp16) | `Context.cpu.precision_mode `__ |
+| CPUDeviceInfo | Set enables to perform the float16 inference | void SetEnableFP16(bool is_fp16) | `Context.cpu.precision_mode `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| CPUDeviceInfo | Get enables to perform the float16 inference | bool GetEnableFP16() const | `Context.cpu.precision_mode `__ |
+| CPUDeviceInfo | Get enables to perform the float16 inference | bool GetEnableFP16() const | `Context.cpu.precision_mode `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| GPUDeviceInfo | Get the type of this DeviceInfoContext | enum DeviceType GetDeviceType() const | `Context.gpu `__ |
+| GPUDeviceInfo | Get the type of this DeviceInfoContext | enum DeviceType GetDeviceType() const | `Context.gpu `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| GPUDeviceInfo | Set device id | void SetDeviceID(uint32_t device_id) | `Context.gpu.device_id `__ |
+| GPUDeviceInfo | Set device id | void SetDeviceID(uint32_t device_id) | `Context.gpu.device_id `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| GPUDeviceInfo | Get the device id | uint32_t GetDeviceID() const | `Context.gpu.device_id `__ |
+| GPUDeviceInfo | Get the device id | uint32_t GetDeviceID() const | `Context.gpu.device_id `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| GPUDeviceInfo | Get the distribution rank id | int GetRankID() const | `Context.gpu.rank_id `__ |
+| GPUDeviceInfo | Get the distribution rank id | int GetRankID() const | `Context.gpu.rank_id `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| GPUDeviceInfo | Get the distribution group size | int GetGroupSize() const | `Context.gpu.group_size `__ |
+| GPUDeviceInfo | Get the distribution group size | int GetGroupSize() const | `Context.gpu.group_size `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| GPUDeviceInfo | Set the precision mode | void SetPrecisionMode(const std::string &precision_mode) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| GPUDeviceInfo | Get the precision mode | std::string GetPrecisionMode() const | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| GPUDeviceInfo | Set enables to perform the float16 inference | void SetEnableFP16(bool is_fp16) | `Context.gpu.precision_mode `__ |
+| GPUDeviceInfo | Set enables to perform the float16 inference | void SetEnableFP16(bool is_fp16) | `Context.gpu.precision_mode `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| GPUDeviceInfo | Get enables to perform the float16 inference | bool GetEnableFP16() const | `Context.gpu.precision_mode `__ |
+| GPUDeviceInfo | Get enables to perform the float16 inference | bool GetEnableFP16() const | `Context.gpu.precision_mode `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| GPUDeviceInfo | Set enables to sharing mem with OpenGL | void SetEnableGLTexture(bool is_enable_gl_texture) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
@@ -98,11 +98,11 @@ Summary of MindSpore Lite API support
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| GPUDeviceInfo | Get current OpenGL display | void \*GetGLDisplay() const | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| AscendDeviceInfo | Get the type of this DeviceInfoContext | enum DeviceType GetDeviceType() const | `Context.ascend `__ |
+| AscendDeviceInfo | Get the type of this DeviceInfoContext | enum DeviceType GetDeviceType() const | `Context.ascend `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| AscendDeviceInfo | Set device id | void SetDeviceID(uint32_t device_id) | `Context.ascend.device_id `__ |
+| AscendDeviceInfo | Set device id | void SetDeviceID(uint32_t device_id) | `Context.ascend.device_id `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| AscendDeviceInfo | Get the device id | uint32_t GetDeviceID() const | `Context.ascend.device_id `__ |
+| AscendDeviceInfo | Get the device id | uint32_t GetDeviceID() const | `Context.ascend.device_id `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| AscendDeviceInfo | Set AIPP configuration file path | void SetInsertOpConfigPath(const std::string &cfg_path) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
@@ -132,9 +132,9 @@ Summary of MindSpore Lite API support
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| AscendDeviceInfo | Get type of model outputs | enum DataType GetOutputType() const | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| AscendDeviceInfo | Set precision mode of model | void SetPrecisionMode(const std::string &precision_mode) | `Context.ascend.precision_mode `__ |
+| AscendDeviceInfo | Set precision mode of model | void SetPrecisionMode(const std::string &precision_mode) | `Context.ascend.precision_mode `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| AscendDeviceInfo | Get precision mode of model | std::string GetPrecisionMode() const | `Context.ascend.precision_mode `__ |
+| AscendDeviceInfo | Get precision mode of model | std::string GetPrecisionMode() const | `Context.ascend.precision_mode `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| AscendDeviceInfo | Set op select implementation mode | void SetOpSelectImplMode(const std::string &op_select_impl_mode) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
@@ -160,7 +160,7 @@ Summary of MindSpore Lite API support
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Model | Build a model from model buffer so that it can run on a device | Status Build(const void \*model_data, size_t data_size, ModelType model_type, const std::shared_ptr &model_context = nullptr) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Model | Load and build a model from model buffer so that it can run on a device | Status Build(const std::string &model_path, ModelType model_type, const std::shared_ptr &model_context = nullptr) | `Model.build_from_file `__ |
+| Model | Load and build a model from model buffer so that it can run on a device | Status Build(const std::string &model_path, ModelType model_type, const std::shared_ptr &model_context = nullptr) | `Model.build_from_file `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Model | Build a model from model buffer so that it can run on a device | Status Build(const void \*model_data, size_t data_size, ModelType model_type, const std::shared_ptr &model_context, const Key &dec_key, const std::string &dec_mode, const std::string &cropto_lib_path) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
@@ -172,11 +172,11 @@ Summary of MindSpore Lite API support
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Model | Build a Transfer Learning model where the backbone weights are fixed and the head weights are trainable | Status BuildTransferLearning(GraphCell backbone, GraphCell head, const std::shared_ptr &context, const std::shared_ptr &train_cfg = nullptr) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Model | Resize the shapes of inputs | Status Resize(const std::vector &inputs, const std::vector > &dims) | `Model.resize `__ |
+| Model | Resize the shapes of inputs | Status Resize(const std::vector &inputs, const std::vector > &dims) | `Model.resize `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Model | Change the size and or content of weight tensors | Status UpdateWeights(const std::vector &new_weights) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Model | Inference model API | Status Predict(const std::vector &inputs, std::vector \*outputs, const MSKernelCallBack &before = nullptr, const MSKernelCallBack &after = nullptr) | `Model.predict `__ |
+| Model | Inference model API | Status Predict(const std::vector &inputs, std::vector \*outputs, const MSKernelCallBack &before = nullptr, const MSKernelCallBack &after = nullptr) | `Model.predict `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Model | Inference model API only with callback | Status Predict(const MSKernelCallBack &before = nullptr, const MSKernelCallBack &after = nullptr) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
@@ -188,11 +188,11 @@ Summary of MindSpore Lite API support
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Model | Check if data preprocess exists in model | bool HasPreprocess() | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Model | Load config file | Status LoadConfig(const std::string &config_path) | Wrapped in the parameter `config_path` of `Model.build_from_file `__ |
+| Model | Load config file | Status LoadConfig(const std::string &config_path) | Wrapped in the parameter `config_path` of `Model.build_from_file `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Model | Update config | Status UpdateConfig(const std::string §ion, const std::pair &config) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Model | Obtains all input tensors of the model | std::vector GetInputs() | `Model.get_inputs `__ |
+| Model | Obtains all input tensors of the model | std::vector GetInputs() | `Model.get_inputs `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Model | Obtains the input tensor of the model by name | MSTensor GetInputByTensorName(const std::string &tensor_name) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
@@ -220,7 +220,7 @@ Summary of MindSpore Lite API support
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Model | Accessor to TrainLoop metric objects | std::vector GetMetrics() | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| Model | Obtains all output tensors of the model | std::vector GetOutputs() | Wrapped in the return value of `Model.predict `__ |
+| Model | Obtains all output tensors of the model | std::vector GetOutputs() | Wrapped in the return value of `Model.predict `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Model | Obtains names of all output tensors of the model | std::vector GetOutputTensorNames() | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
@@ -240,33 +240,33 @@ Summary of MindSpore Lite API support
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Model | Check if the device supports the model | static bool CheckModelSupport(enum DeviceType device_type, ModelType model_type) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| RunnerConfig | Set the number of workers at runtime | void SetWorkersNum(int32_t workers_num) | `Context.parallel.workers_num `__ |
+| RunnerConfig | Set the number of workers at runtime | void SetWorkersNum(int32_t workers_num) | `Context.parallel.workers_num `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| RunnerConfig | Get the current operators parallel workers number setting | int32_t GetWorkersNum() const | `Context.parallel.workers_num `__ |
+| RunnerConfig | Get the current operators parallel workers number setting | int32_t GetWorkersNum() const | `Context.parallel.workers_num `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| RunnerConfig | Set the context at runtime | void SetContext(const std::shared_ptr &context) | Wrapped in `Context.parallel `__ |
+| RunnerConfig | Set the context at runtime | void SetContext(const std::shared_ptr &context) | Wrapped in `Context.parallel `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| RunnerConfig | Get the current context setting | std::shared_ptr GetContext() const | Wrapped in `Context.parallel `__ |
+| RunnerConfig | Get the current context setting | std::shared_ptr GetContext() const | Wrapped in `Context.parallel `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| RunnerConfig | Set the config before runtime | void SetConfigInfo(const std::string §ion, const std::map &config) | `Context.parallel.config_info `__ |
+| RunnerConfig | Set the config before runtime | void SetConfigInfo(const std::string §ion, const std::map &config) | `Context.parallel.config_info `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| RunnerConfig | Get the current config setting | std::map> GetConfigInfo() const | `Context.parallel.config_info `__ |
+| RunnerConfig | Get the current config setting | std::map> GetConfigInfo() const | `Context.parallel.config_info `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| RunnerConfig | Set the config path before runtime | void SetConfigPath(const std::string &config_path) | `Context.parallel.config_path `__ |
+| RunnerConfig | Set the config path before runtime | void SetConfigPath(const std::string &config_path) | `Context.parallel.config_path `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| RunnerConfig | Get the current config path | std::string GetConfigPath() const | `Context.parallel.config_path `__ |
+| RunnerConfig | Get the current config path | std::string GetConfigPath() const | `Context.parallel.config_path `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| ModelParallelRunner | build a model parallel runner from model path so that it can run on a device | Status Init(const std::string &model_path, const std::shared_ptr &runner_config = nullptr) | `Model.parallel_runner.build_from_file `__ |
+| ModelParallelRunner | build a model parallel runner from model path so that it can run on a device | Status Init(const std::string &model_path, const std::shared_ptr &runner_config = nullptr) | `Model.parallel_runner.build_from_file `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| ModelParallelRunner | build a model parallel runner from model buffer so that it can run on a device | Status Init(const void \*model_data, const size_t data_size, const std::shared_ptr &runner_config = nullptr) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| ModelParallelRunner | Obtains all input tensors information of the model | std::vector GetInputs() | `Model.parallel_runner.get_inputs `__ |
+| ModelParallelRunner | Obtains all input tensors information of the model | std::vector GetInputs() | `Model.parallel_runner.get_inputs `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| ModelParallelRunner | Obtains all output tensors information of the model | std::vector GetOutputs() | Wrapped in the return value of `Model.parallel_runner.predict `__ |
+| ModelParallelRunner | Obtains all output tensors information of the model | std::vector GetOutputs() | Wrapped in the return value of `Model.parallel_runner.predict `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| ModelParallelRunner | Inference ModelParallelRunner | Status Predict(const std::vector &inputs, std::vector \*outputs,const MSKernelCallBack &before = nullptr, const MSKernelCallBack &after = nullptr) | `Model.parallel_runner.predict `__ |
+| ModelParallelRunner | Inference ModelParallelRunner | Status Predict(const std::vector &inputs, std::vector \*outputs,const MSKernelCallBack &before = nullptr, const MSKernelCallBack &after = nullptr) | `Model.parallel_runner.predict `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| MSTensor | Creates a MSTensor object, whose data need to be copied before accessed by Model | static inline MSTensor \*CreateTensor(const std::string &name, DataType type, const std::vector &shape, const void \*data, size_t data_len) noexcept | `Tensor `__ |
+| MSTensor | Creates a MSTensor object, whose data need to be copied before accessed by Model | static inline MSTensor \*CreateTensor(const std::string &name, DataType type, const std::vector &shape, const void \*data, size_t data_len) noexcept | `Tensor `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| MSTensor | Creates a MSTensor object, whose data can be directly accessed by Model | static inline MSTensor \*CreateRefTensor(const std::string &name, DataType type, const std::vector &shape, const void \*data, size_t data_len, bool own_data = true) noexcept | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
@@ -280,19 +280,19 @@ Summary of MindSpore Lite API support
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| MSTensor | Destroy an object created by `Clone` , `StringsToTensor` , `CreateRefTensor` or `CreateTensor` | static void DestroyTensorPtr(MSTensor \*tensor) noexcept | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| MSTensor | Obtains the name of the MSTensor | std::string Name() const | `Tensor.name `__ |
+| MSTensor | Obtains the name of the MSTensor | std::string Name() const | `Tensor.name `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| MSTensor | Obtains the data type of the MSTensor | enum DataType DataType() const | `Tensor.dtype `__ |
+| MSTensor | Obtains the data type of the MSTensor | enum DataType DataType() const | `Tensor.dtype `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| MSTensor | Obtains the shape of the MSTensor | const std::vector &Shape() const | `Tensor.shape `__ |
+| MSTensor | Obtains the shape of the MSTensor | const std::vector &Shape() const | `Tensor.shape `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| MSTensor | Obtains the number of elements of the MSTensor | int64_t ElementNum() const | `Tensor.element_num `__ |
+| MSTensor | Obtains the number of elements of the MSTensor | int64_t ElementNum() const | `Tensor.element_num `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| MSTensor | Obtains a shared pointer to the copy of data of the MSTensor | std::shared_ptr Data() const | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| MSTensor | Obtains the pointer to the data of the MSTensor | void \*MutableData() | Wrapped in `Tensor.get_data_to_numpy `__ and `Tensor.set_data_from_numpy `__ |
+| MSTensor | Obtains the pointer to the data of the MSTensor | void \*MutableData() | Wrapped in `Tensor.get_data_to_numpy `__ and `Tensor.set_data_from_numpy `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| MSTensor | Obtains the length of the data of the MSTensor, in bytes | size_t DataSize() const | `Tensor.data_size `__ |
+| MSTensor | Obtains the length of the data of the MSTensor, in bytes | size_t DataSize() const | `Tensor.data_size `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| MSTensor | Get whether the MSTensor data is const data | bool IsConst() const | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
@@ -308,19 +308,19 @@ Summary of MindSpore Lite API support
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| MSTensor | Get the boolean value that indicates whether the MSTensor not equals tensor | bool operator!=(const MSTensor &tensor) const | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| MSTensor | Set the shape of for the MSTensor | void SetShape(const std::vector &shape) | `Tensor.shape `__ |
+| MSTensor | Set the shape of for the MSTensor | void SetShape(const std::vector &shape) | `Tensor.shape `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| MSTensor | Set the data type for the MSTensor | void SetDataType(enum DataType data_type) | `Tensor.dtype `__ |
+| MSTensor | Set the data type for the MSTensor | void SetDataType(enum DataType data_type) | `Tensor.dtype `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| MSTensor | Set the name for the MSTensor | void SetTensorName(const std::string &name) | `Tensor.name `__ |
+| MSTensor | Set the name for the MSTensor | void SetTensorName(const std::string &name) | `Tensor.name `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| MSTensor | Set the Allocator for the MSTensor | void SetAllocator(std::shared_ptr allocator) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| MSTensor | Obtain the Allocator of the MSTensor | std::shared_ptr allocator() const | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| MSTensor | Set the format for the MSTensor | void SetFormat(mindspore::Format format) | `Tensor.format `__ |
+| MSTensor | Set the format for the MSTensor | void SetFormat(mindspore::Format format) | `Tensor.format `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| MSTensor | Obtain the format of the MSTensor | mindspore::Format format() const | `Tensor.format `__ |
+| MSTensor | Obtain the format of the MSTensor | mindspore::Format format() const | `Tensor.format `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| MSTensor | Set the data for the MSTensor | void SetData(void \*data, bool own_data = true) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
@@ -332,15 +332,15 @@ Summary of MindSpore Lite API support
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| MSTensor | Set the quantization parameters for the MSTensor | void SetQuantParams(std::vector quant_params) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| ModelGroup | Construct a ModelGroup object and indicate shared workspace memory or shared weight memory, with default shared workspace memory | ModelGroup(ModelGroupFlag flags = ModelGroupFlag::kShareWorkspace) | `ModelGroup `__ |
+| ModelGroup | Construct a ModelGroup object and indicate shared workspace memory or shared weight memory, with default shared workspace memory | ModelGroup(ModelGroupFlag flags = ModelGroupFlag::kShareWorkspace) | `ModelGroup `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| ModelGroup | When sharing weight memory, add model objects that require shared weight memory | Status AddModel(const std::vector &model_list) | `ModelGroup.add_model `__ |
+| ModelGroup | When sharing weight memory, add model objects that require shared weight memory | Status AddModel(const std::vector &model_list) | `ModelGroup.add_model `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| ModelGroup | When sharing workspace memory, add the path of the model that requires shared workspace memory | Status AddModel(const std::vector &model_path_list) | `ModelGroup.add_model `__ |
+| ModelGroup | When sharing workspace memory, add the path of the model that requires shared workspace memory | Status AddModel(const std::vector &model_path_list) | `ModelGroup.add_model `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| ModelGroup | When sharing workspace memory, add a model buffer that requires shared workspace memory | Status AddModel(const std::vector> &model_buff_list) | |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
-| ModelGroup | When sharing workspace memory, calculate the maximum workspace memory size | Status CalMaxSizeOfWorkspace(ModelType model_type, const std::shared_ptr &ms_context) | `ModelGroup.cal_max_size_of_workspace `__ |
+| ModelGroup | When sharing workspace memory, calculate the maximum workspace memory size | Status CalMaxSizeOfWorkspace(ModelType model_type, const std::shared_ptr &ms_context) | `ModelGroup.cal_max_size_of_workspace `__ |
+---------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
diff --git a/docs/lite/api/source_zh_cn/api_c/context_c.md b/docs/lite/api/source_zh_cn/api_c/context_c.md
index 47e823ce6d..132488fa83 100644
--- a/docs/lite/api/source_zh_cn/api_c/context_c.md
+++ b/docs/lite/api/source_zh_cn/api_c/context_c.md
@@ -1,6 +1,6 @@
# context_c
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_c/context_c.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_c/context_c.md)
```c
#include
@@ -198,7 +198,7 @@ MSDeviceInfoHandle MSDeviceInfoCreate(MSDeviceType device_type)
新建运行设备信息,若创建失败则会返回`nullptr`,并在日志中输出信息。
- 参数
- - `device_type`: 设备类型,具体见[MSDeviceType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_c/types_c.html#msdevicetype)。
+ - `device_type`: 设备类型,具体见[MSDeviceType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_c/types_c.html#msdevicetype)。
- 返回值
diff --git a/docs/lite/api/source_zh_cn/api_c/data_type_c.md b/docs/lite/api/source_zh_cn/api_c/data_type_c.md
index 7d66a2db36..c4678a378f 100644
--- a/docs/lite/api/source_zh_cn/api_c/data_type_c.md
+++ b/docs/lite/api/source_zh_cn/api_c/data_type_c.md
@@ -1,6 +1,6 @@
# data_type_c
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_c/data_type_c.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_c/data_type_c.md)
```C
#include
diff --git a/docs/lite/api/source_zh_cn/api_c/format_c.md b/docs/lite/api/source_zh_cn/api_c/format_c.md
index 29b7dc2bef..82f5351800 100644
--- a/docs/lite/api/source_zh_cn/api_c/format_c.md
+++ b/docs/lite/api/source_zh_cn/api_c/format_c.md
@@ -1,6 +1,6 @@
# format_c
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_c/format_c.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_c/format_c.md)
```C
#include
diff --git a/docs/lite/api/source_zh_cn/api_c/lite_c_example.rst b/docs/lite/api/source_zh_cn/api_c/lite_c_example.rst
index 32a216e74a..d7741e3994 100644
--- a/docs/lite/api/source_zh_cn/api_c/lite_c_example.rst
+++ b/docs/lite/api/source_zh_cn/api_c/lite_c_example.rst
@@ -4,4 +4,4 @@
.. toctree::
:maxdepth: 1
- 极简Demo↗
+ 极简Demo↗
diff --git a/docs/lite/api/source_zh_cn/api_c/model_c.md b/docs/lite/api/source_zh_cn/api_c/model_c.md
index b1e17721fc..acc48d1ae5 100644
--- a/docs/lite/api/source_zh_cn/api_c/model_c.md
+++ b/docs/lite/api/source_zh_cn/api_c/model_c.md
@@ -1,6 +1,6 @@
# model_c
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_c/model_c.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_c/model_c.md)
```C
#include
@@ -145,8 +145,8 @@ MSStatus MSModelBuild(MSModelHandle model, const void* model_data, size_t data_s
- `model`: 指向模型对象的指针。
- `model_data`: 内存中已经加载的模型数据地址。
- `data_size`: 模型数据的长度。
- - `model_type`: 模型文件类型,具体见: [MSModelType](https://mindspore.cn/lite/api/zh-CN/r2.7.1/api_c/types_c.html#msmodeltype)。
- - `model_context`: 模型的上下文环境,具体见: [Context](https://mindspore.cn/lite/api/zh-CN/r2.7.1/api_c/context_c.html)。
+ - `model_type`: 模型文件类型,具体见: [MSModelType](https://mindspore.cn/lite/api/zh-CN/r2.7.2/api_c/types_c.html#msmodeltype)。
+ - `model_context`: 模型的上下文环境,具体见: [Context](https://mindspore.cn/lite/api/zh-CN/r2.7.2/api_c/context_c.html)。
- 返回值
@@ -165,8 +165,8 @@ MSStatus MSModelBuildFromFile(MSModelHandle model, const char* model_path, MSMod
- `model`: 指向模型对象的指针。
- `model_path`: 模型文件路径。
- - `model_type`: 模型文件类型,具体见: [MSModelType](https://mindspore.cn/lite/api/zh-CN/r2.7.1/api_c/types_c.html#msmodeltype)。
- - `model_context`: 模型的上下文环境,具体见: [Context](https://mindspore.cn/lite/api/zh-CN/r2.7.1/api_c/context_c.html)。
+ - `model_type`: 模型文件类型,具体见: [MSModelType](https://mindspore.cn/lite/api/zh-CN/r2.7.2/api_c/types_c.html#msmodeltype)。
+ - `model_context`: 模型的上下文环境,具体见: [Context](https://mindspore.cn/lite/api/zh-CN/r2.7.2/api_c/context_c.html)。
- 返回值
diff --git a/docs/lite/api/source_zh_cn/api_c/status_c.md b/docs/lite/api/source_zh_cn/api_c/status_c.md
index 7e85a39309..c0431835de 100644
--- a/docs/lite/api/source_zh_cn/api_c/status_c.md
+++ b/docs/lite/api/source_zh_cn/api_c/status_c.md
@@ -1,6 +1,6 @@
# status_c
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_c/status_c.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_c/status_c.md)
```C
#include
diff --git a/docs/lite/api/source_zh_cn/api_c/tensor_c.md b/docs/lite/api/source_zh_cn/api_c/tensor_c.md
index 781fb1064a..4537b93d99 100644
--- a/docs/lite/api/source_zh_cn/api_c/tensor_c.md
+++ b/docs/lite/api/source_zh_cn/api_c/tensor_c.md
@@ -1,6 +1,6 @@
# tensor_c
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_c/tensor_c.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_c/tensor_c.md)
```C
#include
@@ -123,7 +123,7 @@ void MSTensorSetDataType(MSTensorHandle tensor, MSDataType type)
MSDataType MSTensorGetDataType(const MSTensorHandle tensor)
```
-获取MSTensor的数据类型,具体数据类型见[MSDataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_c/data_type_c.html#msdatatype)。
+获取MSTensor的数据类型,具体数据类型见[MSDataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_c/data_type_c.html#msdatatype)。
- 参数
- `tensor`: 指向MSTensor的指针。
@@ -171,7 +171,7 @@ void MSTensorSetFormat(MSTensorHandle tensor, MSFormat format)
- 参数
- `tensor`: 指向MSTensor的指针。
- - `format`: 张量的数据排列,具体见[MSFormat](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_c/format_c.html#msformat)。
+ - `format`: 张量的数据排列,具体见[MSFormat](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_c/format_c.html#msformat)。
### MSTensorGetFormat
@@ -183,7 +183,7 @@ MSFormat MSTensorGetFormat(const MSTensorHandle tensor)
- 返回值
- 张量的数据排列,具体见[MSFormat](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_c/format_c.html#msformat)。
+ 张量的数据排列,具体见[MSFormat](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_c/format_c.html#msformat)。
### MSTensorSetData
diff --git a/docs/lite/api/source_zh_cn/api_c/types_c.md b/docs/lite/api/source_zh_cn/api_c/types_c.md
index 3fa48aaffc..9e0c5c4cdc 100644
--- a/docs/lite/api/source_zh_cn/api_c/types_c.md
+++ b/docs/lite/api/source_zh_cn/api_c/types_c.md
@@ -1,6 +1,6 @@
# types_c
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_c/types_c.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_c/types_c.md)
```C
#include
diff --git a/docs/lite/api/source_zh_cn/api_cpp/lite_cpp_example.rst b/docs/lite/api/source_zh_cn/api_cpp/lite_cpp_example.rst
index edaaf3e8b2..561762ee28 100644
--- a/docs/lite/api/source_zh_cn/api_cpp/lite_cpp_example.rst
+++ b/docs/lite/api/source_zh_cn/api_cpp/lite_cpp_example.rst
@@ -4,6 +4,6 @@
.. toctree::
:maxdepth: 1
- 极简Demo↗
- 基于JNI接口的Android应用开发↗
- 高阶用法↗
\ No newline at end of file
+ 极简Demo↗
+ 基于JNI接口的Android应用开发↗
+ 高阶用法↗
\ No newline at end of file
diff --git a/docs/lite/api/source_zh_cn/api_cpp/mindspore.md b/docs/lite/api/source_zh_cn/api_cpp/mindspore.md
index ac1029ed08..d78f19bf76 100644
--- a/docs/lite/api/source_zh_cn/api_cpp/mindspore.md
+++ b/docs/lite/api/source_zh_cn/api_cpp/mindspore.md
@@ -1,6 +1,6 @@
# mindspore
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_cpp/mindspore.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_cpp/mindspore.md)
## 接口汇总
@@ -38,8 +38,8 @@
|--------------------------------------------------|---------------------------------------------------|--------|--------|
| [MSTensor](#mstensor) | MindSpore中的张量。 | √ | √ |
| [QuantParam](#quantparam) | MSTensor中的一组量化参数。 | √ | √ |
-| [DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_datatype.html) | MindSpore MSTensor保存的数据支持的类型。 | √ | √ |
-| [Format](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_format.html) | MindSpore MSTensor保存的数据支持的排列格式。 | √ | √ |
+| [DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_datatype.html) | MindSpore MSTensor保存的数据支持的类型。 | √ | √ |
+| [Format](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_format.html) | MindSpore MSTensor保存的数据支持的排列格式。 | √ | √ |
| [Allocator](#allocator-1) | 内存管理基类。 | √ | √ |
### 模型分组
@@ -157,9 +157,9 @@ Context的数据。
| [bool GetEnableParallel() const](#getenableparallel) | ✕ | √ |
| [void SetBuiltInDelegate(DelegateMode mode)](#setbuiltindelegate) | ✕ | √ |
| [DelegateMode GetBuiltInDelegate() const](#getbuiltindelegate) | ✕ | √ |
-| [void set_delegate(const std::shared_ptr\ &delegate)](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#set-delegate) | ✕ | √ |
+| [void set_delegate(const std::shared_ptr\ &delegate)](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#set-delegate) | ✕ | √ |
| [void SetDelegate(const std::shared_ptr\ &delegate)](#setdelegate) | ✕ | √ |
-| [std::shared_ptr\ get_delegate() const](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#get-delegate) | ✕ | √ |
+| [std::shared_ptr\ get_delegate() const](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#get-delegate) | ✕ | √ |
| [std::shared_ptr\ GetDelegate() const](#getdelegate) | ✕ | √ |
| [void SetMultiModalHW(bool float_mode)](#setmultimodalhw) | ✕ | √ |
| [bool GetMultiModalHW() const](#getmultimodalhw) | ✕ | √ |
@@ -2244,7 +2244,7 @@ Status Finalize()
## ModelExecutor
-\#include <[multi_model_runner.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/api/multi_model_runner.h)>
+\#include <[multi_model_runner.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/api/multi_model_runner.h)>
ModelExecutor定义了对Model的封装,用于调度多个Model的推理。
@@ -2326,7 +2326,7 @@ std::vector GetOutputs() const
## MultiModelRunner
-\#include <[multi_model_runner.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/api/multi_model_runner.h)>
+\#include <[multi_model_runner.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/api/multi_model_runner.h)>
MultiModelRunner用于创建包含多个Model的mindir,并提供调度多个模型的方式。
@@ -2612,10 +2612,10 @@ void DestroyTensorPtr(MSTensor *tensor) noexcept
| [bool IsConst() const](#isconst) | √ | √ |
| [bool IsDevice() const](#isdevice) | √ | ✕ |
| [MSTensor *Clone() const](#clone) | √ | √ |
-| [bool operator==(std::nullptr_t) const](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#operatorstd-nullptr-t) | √ | √ |
-| [bool operator!=(std::nullptr_t) const](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#operatorstd-nullptr-t-1) | √ | √ |
-| [bool operator!=(const MSTensor &tensor) const](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#operatorconst-mstensor-tensor) | √ | √ |
-| [bool operator==(const MSTensor &tensor) const](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#operatorconst-mstensor-tensor-1) | √ | √ |
+| [bool operator==(std::nullptr_t) const](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#operatorstd-nullptr-t) | √ | √ |
+| [bool operator!=(std::nullptr_t) const](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#operatorstd-nullptr-t-1) | √ | √ |
+| [bool operator!=(const MSTensor &tensor) const](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#operatorconst-mstensor-tensor) | √ | √ |
+| [bool operator==(const MSTensor &tensor) const](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#operatorconst-mstensor-tensor-1) | √ | √ |
| [void SetShape(const std::vector\ &shape)](#setshape) | √ | √ |
| [void SetDataType(enum DataType data_type)](#setdatatype) | √ | √ |
| [void SetTensorName(const std::string &name)](#settensorname) | √ | √ |
@@ -3176,7 +3176,7 @@ typedef enum {
\#include <[delegate.h](https://gitee.com/mindspore/mindspore/blob/v2.7.1/include/api/delegate.h)>
-定义了MindSpore Lite [Kernel](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_kernel.html#mindspore-kernel)列表的迭代器。
+定义了MindSpore Lite [Kernel](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_kernel.html#mindspore-kernel)列表的迭代器。
```cpp
using KernelIter = std::vector::iterator
@@ -3210,7 +3210,7 @@ DelegateModel(std::vector *kernels, const std::vector *kernels_
```
-[**Kernel**](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_kernel.html#kernel)的列表,保存模型的所有算子。
+[**Kernel**](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_kernel.html#kernel)的列表,保存模型的所有算子。
#### inputs_
@@ -3218,7 +3218,7 @@ std::vector *kernels_
const std::vector &inputs_
```
-[**MSTensor**](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)的列表,保存这个算子的输入tensor。
+[**MSTensor**](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)的列表,保存这个算子的输入tensor。
#### outputs_
@@ -3226,7 +3226,7 @@ const std::vector &inputs_
const std::vector &outputs
```
-[**MSTensor**](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)的列表,保存这个算子的输出tensor。
+[**MSTensor**](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)的列表,保存这个算子的输出tensor。
#### primitives_
@@ -3234,7 +3234,7 @@ const std::vector &outputs
const std::map &primitives_
```
-[**Kernel**](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_kernel.html#kernel)和**schema::Primitive**的Map,保存所有算子的属性。
+[**Kernel**](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_kernel.html#kernel)和**schema::Primitive**的Map,保存所有算子的属性。
#### version_
@@ -3326,7 +3326,7 @@ const std::vector &inputs()
- 返回值
- [**MSTensor**](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)的列表。
+ [**MSTensor**](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)的列表。
#### outputs
@@ -3338,7 +3338,7 @@ const std::vector &outputs()
- 返回值
- [**MSTensor**](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)的列表。
+ [**MSTensor**](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)的列表。
#### GetVersion
@@ -4221,11 +4221,11 @@ inline Status(const StatusCode code, int line_of_code, const char *file_name, co
| [inline std::string GetErrDescription() const](#geterrdescription) | √ | √ |
| [inline std::string SetErrDescription(const std::string &err_description)](#seterrdescription) | √ | √ |
| [inline void SetStatusMsg(const std::string &status_msg)](#setstatusmsg) | √ | √ |
-| [friend std::ostream &operator\<\<(std::ostream &os, const Status &s)](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#operator<
+\#include <[converter_context.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/converter_context.h)>
**enum**类型变量,定义MindSpore Lite转换支持的框架类型。
@@ -32,7 +32,7 @@
## ConverterParameters
-\#include <[converter_context.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/converter_context.h)>
+\#include <[converter_context.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/converter_context.h)>
**struct**类型结构体,定义模型解析时的转换参数,用于模型解析时的只读参数。
@@ -47,7 +47,7 @@ struct ConverterParameters {
## ConverterContext
-\#include <[converter_context.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/converter_context.h)>
+\#include <[converter_context.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/converter_context.h)>
模型转换过程中,基本信息的设置与获取。
@@ -113,7 +113,7 @@ static std::map GetConfigInfo(const std::string §i
## NodeParser
-\#include <[node_parser.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/node_parser.h)>
+\#include <[node_parser.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/node_parser.h)>
op节点的解析基类。
@@ -216,7 +216,7 @@ tflite节点解析接口函数。
## NodeParserPtr
-\#include <[node_parser.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/node_parser.h)>
+\#include <[node_parser.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/node_parser.h)>
NodeParser类的共享智能指针类型。
@@ -226,7 +226,7 @@ using NodeParserPtr = std::shared_ptr;
## ModelParser
-\#include <[model_parser.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/model_parser.h)>
+\#include <[model_parser.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/model_parser.h)>
解析原始模型的基类。
@@ -258,7 +258,7 @@ api::FuncGraphPtr Parse(const converter::ConverterParameters &flags);
- 参数
- - `flags`: 解析模型时基本信息,具体见[ConverterParameters](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_converter.html#converterparameters)。
+ - `flags`: 解析模型时基本信息,具体见[ConverterParameters](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_converter.html#converterparameters)。
- 返回值
diff --git a/docs/lite/api/source_zh_cn/api_cpp/mindspore_datatype.md b/docs/lite/api/source_zh_cn/api_cpp/mindspore_datatype.md
index 398d3d98d0..bbb5e52529 100644
--- a/docs/lite/api/source_zh_cn/api_cpp/mindspore_datatype.md
+++ b/docs/lite/api/source_zh_cn/api_cpp/mindspore_datatype.md
@@ -1,6 +1,6 @@
# DataType
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_cpp/mindspore_datatype.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_cpp/mindspore_datatype.md)
以下表格描述了MindSpore MSTensor保存的数据支持的类型。
diff --git a/docs/lite/api/source_zh_cn/api_cpp/mindspore_format.md b/docs/lite/api/source_zh_cn/api_cpp/mindspore_format.md
index 7212e18abd..d635a3e52a 100644
--- a/docs/lite/api/source_zh_cn/api_cpp/mindspore_format.md
+++ b/docs/lite/api/source_zh_cn/api_cpp/mindspore_format.md
@@ -1,6 +1,6 @@
# mindspore::Format
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_cpp/mindspore_format.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_cpp/mindspore_format.md)
以下表格描述了MindSpore MSTensor保存的数据支持的排列格式。
diff --git a/docs/lite/api/source_zh_cn/api_cpp/mindspore_kernel.md b/docs/lite/api/source_zh_cn/api_cpp/mindspore_kernel.md
index 94c57c6619..3e4e4aa0a4 100644
--- a/docs/lite/api/source_zh_cn/api_cpp/mindspore_kernel.md
+++ b/docs/lite/api/source_zh_cn/api_cpp/mindspore_kernel.md
@@ -1,6 +1,6 @@
# mindspore::kernel
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_cpp/mindspore_kernel.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_cpp/mindspore_kernel.md)
## 接口汇总
@@ -32,13 +32,13 @@ Kernel的默认与带参构造函数,构造Kernel实例。
- 参数
- - `inputs`: 算子输入[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)。
+ - `inputs`: 算子输入[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)。
- - `outputs`: 算子输出[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)。
+ - `outputs`: 算子输出[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)。
- `primitive`: 算子经由flatbuffers反序化为Primitive后的结果。
- - `ctx`: 算子的上下文[Context](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#context)。
+ - `ctx`: 算子的上下文[Context](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#context)。
### 析构函数
@@ -59,7 +59,7 @@ virtual int InferShape()
```
在用户调用`Model::Build`接口时,或是模型推理中需要推理算子形状时,会调用到该接口。
-在自定义算子场景中,用户可以覆写该接口,实现自定义算子的形状推理逻辑。详见[自定义算子章节](https://www.mindspore.cn/lite/docs/zh-CN/r2.7.1/advanced/third_party/register_kernel.html)。
+在自定义算子场景中,用户可以覆写该接口,实现自定义算子的形状推理逻辑。详见[自定义算子章节](https://www.mindspore.cn/lite/docs/zh-CN/r2.7.2/advanced/third_party/register_kernel.html)。
在`InferShape`函数中,一般需要实现算子的形状、数据类型和数据排布的推理逻辑。
- 返回值
@@ -84,7 +84,7 @@ virtual schema::QuantType quant_type()
## KernelInterface
-\#include <[kernel_interface.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/kernel_interface.h)>
+\#include <[kernel_interface.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/kernel_interface.h)>
算子扩展能力基类。
@@ -117,9 +117,9 @@ virtual Status Infer(std::vector *inputs, std::vector *inputs, std::vector &in_tensors) { th
- 参数
- - `in_tensors`: 算子的所有输入[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)列表。
+ - `in_tensors`: 算子的所有输入[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)列表。
#### set_input
@@ -251,7 +251,7 @@ virtual void set_input(mindspore::MSTensor in_tensor, int index) { this->inputs_
- 参数
- - `in_tensor`: 算子的输入[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)。
+ - `in_tensor`: 算子的输入[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)。
- `index`: 算子输入在所有输入中的下标,从0开始计数。
@@ -265,7 +265,7 @@ virtual void set_outputs(const std::vector &out_tensors) {
- 参数
- - `out_tensors`: 算子的所有输出[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)列表。
+ - `out_tensors`: 算子的所有输出[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)列表。
#### set_output
@@ -277,7 +277,7 @@ virtual void set_output(mindspore::MSTensor out_tensor, int index) { this->outpu
- 参数
- - `out_tensor`: 算子的输出[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)。
+ - `out_tensor`: 算子的输出[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)。
- `index`: 算子输出在所有输出中的下标,从0开始计数。
@@ -287,7 +287,7 @@ virtual void set_output(mindspore::MSTensor out_tensor, int index) { this->outpu
virtual const std::vector &inputs()
```
-返回算子的所有输入[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)列表。
+返回算子的所有输入[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)列表。
- 返回值
@@ -299,7 +299,7 @@ virtual const std::vector &inputs()
virtual const std::vector &outputs()
```
-返回算子的所有输出[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)列表。
+返回算子的所有输出[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)列表。
- 返回值
@@ -335,7 +335,7 @@ void set_name(const std::string &name)
const lite::Context *context() const
```
-返回算子对应的[Context](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#context)。
+返回算子对应的[Context](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#context)。
- 返回值
diff --git a/docs/lite/api/source_zh_cn/api_cpp/mindspore_registry.md b/docs/lite/api/source_zh_cn/api_cpp/mindspore_registry.md
index e0a710e6db..124f1fe9cf 100644
--- a/docs/lite/api/source_zh_cn/api_cpp/mindspore_registry.md
+++ b/docs/lite/api/source_zh_cn/api_cpp/mindspore_registry.md
@@ -1,32 +1,32 @@
# mindspore::registry
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_cpp/mindspore_registry.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_cpp/mindspore_registry.md)
## 接口汇总
| 类名 | 描述 |
| --- | --- |
| [NodeParserRegistry](#nodeparserregistry) | 扩展Node解析的注册类。|
-| [REG_NODE_PARSER](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_registry.html#reg-node-parser) | 注册扩展Node解析。|
+| [REG_NODE_PARSER](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_registry.html#reg-node-parser) | 注册扩展Node解析。|
| [ModelParserRegistry](#modelparserregistry) | 扩展Model解析的注册类。|
-| [REG_MODEL_PARSER](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_registry.html#reg-model-parser) | 注册扩展Model解析。|
+| [REG_MODEL_PARSER](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_registry.html#reg-model-parser) | 注册扩展Model解析。|
| [PassBase](#passbase) | Pass的基类。|
| [PassPosition](#passposition) | 扩展Pass的运行位置。|
| [PassRegistry](#passregistry) | 扩展Pass注册构造类。|
-| [REG_PASS](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_registry.html#reg-pass) | 注册扩展Pass。|
-| [REG_SCHEDULED_PASS](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_registry.html#reg-scheduled-pass) | 注册扩展Pass的调度顺序。|
+| [REG_PASS](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_registry.html#reg-pass) | 注册扩展Pass。|
+| [REG_SCHEDULED_PASS](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_registry.html#reg-scheduled-pass) | 注册扩展Pass的调度顺序。|
| [RegisterKernel](#registerkernel) | 算子注册实现类。|
| [KernelReg](#kernelreg) | 算子注册构造类。|
-| [REGISTER_KERNEL](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_registry.html#register-kernel) | 注册算子。|
-| [REGISTER_CUSTOM_KERNEL](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_registry.html#register-custom-kernel) | Custom算子注册。|
+| [REGISTER_KERNEL](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_registry.html#register-kernel) | 注册算子。|
+| [REGISTER_CUSTOM_KERNEL](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_registry.html#register-custom-kernel) | Custom算子注册。|
| [RegisterKernelInterface](#registerkernelinterface) | 算子扩展能力注册实现类。|
| [KernelInterfaceReg](#kernelinterfacereg) | 算子扩展能力注册构造类。|
-| [REGISTER_KERNEL_INTERFACE](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_registry.html#register-kernel-interface) | 注册算子扩展能力。|
-| [REGISTER_CUSTOM_KERNEL_INTERFACE](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_registry.html#register-custom-kernel-interface) | 注册Custom算子扩展能力。|
+| [REGISTER_KERNEL_INTERFACE](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_registry.html#register-kernel-interface) | 注册算子扩展能力。|
+| [REGISTER_CUSTOM_KERNEL_INTERFACE](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_registry.html#register-custom-kernel-interface) | 注册Custom算子扩展能力。|
## NodeParserRegistry
-\#include <[node_parser_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/node_parser_registry.h)>
+\#include <[node_parser_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/node_parser_registry.h)>
NodeParserRegistry类用于注册及获取NodeParser类型的共享智能指针。
@@ -41,11 +41,11 @@ NodeParserRegistry(converter::FmkType fmk_type, const std::string &node_type,
- 参数
- - `fmk_type`: 框架类型,具体见[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_converter.html#fmktype)说明。
+ - `fmk_type`: 框架类型,具体见[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_converter.html#fmktype)说明。
- `node_type`: 节点的类型。
- - `node_parser`: NodeParser类型的共享智能指针实例,具体见[NodeParserPtr](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_converter.html#nodeparserptr)说明。
+ - `node_parser`: NodeParser类型的共享智能指针实例,具体见[NodeParserPtr](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_converter.html#nodeparserptr)说明。
### ~NodeParserRegistry
@@ -67,13 +67,13 @@ static converter::NodeParserPtr GetNodeParser(converter::FmkType fmk_type, const
- 参数
- - `fmk_type`: 框架类型,具体见[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_converter.html#fmktype)说明。
+ - `fmk_type`: 框架类型,具体见[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_converter.html#fmktype)说明。
- `node_type`: 节点的类型。
## REG_NODE_PARSER
-\#include <[node_parser_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/node_parser_registry.h)>
+\#include <[node_parser_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/node_parser_registry.h)>
```c++
#define REG_NODE_PARSER(fmk_type, node_type, node_parser)
@@ -83,25 +83,25 @@ static converter::NodeParserPtr GetNodeParser(converter::FmkType fmk_type, const
- 参数
- - `fmk_type`: 框架类型,具体见[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_converter.html#fmktype)说明。
+ - `fmk_type`: 框架类型,具体见[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_converter.html#fmktype)说明。
- `node_type`: 节点的类型。
- - `node_parser`: NodeParser类型的共享智能指针实例,具体见[NodeParserPtr](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_converter.html#nodeparserptr)说明。
+ - `node_parser`: NodeParser类型的共享智能指针实例,具体见[NodeParserPtr](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_converter.html#nodeparserptr)说明。
## ModelParserCreator
-\#include <[model_parser_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/model_parser_registry.h)>
+\#include <[model_parser_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/model_parser_registry.h)>
```c++
typedef converter::ModelParser *(*ModelParserCreator)()
```
-创建[ModelParser](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_converter.html#modelparser)的函数原型声明。
+创建[ModelParser](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_converter.html#modelparser)的函数原型声明。
## ModelParserRegistry
-\#include <[model_parser_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/model_parser_registry.h)>
+\#include <[model_parser_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/model_parser_registry.h)>
ModelParserRegistry类用于注册及获取ModelParserCreator类型的函数指针。
@@ -115,7 +115,7 @@ ModelParserRegistry(FmkType fmk, ModelParserCreator creator)
- 参数
- - `fmk`: 框架类型,具体见[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_converter.html#fmktype)说明。
+ - `fmk`: 框架类型,具体见[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_converter.html#fmktype)说明。
- `creator`: ModelParserCreator类型的函数指针,具体见[ModelParserCreator](#modelparsercreator)说明。
@@ -139,11 +139,11 @@ static ModelParser *GetModelParser(FmkType fmk)
- 参数
- - `fmk`: 框架类型,具体见[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_converter.html#fmktype)说明。
+ - `fmk`: 框架类型,具体见[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_converter.html#fmktype)说明。
## REG_MODEL_PARSER
-\#include <[model_parser_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/model_parser_registry.h)>
+\#include <[model_parser_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/model_parser_registry.h)>
```c++
#define REG_MODEL_PARSER(fmk, parserCreator)
@@ -153,15 +153,15 @@ static ModelParser *GetModelParser(FmkType fmk)
- 参数
- - `fmk`: 框架类型,具体见[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_converter.html#fmktype)说明。
+ - `fmk`: 框架类型,具体见[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_converter.html#fmktype)说明。
- `creator`: ModelParserCreator类型的函数指针,具体见[ModelParserCreator](#modelparsercreator)说明。
-> 用户自定义的ModelParser,框架类型必须满足设定支持的框架类型[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_converter.html#fmktype)。
+> 用户自定义的ModelParser,框架类型必须满足设定支持的框架类型[FmkType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_converter.html#fmktype)。
## PassBase
-\#include <[pass_base.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/pass_base.h)>
+\#include <[pass_base.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/pass_base.h)>
PassBase定义了图优化的基类,以供用户继承并自定义图优化算法。
@@ -201,7 +201,7 @@ virtual bool Execute(const api::FuncGraphPtr &func_graph) = 0;
## PassBasePtr
-\#include <[pass_base.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/pass_base.h)>
+\#include <[pass_base.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/pass_base.h)>
PassBase类的共享智能指针类型。
@@ -211,7 +211,7 @@ using PassBasePtr = std::shared_ptr
## PassPosition
-\#include <[pass_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/pass_registry.h)>
+\#include <[pass_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/pass_registry.h)>
**enum**类型变量,定义扩展Pass的运行位置。
@@ -224,7 +224,7 @@ enum PassPosition {
## PassRegistry
-\#include <[pass_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/pass_registry.h)>
+\#include <[pass_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/pass_registry.h)>
PassRegistry类用于注册及获取Pass类实例。
@@ -290,7 +290,7 @@ static PassBasePtr GetPassFromStoreRoom(const std::string &pass_name)
## REG_PASS
-\#include <[pass_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/pass_registry.h)>
+\#include <[pass_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/pass_registry.h)>
```c++
#define REG_PASS(name, pass)
@@ -306,7 +306,7 @@ static PassBasePtr GetPassFromStoreRoom(const std::string &pass_name)
## REG_SCHEDULED_PASS
-\#include <[pass_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/pass_registry.h)>
+\#include <[pass_registry.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/pass_registry.h)>
```c++
#define REG_SCHEDULED_PASS(position, names)
@@ -322,7 +322,7 @@ static PassBasePtr GetPassFromStoreRoom(const std::string &pass_name)
> MindSpore Lite开放了部分内置Pass,请见以下说明。用户可以在`names`参数中添加内置Pass的命名标识,以在指定运行处调用内置Pass。
>
-> - `ConstFoldPass`: 将输入均是常量的节点进行离线计算,导出的模型将不含该节点。特别地,针对shape算子,在[inputShape](https://www.mindspore.cn/lite/docs/zh-CN/r2.7.1/converter/converter_tool.html#参数说明)给定的情形下,也会触发预计算。
+> - `ConstFoldPass`: 将输入均是常量的节点进行离线计算,导出的模型将不含该节点。特别地,针对shape算子,在[inputShape](https://www.mindspore.cn/lite/docs/zh-CN/r2.7.2/converter/converter_tool.html#参数说明)给定的情形下,也会触发预计算。
> - `DumpGraph`: 导出当前状态下的模型。请确保当前模型为NHWC或者NCHW格式的模型,例如卷积算子等。
> - `ToNCHWFormat`: 将当前状态下的模型转换为NCHW的格式,例如,四维的图输入、卷积算子等。
> - `ToNHWCFormat`: 将当前状态下的模型转换为NHWC的格式,例如,四维的图输入、卷积算子等。
@@ -334,7 +334,7 @@ static PassBasePtr GetPassFromStoreRoom(const std::string &pass_name)
## KernelDesc
-\#include <[registry/register_kernel.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/register_kernel.h)>
+\#include <[registry/register_kernel.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/register_kernel.h)>
**struct**类型结构体,定义扩展kernel的基本属性。
@@ -349,7 +349,7 @@ struct KernelDesc {
## RegisterKernel
-\#include <[registry/register_kernel.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/register_kernel.h)>
+\#include <[registry/register_kernel.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/register_kernel.h)>
### CreateKernel
@@ -363,13 +363,13 @@ using CreateKernel = std::function(
- 参数
- - `inputs`: 算子输入[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)。
+ - `inputs`: 算子输入[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)。
- - `outputs`: 算子输出[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#mstensor)。
+ - `outputs`: 算子输出[MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#mstensor)。
- `primitive`: 算子经由flatbuffers反序化为Primitive后的结果。
- - `ctx`: 算子的上下文[Context](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html#context)。
+ - `ctx`: 算子的上下文[Context](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html#context)。
### 公有成员函数
@@ -387,9 +387,9 @@ static Status RegKernel(const std::string &arch, const std::string &provider, Da
- `provider`: 生产商名,由用户自定义。
- - `data_type`: 算子支持的数据类型,具体见[DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_datatype.html)。
+ - `data_type`: 算子支持的数据类型,具体见[DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_datatype.html)。
- - `op_type`: 算子类型,定义在[ops.fbs](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/schema/ops.fbs)中,编绎时会生成到ops_generated.h,该文件可以在发布件中获取。
+ - `op_type`: 算子类型,定义在[ops.fbs](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/schema/ops.fbs)中,编绎时会生成到ops_generated.h,该文件可以在发布件中获取。
- `creator`: 创建算子的函数指针,具体见[CreateKernel](#createkernel)的说明。
@@ -407,7 +407,7 @@ Custom算子注册。
- `provider`: 生产商名,由用户自定义。
- - `data_type`: 算子支持的数据类型,具体见[DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_datatype.html)。
+ - `data_type`: 算子支持的数据类型,具体见[DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_datatype.html)。
- `type`: 算子类型,由用户自定义,确保唯一即可。
@@ -429,7 +429,7 @@ static CreateKernel GetCreator(const schema::Primitive *primitive, KernelDesc *d
## KernelReg
-\#include <[registry/register_kernel.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/register_kernel.h)>
+\#include <[registry/register_kernel.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/register_kernel.h)>
### ~KernelReg
@@ -453,9 +453,9 @@ KernelReg(const std::string &arch, const std::string &provider, DataType data_ty
- `provider`: 生产商名,由用户自定义。
- - `data_type`: 算子支持的数据类型,具体见[DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_datatype.html)。
+ - `data_type`: 算子支持的数据类型,具体见[DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_datatype.html)。
- - `op_type`: 算子类型,定义在[ops.fbs](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/schema/ops.fbs)中,编绎时会生成到ops_generated.h,该文件可以在发布件中获取。
+ - `op_type`: 算子类型,定义在[ops.fbs](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/schema/ops.fbs)中,编绎时会生成到ops_generated.h,该文件可以在发布件中获取。
- `creator`: 创建算子的函数指针,具体见[CreateKernel](#createkernel)的说明。
@@ -471,7 +471,7 @@ KernelReg(const std::string &arch, const std::string &provider, DataType data_ty
- `provider`: 生产商名,由用户自定义。
- - `data_type`: 算子支持的数据类型,具体见[DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_datatype.html)。
+ - `data_type`: 算子支持的数据类型,具体见[DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_datatype.html)。
- `op_type`: 算子类型,由用户自定义,确保唯一即可。
@@ -491,9 +491,9 @@ KernelReg(const std::string &arch, const std::string &provider, DataType data_ty
- `provider`: 生产商名,由用户自定义。
- - `data_type`: 算子支持的数据类型,具体见[DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_datatype.html)。
+ - `data_type`: 算子支持的数据类型,具体见[DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_datatype.html)。
- - `op_type`: 算子类型,定义在[ops.fbs](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/schema/ops.fbs)中,编绎时会生成到ops_generated.h,该文件可以在发布件中获取。
+ - `op_type`: 算子类型,定义在[ops.fbs](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/schema/ops.fbs)中,编绎时会生成到ops_generated.h,该文件可以在发布件中获取。
- `creator`: 创建算子的函数指针,具体见[CreateKernel](#createkernel)的说明。
@@ -511,7 +511,7 @@ KernelReg(const std::string &arch, const std::string &provider, DataType data_ty
- `provider`: 生产商名,由用户自定义。
- - `data_type`: 算子支持的数据类型,具体见[DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore_datatype.html)。
+ - `data_type`: 算子支持的数据类型,具体见[DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore_datatype.html)。
- `op_type`: 算子类型,由用户自定义,确保唯一即可。
@@ -519,7 +519,7 @@ KernelReg(const std::string &arch, const std::string &provider, DataType data_ty
## KernelInterfaceCreator
-\#include <[registry/register_kernel_interface.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/register_kernel_interface.h)>
+\#include <[registry/register_kernel_interface.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/register_kernel_interface.h)>
定义创建算子的函数指针类型。
@@ -529,7 +529,7 @@ using KernelInterfaceCreator = std::function
+\#include <[registry/register_kernel_interface.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/register_kernel_interface.h)>
算子扩展能力注册实现类。
@@ -563,7 +563,7 @@ static Status Reg(const std::string &provider, int op_type, const KernelInterfac
- `provider`: 生产商,由用户自定义。
- - `op_type`: 算子类型,定义在[ops.fbs](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/schema/ops.fbs)中,编绎时会生成到ops_generated.h,该文件可以在发布件中获取。
+ - `op_type`: 算子类型,定义在[ops.fbs](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/schema/ops.fbs)中,编绎时会生成到ops_generated.h,该文件可以在发布件中获取。
- `creator`: KernelInterface的创建函数,详细见[KernelInterfaceCreator](#kernelinterfacecreator)的说明。
@@ -585,7 +585,7 @@ static std::shared_ptr GetKernelInterface(const std::st
## KernelInterfaceReg
-\#include <[registry/register_kernel_interface.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/register_kernel_interface.h)>
+\#include <[registry/register_kernel_interface.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/register_kernel_interface.h)>
算子扩展能力注册构造类。
@@ -601,7 +601,7 @@ KernelInterfaceReg(const std::string &provider, int op_type, const KernelInterfa
- `provider`: 生产商,由用户自定义。
- - `op_type`: 算子类型,定义在[ops.fbs](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/schema/ops.fbs)中,编绎时会生成到ops_generated.h,该文件可以在发布件中获取。
+ - `op_type`: 算子类型,定义在[ops.fbs](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/schema/ops.fbs)中,编绎时会生成到ops_generated.h,该文件可以在发布件中获取。
- `creator`: KernelInterface的创建函数,详细见[KernelInterfaceCreator](#kernelinterfacecreator)的说明。
@@ -621,7 +621,7 @@ KernelInterfaceReg(const std::string &provider, const std::string &op_type, cons
## REGISTER_KERNEL_INTERFACE
-\#include <[registry/register_kernel_interface.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/register_kernel_interface.h)>
+\#include <[registry/register_kernel_interface.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/register_kernel_interface.h)>
注册KernelInterface的实现。
@@ -633,13 +633,13 @@ KernelInterfaceReg(const std::string &provider, const std::string &op_type, cons
- `provider`: 生产商,由用户自定义。
- - `op_type`: 算子类型,定义在[ops.fbs](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/schema/ops.fbs)中,编绎时会生成到ops_generated.h,该文件可以在发布件中获取。
+ - `op_type`: 算子类型,定义在[ops.fbs](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/schema/ops.fbs)中,编绎时会生成到ops_generated.h,该文件可以在发布件中获取。
- `creator`: 创建KernelInterface的函数指针,具体见[KernelInterfaceCreator](#kernelinterfacecreator)的说明。
## REGISTER_CUSTOM_KERNEL_INTERFACE
-\#include <[registry/register_kernel_interface.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/register_kernel_interface.h)>
+\#include <[registry/register_kernel_interface.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/register_kernel_interface.h)>
注册Custom算子对应的KernelInterface实现。
diff --git a/docs/lite/api/source_zh_cn/api_cpp/mindspore_registry_opencl.md b/docs/lite/api/source_zh_cn/api_cpp/mindspore_registry_opencl.md
index 3ee73463bc..fe3360556b 100644
--- a/docs/lite/api/source_zh_cn/api_cpp/mindspore_registry_opencl.md
+++ b/docs/lite/api/source_zh_cn/api_cpp/mindspore_registry_opencl.md
@@ -1,6 +1,6 @@
# mindspore::registry::opencl
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_cpp/mindspore_registry_opencl.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_cpp/mindspore_registry_opencl.md)
## 接口汇总
@@ -10,7 +10,7 @@
## OpenCLRuntimeWrapper
-\#include <[include/registry/opencl_runtime_wrapper.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/include/registry/opencl_runtime_wrapper.h)>
+\#include <[include/registry/opencl_runtime_wrapper.h](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/include/registry/opencl_runtime_wrapper.h)>
OpenCLRuntimeWrapper类包装了内部OpenCL的相关接口,用于支持南向GPU算子的开发。
@@ -134,7 +134,7 @@ Status SyncCommandQueue()
std::shared_ptr GetAllocator()
```
-获取GPU内存分配器的智能指针。通过[Allocator接口](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_cpp/mindspore.html),可申请GPU内存,用于OpenCL内核的运算。
+获取GPU内存分配器的智能指针。通过[Allocator接口](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_cpp/mindspore.html),可申请GPU内存,用于OpenCL内核的运算。
#### MapBuffer
diff --git a/docs/lite/api/source_zh_cn/api_java/ascend_device_info.md b/docs/lite/api/source_zh_cn/api_java/ascend_device_info.md
index 9bfd68ace9..97fb5f6e72 100644
--- a/docs/lite/api/source_zh_cn/api_java/ascend_device_info.md
+++ b/docs/lite/api/source_zh_cn/api_java/ascend_device_info.md
@@ -1,6 +1,6 @@
# AscendDeviceInfo
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_java/ascend_device_info.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_java/ascend_device_info.md)
```java
import com.mindspore.config.AscendDeviceInfo;
diff --git a/docs/lite/api/source_zh_cn/api_java/class_list.md b/docs/lite/api/source_zh_cn/api_java/class_list.md
index c490cf5ef0..1580533cb5 100644
--- a/docs/lite/api/source_zh_cn/api_java/class_list.md
+++ b/docs/lite/api/source_zh_cn/api_java/class_list.md
@@ -1,20 +1,20 @@
# 类列表
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_java/class_list.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_java/class_list.md)
| 包 | 类 | 描述 | 云侧推理是否支持 | 端侧推理是否支持 |
| ------------------------- | ------------------------------------------------------------ | ------------------------------------------------------------ |--------|--------|
-| com.mindspore | [Model](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_java/model.html) | Model定义了MindSpore中的模型,用于计算图的编译和执行。 | √ | √ |
-| com.mindspore.config | [MSContext](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_java/mscontext.html) | MSContext用于保存执行期间的上下文。 | √ | √ |
-| com.mindspore | [MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_java/mstensor.html) | MSTensor定义了MindSpore中的张量。 | √ | √ |
-| com.mindspore | [ModelParallelRunner](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_java/model_parallel_runner.html) | 定义了MindSpore Lite并发推理。 | √ | ✕ |
-| com.mindspore.config | [RunnerConfig](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_java/runner_config.html) | RunnerConfig定义并发推理的配置参数。 | √ | ✕ |
-| com.mindspore | [Graph](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_java/graph.html) | Graph定义了MindSpore中的计算图。 | ✕ | √ |
-| com.mindspore.config | [CpuBindMode](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_java/mscontext.html#cpubindmode) | CpuBindMode定义了CPU绑定模式。 | √ | √ |
-| com.mindspore.config | [DeviceType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_java/mscontext.html#devicetype) | DeviceType定义了后端设备类型。 | √ | √ |
-| com.mindspore.config | [DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_java/mstensor.html#datatype) | DataType定义了所支持的数据类型。 | √ | √ |
-| com.mindspore.config | [Version](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_java/version.html) | Version用于获取MindSpore的版本信息。 | √ | √ |
-| com.mindspore.config | [ModelType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_java/model.html#modeltype) | ModelType定义了模型文件的类型。 | √ | √ |
-| com.mindspore.config | [AscendDeviceInfo](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_java/ascend_device_info.html) | MindSpore Lite用于昇腾硬件推理的配置参数。 | √ | ✕ |
-| com.mindspore.config | [TrainCfg](https://www.mindspore.cn/lite/api/zh-CN/r2.7.1/api_java/train_cfg.html) | 用于端上模型训练的配置参数。 | ✕ | √ |
+| com.mindspore | [Model](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_java/model.html) | Model定义了MindSpore中的模型,用于计算图的编译和执行。 | √ | √ |
+| com.mindspore.config | [MSContext](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_java/mscontext.html) | MSContext用于保存执行期间的上下文。 | √ | √ |
+| com.mindspore | [MSTensor](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_java/mstensor.html) | MSTensor定义了MindSpore中的张量。 | √ | √ |
+| com.mindspore | [ModelParallelRunner](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_java/model_parallel_runner.html) | 定义了MindSpore Lite并发推理。 | √ | ✕ |
+| com.mindspore.config | [RunnerConfig](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_java/runner_config.html) | RunnerConfig定义并发推理的配置参数。 | √ | ✕ |
+| com.mindspore | [Graph](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_java/graph.html) | Graph定义了MindSpore中的计算图。 | ✕ | √ |
+| com.mindspore.config | [CpuBindMode](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_java/mscontext.html#cpubindmode) | CpuBindMode定义了CPU绑定模式。 | √ | √ |
+| com.mindspore.config | [DeviceType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_java/mscontext.html#devicetype) | DeviceType定义了后端设备类型。 | √ | √ |
+| com.mindspore.config | [DataType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_java/mstensor.html#datatype) | DataType定义了所支持的数据类型。 | √ | √ |
+| com.mindspore.config | [Version](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_java/version.html) | Version用于获取MindSpore的版本信息。 | √ | √ |
+| com.mindspore.config | [ModelType](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_java/model.html#modeltype) | ModelType定义了模型文件的类型。 | √ | √ |
+| com.mindspore.config | [AscendDeviceInfo](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_java/ascend_device_info.html) | MindSpore Lite用于昇腾硬件推理的配置参数。 | √ | ✕ |
+| com.mindspore.config | [TrainCfg](https://www.mindspore.cn/lite/api/zh-CN/r2.7.2/api_java/train_cfg.html) | 用于端上模型训练的配置参数。 | ✕ | √ |
diff --git a/docs/lite/api/source_zh_cn/api_java/graph.md b/docs/lite/api/source_zh_cn/api_java/graph.md
index 8671d37fb7..b7929eaa5f 100644
--- a/docs/lite/api/source_zh_cn/api_java/graph.md
+++ b/docs/lite/api/source_zh_cn/api_java/graph.md
@@ -1,6 +1,6 @@
# Graph
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_java/graph.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_java/graph.md)
```java
import com.mindspore.Graph;
diff --git a/docs/lite/api/source_zh_cn/api_java/lite_java_example.rst b/docs/lite/api/source_zh_cn/api_java/lite_java_example.rst
index fe574a73fa..5f37deb67a 100644
--- a/docs/lite/api/source_zh_cn/api_java/lite_java_example.rst
+++ b/docs/lite/api/source_zh_cn/api_java/lite_java_example.rst
@@ -4,6 +4,6 @@
.. toctree::
:maxdepth: 1
- 极简Demo↗
- 基于Java接口的Android应用开发↗
- 高阶用法↗
\ No newline at end of file
+ 极简Demo↗
+ 基于Java接口的Android应用开发↗
+ 高阶用法↗
\ No newline at end of file
diff --git a/docs/lite/api/source_zh_cn/api_java/model.md b/docs/lite/api/source_zh_cn/api_java/model.md
index dc52ccacda..3e55fa495e 100644
--- a/docs/lite/api/source_zh_cn/api_java/model.md
+++ b/docs/lite/api/source_zh_cn/api_java/model.md
@@ -1,6 +1,6 @@
# Model
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_java/model.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_java/model.md)
```java
import com.mindspore.Model;
diff --git a/docs/lite/api/source_zh_cn/api_java/model_parallel_runner.md b/docs/lite/api/source_zh_cn/api_java/model_parallel_runner.md
index 55ae5b5930..1062f01ac0 100644
--- a/docs/lite/api/source_zh_cn/api_java/model_parallel_runner.md
+++ b/docs/lite/api/source_zh_cn/api_java/model_parallel_runner.md
@@ -1,6 +1,6 @@
# ModelParallelRunner
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_java/model_parallel_runner.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_java/model_parallel_runner.md)
```java
import com.mindspore.config.RunnerConfig;
diff --git a/docs/lite/api/source_zh_cn/api_java/mscontext.md b/docs/lite/api/source_zh_cn/api_java/mscontext.md
index cae3f0d3c0..405a29cd35 100644
--- a/docs/lite/api/source_zh_cn/api_java/mscontext.md
+++ b/docs/lite/api/source_zh_cn/api_java/mscontext.md
@@ -1,6 +1,6 @@
# MSContext
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_java/mscontext.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_java/mscontext.md)
```java
import com.mindspore.config.MSContext;
@@ -54,7 +54,7 @@ public boolean init(int threadNum, int cpuBindMode)
- 参数
- `threadNum`: 线程数。
- - `cpuBindMode`: CPU绑定模式,`cpuBindMode`在[com.mindspore.config.CpuBindMode](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/java/src/main/java/com/mindspore/config/CpuBindMode.java)中定义。
+ - `cpuBindMode`: CPU绑定模式,`cpuBindMode`在[com.mindspore.config.CpuBindMode](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/java/src/main/java/com/mindspore/config/CpuBindMode.java)中定义。
- 返回值
@@ -69,7 +69,7 @@ public boolean init(int threadNum, int cpuBindMode, boolean isEnableParallel)
- 参数
- `threadNum`: 线程数。
- - `cpuBindMode`: CPU绑定模式,`cpuBindMode`在[com.mindspore.config.CpuBindMode](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/java/src/main/java/com/mindspore/config/CpuBindMode.java)中定义。
+ - `cpuBindMode`: CPU绑定模式,`cpuBindMode`在[com.mindspore.config.CpuBindMode](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/java/src/main/java/com/mindspore/config/CpuBindMode.java)中定义。
- `isEnableParallel`: 是否开启异构并行。
- 返回值
@@ -86,7 +86,7 @@ public boolean addDeviceInfo(int deviceType, boolean isEnableFloat16)
- 参数
- - `deviceType`: 设备类型,`deviceType`在[com.mindspore.config.DeviceType](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/java/src/main/java/com/mindspore/config/DeviceType.java)中定义。
+ - `deviceType`: 设备类型,`deviceType`在[com.mindspore.config.DeviceType](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/java/src/main/java/com/mindspore/config/DeviceType.java)中定义。
- `isEnableFloat16`: 是否开启fp16。
- 返回值
@@ -101,7 +101,7 @@ public boolean addDeviceInfo(int deviceType, boolean isEnableFloat16, int npuFre
- 参数
- - `deviceType`: 设备类型,`deviceType`在[com.mindspore.config.DeviceType](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/java/src/main/java/com/mindspore/config/DeviceType.java)中定义。
+ - `deviceType`: 设备类型,`deviceType`在[com.mindspore.config.DeviceType](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/java/src/main/java/com/mindspore/config/DeviceType.java)中定义。
- `isEnableFloat16`: 是否开启fp16。
- `npuFreq`: NPU运行频率,仅当deviceType为npu才需要。
diff --git a/docs/lite/api/source_zh_cn/api_java/mstensor.md b/docs/lite/api/source_zh_cn/api_java/mstensor.md
index 959c3e1ef3..d07d9826b6 100644
--- a/docs/lite/api/source_zh_cn/api_java/mstensor.md
+++ b/docs/lite/api/source_zh_cn/api_java/mstensor.md
@@ -1,6 +1,6 @@
# MSTensor
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_java/mstensor.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_java/mstensor.md)
```java
import com.mindspore.MSTensor;
@@ -86,7 +86,7 @@ public int[] getShape()
public int getDataType()
```
-DataType在[com.mindspore.DataType](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.1/mindspore-lite/java/src/main/java/com/mindspore/config/DataType.java)中定义。
+DataType在[com.mindspore.DataType](https://gitee.com/mindspore/mindspore-lite/blob/r2.7.2/mindspore-lite/java/src/main/java/com/mindspore/config/DataType.java)中定义。
- 返回值
diff --git a/docs/lite/api/source_zh_cn/api_java/runner_config.md b/docs/lite/api/source_zh_cn/api_java/runner_config.md
index 3033b83cba..f0c92b6cb3 100644
--- a/docs/lite/api/source_zh_cn/api_java/runner_config.md
+++ b/docs/lite/api/source_zh_cn/api_java/runner_config.md
@@ -1,6 +1,6 @@
# RunnerConfig
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_java/runner_config.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_java/runner_config.md)
RunnerConfig定义了MindSpore Lite并发推理的配置参数。
diff --git a/docs/lite/api/source_zh_cn/api_java/train_cfg.md b/docs/lite/api/source_zh_cn/api_java/train_cfg.md
index e9d1540c47..95407ba090 100644
--- a/docs/lite/api/source_zh_cn/api_java/train_cfg.md
+++ b/docs/lite/api/source_zh_cn/api_java/train_cfg.md
@@ -1,6 +1,6 @@
# TrainCfg
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_java/train_cfg.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_java/train_cfg.md)
```java
import com.mindspore.config.TrainCfg;
diff --git a/docs/lite/api/source_zh_cn/api_java/version.md b/docs/lite/api/source_zh_cn/api_java/version.md
index 267d78a39b..4881d380f1 100644
--- a/docs/lite/api/source_zh_cn/api_java/version.md
+++ b/docs/lite/api/source_zh_cn/api_java/version.md
@@ -1,6 +1,6 @@
# Version
-[](https://gitee.com/mindspore/docs/blob/r2.7.1/docs/lite/api/source_zh_cn/api_java/version.md)
+[](https://gitee.com/mindspore/docs/blob/r2.7.2/docs/lite/api/source_zh_cn/api_java/version.md)
```java
import com.mindspore.config.Version;
diff --git a/docs/lite/api/source_zh_cn/index.rst b/docs/lite/api/source_zh_cn/index.rst
index aea0c9d775..b278970aeb 100644
--- a/docs/lite/api/source_zh_cn/index.rst
+++ b/docs/lite/api/source_zh_cn/index.rst
@@ -12,21 +12,21 @@ MindSpore Lite API 支持情况汇总
+---------------------+---------------------------------------------------------------------------------------------------------+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| 类名 | 接口说明 | C++ 接口 | Python 接口 |
+=====================+=========================================================================================================+==========================================================================================================================================================================================================================+============================================================================================================================================================================================================================================================================================================================================================================+
-| Context | 设置运行时的线程数 | void SetThreadNum(int32_t thread_num) | `Context.cpu.thread_num `__ |
+| Context | 设置运行时的线程数 | void SetThreadNum(int32_t thread_num) | `Context.cpu.thread_num