From c819356b2b27cdd0d99e934163b59c6415edcb7e Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E5=AE=A6=E6=99=93=E7=8E=B2?= <3174348550@qq.com> Date: Mon, 8 Sep 2025 14:11:40 +0800 Subject: [PATCH] modify contents --- .../source_en/model_infer/ms_infer/ms_infer_quantization.md | 2 +- .../source_zh_cn/model_infer/ms_infer/ms_infer_quantization.md | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/tutorials/source_en/model_infer/ms_infer/ms_infer_quantization.md b/tutorials/source_en/model_infer/ms_infer/ms_infer_quantization.md index 5d20e83bcf..30a3674191 100644 --- a/tutorials/source_en/model_infer/ms_infer/ms_infer_quantization.md +++ b/tutorials/source_en/model_infer/ms_infer/ms_infer_quantization.md @@ -6,7 +6,7 @@ MindSpore is an all-scenario AI framework. When a model is deployed on the device or other lightweight devices, it may be subject to memory, power consumption, and latency. Therefore, the model needs to be compressed before deployment. -[MindSpore Golden Stick](https://www.mindspore.cn/golden_stick/docs/en/master/index.html) provides the model compression capability of MindSpore. MindSpore Golden Stick is a set of model compression algorithms jointly designed and developed by Huawei Noah's Ark team and Huawei MindSpore team. It provides a series of model compression algorithms for MindSpore, supporting quantization modes such as A16W8, A16W4, A8W8, and KVCache. For details, see [MindSpore Golden Stick](https://www.mindspore.cn/golden_stick/docs/en/master/index.html). +[MindSpore Golden Stick](https://www.mindspore.cn/golden_stick/docs/en/master/index.html) provides the model compression capability of MindSpore. MindSpore Golden Stick is a set of model compression algorithms jointly designed and developed by Huawei Noah's Ark team and MindSpore team. It provides a series of model compression algorithms for MindSpore, supporting quantization modes such as A16W8, A16W4, A8W8, and KVCache. For details, see [MindSpore Golden Stick](https://www.mindspore.cn/golden_stick/docs/en/master/index.html). ## Basic Model Quantization Process diff --git a/tutorials/source_zh_cn/model_infer/ms_infer/ms_infer_quantization.md b/tutorials/source_zh_cn/model_infer/ms_infer/ms_infer_quantization.md index 6997d8ce6d..3a3e409340 100644 --- a/tutorials/source_zh_cn/model_infer/ms_infer/ms_infer_quantization.md +++ b/tutorials/source_zh_cn/model_infer/ms_infer/ms_infer_quantization.md @@ -6,7 +6,7 @@ MindSpore是一个全场景的AI框架。当模型部署到端侧或者其他轻量化设备上时,对于部署的内存、功耗、时延等有各种限制,因此在部署前需要对模型进行压缩。 -MindSpore的模型压缩能力由 [MindSpore Golden Stick](https://www.mindspore.cn/golden_stick/docs/zh-CN/master/index.html) 提供,MindSpore Golden Stick是华为诺亚团队和华为MindSpore团队联合设计开发的一个模型压缩算法集,为MindSpore提供了一系列模型压缩算法,支持A16W8、A16W4、A8W8和KVCache等量化方式。详细资料可前往 [MindSpore Golden Stick官方资料](https://www.mindspore.cn/golden_stick/docs/zh-CN/master/index.html) 查看。 +MindSpore的模型压缩能力由 [MindSpore Golden Stick](https://www.mindspore.cn/golden_stick/docs/zh-CN/master/index.html) 提供,MindSpore Golden Stick是华为诺亚团队和MindSpore团队联合设计开发的一个模型压缩算法集,为MindSpore提供了一系列模型压缩算法,支持A16W8、A16W4、A8W8和KVCache等量化方式。详细资料可前往 [MindSpore Golden Stick官方资料](https://www.mindspore.cn/golden_stick/docs/zh-CN/master/index.html) 查看。 ## 模型量化基本流程 -- Gitee