From fa894c5e9f9f238ffcd89312ac82c890c9237eb3 Mon Sep 17 00:00:00 2001 From: kongziyi <1045916357@qq.com> Date: Wed, 6 Aug 2025 15:20:51 +0800 Subject: [PATCH] =?UTF-8?q?=E3=80=90master=E3=80=91=E6=96=B0=E5=A2=9EMcore?= =?UTF-8?q?=E9=85=8D=E7=BD=AEROPE=E3=80=81SwiGLU=E8=9E=8D=E5=90=88?= =?UTF-8?q?=E7=AE=97=E5=AD=90=E5=8F=82=E6=95=B0=E8=AF=B4=E6=98=8E?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- docs/mindformers/docs/source_en/feature/configuration.md | 2 ++ docs/mindformers/docs/source_zh_cn/feature/configuration.md | 2 ++ 2 files changed, 4 insertions(+) diff --git a/docs/mindformers/docs/source_en/feature/configuration.md b/docs/mindformers/docs/source_en/feature/configuration.md index d267a91d06..3638e481b0 100644 --- a/docs/mindformers/docs/source_en/feature/configuration.md +++ b/docs/mindformers/docs/source_en/feature/configuration.md @@ -73,6 +73,8 @@ Since the configuration will vary from model to model, only the generic configur | model.model_config.output_scores | Set to include score before the input softmax for each forward generation when returning the result as a dictionary, defaults to `False`. | bool | | model.model_config.output_logits | Set to include the logits output by the model at each forward generation when returning results as a dictionary, defaults to `False`. | bool | | model.model_config.layers_per_stage | Set the number of transformer layers assigned to each stage when enabling the pipeline stage, default is `None`, which means the transformer layers are evenly distributed across each stage. The set value is a list of integers with a length equal to the number of pipeline stages, where the i-th element indicates the number of transformer layers assigned to the i-th stage. | list | +| model.model_config.bias_swiglu_fusion | Whether to use the swiglu fusion operator, defaults to `False`. | bool | +| model.model_config.apply_rope_fusion | Whether to use the fused RoPE kernel, defaults to `False`. | bool | ### MoE Configuration diff --git a/docs/mindformers/docs/source_zh_cn/feature/configuration.md b/docs/mindformers/docs/source_zh_cn/feature/configuration.md index 42cb2ed062..b0f999c50e 100644 --- a/docs/mindformers/docs/source_zh_cn/feature/configuration.md +++ b/docs/mindformers/docs/source_zh_cn/feature/configuration.md @@ -73,6 +73,8 @@ Context配置主要用于指定[mindspore.set_context](https://www.mindspore.cn/ | model.model_config.output_scores | 是否以字典形式返回结果时,包含每次前向生成时的输入softmax前的分数,默认为`False`。 | bool | | model.model_config.output_logits | 是否以字典形式返回结果时,包含每次前向生成时模型输出的logits,默认为`False`。 | bool | | model.model_config.layers_per_stage | 设置开启pipeline stage时,每个stage分配到的transformer层数,默认为`None`,表示每个stage平均分配。设置的值为一个长度为pipeline stage数量的整数列表,第i位表示第i个stage被分配到的transformer层数。 | list | +| model.model_config.bias_swiglu_fusion | 是否使用swiglu融合算子,默认为`False`。 | bool | +| model.model_config.apply_rope_fusion | 是否使用RoPE融合算子,默认为`False`。 | bool | ### MoE配置 -- Gitee