diff --git a/docs/mindformers/docs/source_en/feature/configuration.md b/docs/mindformers/docs/source_en/feature/configuration.md index d267a91d0699a40e65ad79c2ebf54a35be899868..3638e481b0a7e6281923066054fe7a6209658bbd 100644 --- a/docs/mindformers/docs/source_en/feature/configuration.md +++ b/docs/mindformers/docs/source_en/feature/configuration.md @@ -73,6 +73,8 @@ Since the configuration will vary from model to model, only the generic configur | model.model_config.output_scores | Set to include score before the input softmax for each forward generation when returning the result as a dictionary, defaults to `False`. | bool | | model.model_config.output_logits | Set to include the logits output by the model at each forward generation when returning results as a dictionary, defaults to `False`. | bool | | model.model_config.layers_per_stage | Set the number of transformer layers assigned to each stage when enabling the pipeline stage, default is `None`, which means the transformer layers are evenly distributed across each stage. The set value is a list of integers with a length equal to the number of pipeline stages, where the i-th element indicates the number of transformer layers assigned to the i-th stage. | list | +| model.model_config.bias_swiglu_fusion | Whether to use the swiglu fusion operator, defaults to `False`. | bool | +| model.model_config.apply_rope_fusion | Whether to use the fused RoPE kernel, defaults to `False`. | bool | ### MoE Configuration diff --git a/docs/mindformers/docs/source_zh_cn/feature/configuration.md b/docs/mindformers/docs/source_zh_cn/feature/configuration.md index 42cb2ed0628425d8bead1d0224a9d54d0df45196..b0f999c50e514fb3ea20321bdd22e0d2854ddc9d 100644 --- a/docs/mindformers/docs/source_zh_cn/feature/configuration.md +++ b/docs/mindformers/docs/source_zh_cn/feature/configuration.md @@ -73,6 +73,8 @@ Context配置主要用于指定[mindspore.set_context](https://www.mindspore.cn/ | model.model_config.output_scores | 是否以字典形式返回结果时,包含每次前向生成时的输入softmax前的分数,默认为`False`。 | bool | | model.model_config.output_logits | 是否以字典形式返回结果时,包含每次前向生成时模型输出的logits,默认为`False`。 | bool | | model.model_config.layers_per_stage | 设置开启pipeline stage时,每个stage分配到的transformer层数,默认为`None`,表示每个stage平均分配。设置的值为一个长度为pipeline stage数量的整数列表,第i位表示第i个stage被分配到的transformer层数。 | list | +| model.model_config.bias_swiglu_fusion | 是否使用swiglu融合算子,默认为`False`。 | bool | +| model.model_config.apply_rope_fusion | 是否使用RoPE融合算子,默认为`False`。 | bool | ### MoE配置