From 6fd2219f06ce822d1f06d4385c64f36de3589c9f Mon Sep 17 00:00:00 2001 From: little_nik Date: Thu, 10 Jul 2025 14:50:18 +0800 Subject: [PATCH] support ring attention for context parallel --- mindspeed_llm/mindspore/mindspore_adaptor.py | 9 +++++++++ 1 file changed, 9 insertions(+) diff --git a/mindspeed_llm/mindspore/mindspore_adaptor.py b/mindspeed_llm/mindspore/mindspore_adaptor.py index 934bad0bf..be2ceccdc 100644 --- a/mindspeed_llm/mindspore/mindspore_adaptor.py +++ b/mindspeed_llm/mindspore/mindspore_adaptor.py @@ -350,6 +350,15 @@ class MindSporeAdaptation(MegatronAdaptationABC): from mindspeed_llm.mindspore.core.transformer.dot_product_attention import flash_attention_forward MindSporeAdaptation.register('mindspeed_llm.core.transformer.dot_product_attention.flash_attention_forward', flash_attention_forward) + from mindspeed.mindspore.core.context_parallel.utils import general_out_update + from mindspeed.mindspore.ops.fusion_attention_v2 import npu_fusion_attention, npu_fusion_attention_grad + from mindspeed.mindspore.ops.npu_ring_attention_update import npu_ring_attention_update + MindSporeAdaptation.register('mindspeed.core.context_parallel.utils.general_out_update', general_out_update) + MindSporeAdaptation.register('mindspeed.core.context_parallel.ring_context_parallel.general_out_update', general_out_update) + MindSporeAdaptation.register('mindspeed.ops.fusion_attention_v2.npu_fusion_attention', npu_fusion_attention) + MindSporeAdaptation.register('mindspeed.ops.fusion_attention_v2.npu_fusion_attention_grad', npu_fusion_attention_grad) + MindSporeAdaptation.register('mindspeed.ops.npu_ring_attention_update.npu_ring_attention_update', npu_ring_attention_update) + @staticmethod def reparse_args(): """ -- Gitee