diff --git a/mindspeed_llm/mindspore/mindspore_adaptor.py b/mindspeed_llm/mindspore/mindspore_adaptor.py index 934bad0bfa7ac02681f700b40bd3123279371b73..be2ceccdcb38278352f6ade68b32b363c1e89d9f 100644 --- a/mindspeed_llm/mindspore/mindspore_adaptor.py +++ b/mindspeed_llm/mindspore/mindspore_adaptor.py @@ -350,6 +350,15 @@ class MindSporeAdaptation(MegatronAdaptationABC): from mindspeed_llm.mindspore.core.transformer.dot_product_attention import flash_attention_forward MindSporeAdaptation.register('mindspeed_llm.core.transformer.dot_product_attention.flash_attention_forward', flash_attention_forward) + from mindspeed.mindspore.core.context_parallel.utils import general_out_update + from mindspeed.mindspore.ops.fusion_attention_v2 import npu_fusion_attention, npu_fusion_attention_grad + from mindspeed.mindspore.ops.npu_ring_attention_update import npu_ring_attention_update + MindSporeAdaptation.register('mindspeed.core.context_parallel.utils.general_out_update', general_out_update) + MindSporeAdaptation.register('mindspeed.core.context_parallel.ring_context_parallel.general_out_update', general_out_update) + MindSporeAdaptation.register('mindspeed.ops.fusion_attention_v2.npu_fusion_attention', npu_fusion_attention) + MindSporeAdaptation.register('mindspeed.ops.fusion_attention_v2.npu_fusion_attention_grad', npu_fusion_attention_grad) + MindSporeAdaptation.register('mindspeed.ops.npu_ring_attention_update.npu_ring_attention_update', npu_ring_attention_update) + @staticmethod def reparse_args(): """