From c6ab27ba65f194599a12abb2379fa01c33c01995 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E7=8E=8B=E5=87=AF=E5=AE=87?= Date: Wed, 6 Aug 2025 15:53:52 +0800 Subject: [PATCH] [built-in][mm]close internal_format to avoid Nz error --- PyTorch/built-in/mm/CLIP_for_PyTorch/run_clip.py | 2 ++ .../src/transformers/models/roberta/modeling_roberta.py | 2 +- 2 files changed, 3 insertions(+), 1 deletion(-) diff --git a/PyTorch/built-in/mm/CLIP_for_PyTorch/run_clip.py b/PyTorch/built-in/mm/CLIP_for_PyTorch/run_clip.py index 3375cb41b7..e0e325c48a 100644 --- a/PyTorch/built-in/mm/CLIP_for_PyTorch/run_clip.py +++ b/PyTorch/built-in/mm/CLIP_for_PyTorch/run_clip.py @@ -53,6 +53,8 @@ from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version from transformers.utils.versions import require_version +torch_npu.npu.config.allow_internal_format = False + logger = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. diff --git a/PyTorch/built-in/mm/CLIP_for_PyTorch/transformers/src/transformers/models/roberta/modeling_roberta.py b/PyTorch/built-in/mm/CLIP_for_PyTorch/transformers/src/transformers/models/roberta/modeling_roberta.py index 1741476cf3..df8f0b115c 100644 --- a/PyTorch/built-in/mm/CLIP_for_PyTorch/transformers/src/transformers/models/roberta/modeling_roberta.py +++ b/PyTorch/built-in/mm/CLIP_for_PyTorch/transformers/src/transformers/models/roberta/modeling_roberta.py @@ -867,7 +867,7 @@ class RobertaModel(RobertaPreTrainedModel): output_hidden_states=output_hidden_states, return_dict=return_dict, ) - sequence_output = encoder_outputs[0].view(bs, from_seq_len, -1) + sequence_output = encoder_outputs[0].view(bs, from_seq_len, -1).clone() pooled_output = self.pooler(sequence_output) if self.pooler is not None else None if not return_dict: -- Gitee