From 2965b204593df9d5652313386ec280ffbfd1753b Mon Sep 17 00:00:00 2001 From: Manuel Faysse <43467008+ManuelFay@users.noreply.github.com> Date: Tue, 4 Jun 2024 16:46:19 +0200 Subject: [PATCH] add no split modules for xlmrobertaxl (#31223) --- .../models/xlm_roberta_xl/modeling_xlm_roberta_xl.py | 1 + 1 file changed, 1 insertion(+) diff --git a/src/transformers/models/xlm_roberta_xl/modeling_xlm_roberta_xl.py b/src/transformers/models/xlm_roberta_xl/modeling_xlm_roberta_xl.py index bd8ec70e8..7534eb4da 100644 --- a/src/transformers/models/xlm_roberta_xl/modeling_xlm_roberta_xl.py +++ b/src/transformers/models/xlm_roberta_xl/modeling_xlm_roberta_xl.py @@ -572,6 +572,7 @@ class XLMRobertaXLPreTrainedModel(PreTrainedModel): config_class = XLMRobertaXLConfig base_model_prefix = "roberta" + _no_split_modules = ["XLMRobertaXLEmbeddings", "XLMRobertaXLSelfAttention"] # Copied from transformers.models.bert.modeling_bert.BertPreTrainedModel._init_weights def _init_weights(self, module):