mirror of
https://github.com/saymrwulf/transformers.git
synced 2026-05-14 20:58:08 +00:00
add no split modules for xlmrobertaxl (#31223)
This commit is contained in:
parent
821b772ab9
commit
2965b20459
1 changed files with 1 additions and 0 deletions
|
|
@ -572,6 +572,7 @@ class XLMRobertaXLPreTrainedModel(PreTrainedModel):
|
|||
|
||||
config_class = XLMRobertaXLConfig
|
||||
base_model_prefix = "roberta"
|
||||
_no_split_modules = ["XLMRobertaXLEmbeddings", "XLMRobertaXLSelfAttention"]
|
||||
|
||||
# Copied from transformers.models.bert.modeling_bert.BertPreTrainedModel._init_weights
|
||||
def _init_weights(self, module):
|
||||
|
|
|
|||
Loading…
Reference in a new issue