mirror of
https://github.com/saymrwulf/transformers.git
synced 2026-05-14 20:58:08 +00:00
Fix FSMT weight sharing (#26292)
This commit is contained in:
parent
da971b2271
commit
26ba56ccbd
2 changed files with 21 additions and 2 deletions
|
|
@ -1056,8 +1056,9 @@ class FSMTModel(PretrainedFSMTModel):
|
|||
return self.decoder
|
||||
|
||||
def _tie_weights(self):
|
||||
self._tie_or_clone_weights(self.decoder.embed_tokens, self.get_input_embeddings())
|
||||
self._tie_or_clone_weights(self.decoder.output_projection, self.get_input_embeddings())
|
||||
if self.config.tie_word_embeddings:
|
||||
self._tie_or_clone_weights(self.decoder.embed_tokens, self.get_input_embeddings())
|
||||
self._tie_or_clone_weights(self.decoder.output_projection, self.get_input_embeddings())
|
||||
|
||||
@add_start_docstrings_to_model_forward(FSMT_INPUTS_DOCSTRING)
|
||||
@add_code_sample_docstrings(
|
||||
|
|
|
|||
|
|
@ -273,6 +273,8 @@ class FSMTModelTest(ModelTesterMixin, GenerationTesterMixin, PipelineTesterMixin
|
|||
|
||||
def test_ensure_weights_are_shared(self):
|
||||
config, inputs_dict = self.model_tester.prepare_config_and_inputs()
|
||||
|
||||
config.tie_word_embeddings = True
|
||||
model = FSMTForConditionalGeneration(config)
|
||||
|
||||
# FSMT shares three weights.
|
||||
|
|
@ -288,6 +290,22 @@ class FSMTModelTest(ModelTesterMixin, GenerationTesterMixin, PipelineTesterMixin
|
|||
1,
|
||||
)
|
||||
|
||||
config.tie_word_embeddings = False
|
||||
model = FSMTForConditionalGeneration(config)
|
||||
|
||||
# FSMT shares three weights.
|
||||
# Not an issue to not have these correctly tied for torch.load, but it is an issue for safetensors.
|
||||
self.assertEqual(
|
||||
len(
|
||||
{
|
||||
model.get_output_embeddings().weight.data_ptr(),
|
||||
model.get_input_embeddings().weight.data_ptr(),
|
||||
model.base_model.decoder.output_projection.weight.data_ptr(),
|
||||
}
|
||||
),
|
||||
2,
|
||||
)
|
||||
|
||||
@unittest.skip("can't be implemented for FSMT due to dual vocab.")
|
||||
def test_resize_tokens_embeddings(self):
|
||||
pass
|
||||
|
|
|
|||
Loading…
Reference in a new issue