diff --git a/src/transformers/trainer_seq2seq.py b/src/transformers/trainer_seq2seq.py index 08280df10..17ad27bbb 100644 --- a/src/transformers/trainer_seq2seq.py +++ b/src/transformers/trainer_seq2seq.py @@ -277,7 +277,7 @@ class Seq2SeqTrainer(Trainer): self.model.generation_config._from_model_config = False # Retrieves GenerationConfig from model.generation_config - gen_config = model.generation_config + gen_config = self.model.generation_config # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_config.max_length: generated_tokens = self._pad_tensors_to_max_len(generated_tokens, gen_config.max_length)