diff --git a/model_cards/mrm8488/t5-base-finetuned-squadv2/README.md b/model_cards/mrm8488/t5-base-finetuned-squadv2/README.md index f199273e7..d842e6562 100644 --- a/model_cards/mrm8488/t5-base-finetuned-squadv2/README.md +++ b/model_cards/mrm8488/t5-base-finetuned-squadv2/README.md @@ -51,13 +51,13 @@ The training script is a slightly modified version of [this one](https://colab.r ## Model in Action 🚀 ```python -from transformers import AutoModelWithLMHead, AutoTokenizer +from transformers import AutoModelForSeq2SeqLM, AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("mrm8488/t5-base-finetuned-squadv2") -model = AutoModelWithLMHead.from_pretrained("mrm8488/t5-base-finetuned-squadv2") +model = AutoModelForSeq2SeqLM.from_pretrained("mrm8488/t5-base-finetuned-squadv2") def get_answer(question, context): - input_text = "question: %s context: %s " % (question, context) + input_text = "question: %s context: %s" % (question, context) features = tokenizer([input_text], return_tensors='pt') output = model.generate(input_ids=features['input_ids'],