Edit model card

GPT-2-medium fine-tuned for Sentiment Analysis πŸ‘πŸ‘Ž

OpenAI's GPT-2 medium fine-tuned on SST-2 dataset for Sentiment Analysis downstream task.

Details of GPT-2

The GPT-2 model was presented in Language Models are Unsupervised Multitask Learners by Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever

Model fine-tuning πŸ‹οΈβ€

The model has been finetuned for 10 epochs on standard hyperparameters

Val set metrics 🧾

           |precision | recall  | f1-score |support|
|----------|----------|---------|----------|-------|
|negative  |     0.92 |     0.92|      0.92|   428 |
|positive  |     0.92 |     0.93|      0.92|   444 |
|----------|----------|---------|----------|-------|
|accuracy|            |         |      0.92|   872 |
|macro avg|       0.92|     0.92|      0.92|   872 |
|weighted avg|    0.92|     0.92|      0.92|   872 |

Model in Action πŸš€

from transformers import GPT2Tokenizer, GPT2ForSequenceClassification

tokenizer = GPT2Tokenizer.from_pretrained("michelecafagna26/gpt2-medium-finetuned-sst2-sentiment")
model = GPT2ForSequenceClassification.from_pretrained("michelecafagna26/gpt2-medium-finetuned-sst2-sentiment")

inputs = tokenizer("I love it", return_tensors="pt")

model(**inputs).logits.argmax(axis=1)

# 1: Positive, 0: Negative
# Output: tensor([1])

This model card is based on "mrm8488/t5-base-finetuned-imdb-sentiment" by Manuel Romero/@mrm8488

Downloads last month
245
Safetensors
Model size
380M params
Tensor type
F32
Β·
U8
Β·
Inference Examples
Inference API (serverless) is not available, repository is disabled.

Dataset used to train michelecafagna26/gpt2-medium-finetuned-sst2-sentiment