T5-base fine-tuned for Sentiment Analysis ππ
Google's T5 base fine-tuned on SST-2 dataset for Sentiment Analysis downstream task.
Details of T5
The T5 model was presented in Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer by Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, Peter J. Liu
Model fine-tuning ποΈβ
The model has been finetuned for 10 epochs on standard hyperparameters
Val set metrics π§Ύ
|precision | recall | f1-score |support|
|----------|----------|---------|----------|-------|
|negative | 0.95 | 0.95| 0.95| 428 |
|positive | 0.94 | 0.96| 0.95| 444 |
|----------|----------|---------|----------|-------|
|accuracy| | | 0.95| 872 |
|macro avg| 0.95| 0.95| 0.95| 872 |
|weighted avg| 0.95| 0.95| 0.95 | 872 |
Model in Action π
from transformers import T5Tokenizer, T5ForConditionalGeneration
tokenizer = T5Tokenizer.from_pretrained("t5-finetune-sst2")
model = T5ForConditionalGeneration.from_pretrained("t5-finetune-sst2")
def get_sentiment(text):
inputs = tokenizer("sentiment: " + text, max_length=128, truncation=True, return_tensors="pt").input_ids
preds = model.generate(inputs)
decoded_preds = tokenizer.batch_decode(sequences=preds, skip_special_tokens=True)
return decoded_preds
get_sentiment("This movie is awesome")
# labels are 'p' for 'positive' and 'n' for 'negative'
# Output: ['p']
This model card is based on "mrm8488/t5-base-finetuned-imdb-sentiment" by Manuel Romero/@mrm8488
- Downloads last month
- 639
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.