|
--- |
|
datasets: |
|
- syedkhalid076/Sentiment-Analysis-Over-sampled |
|
language: |
|
- en |
|
metrics: |
|
- accuracy: 0.9019906657776932 |
|
- accuracy |
|
model_name: RoBERTa Sentiment Analysis Model v2 |
|
base_model: roberta-base |
|
library_name: transformers |
|
tags: |
|
- Text Classification |
|
- Transformers |
|
- Safetensors |
|
- English |
|
- roberta |
|
- Inference Endpoints |
|
pipeline_tag: text-classification |
|
--- |
|
|
|
|
|
# RoBERTa Sentiment Analysis Model v2 |
|
|
|
This repository hosts a fine-tuned [RoBERTa](https://huggingface.co/roberta-base) model for sentiment analysis. The model classifies text into three categories: **Negative (0)**, **Neutral (1)**, and **Positive (2)**. It has been fine-tuned on the [syedkhalid076/Sentiment-Analysis-Over-sampled](https://huggingface.co/datasets/syedkhalid076/Sentiment-Analysis-Over-sampled) dataset and achieves high accuracy. |
|
The Model is Trained specifically for Feedback Sentiment Analysis for UX Research, but it does perform well on other Sentiment Analysis tasks. |
|
|
|
--- |
|
|
|
## Model Details |
|
|
|
- **Base Model**: [RoBERTa-base](https://huggingface.co/roberta-base) |
|
- **Number of Labels**: 3 (0:Negative, 1:Neutral, 2:Positive) |
|
- **Model Size**: 125M parameters |
|
- **Language**: English (`en`) |
|
- **Metrics**: Accuracy: **90.20%** |
|
- **Tensor Type**: FP32 |
|
- **Dataset**: [syedkhalid076/Sentiment-Analysis-Over-sampled](https://huggingface.co/datasets/syedkhalid076/Sentiment-Analysis-Over-sampled) |
|
- **Library**: [Transformers](https://github.com/huggingface/transformers) |
|
- **File Format**: [Safetensors](https://github.com/huggingface/safetensors) |
|
|
|
--- |
|
|
|
## Features |
|
|
|
- **Text Classification**: Identify the sentiment of input text as Negative, Neutral, or Positive. |
|
- **High Accuracy**: Achieves 90.20% accuracy on the evaluation dataset. |
|
- **Hosted on Hugging Face**: Ready-to-use inference endpoints for quick deployment. |
|
- **Efficient Inference**: Lightweight and efficient, supporting FP32 tensors. |
|
|
|
--- |
|
|
|
## Installation |
|
|
|
To use this model, ensure you have the `transformers` library installed: |
|
|
|
```bash |
|
pip install transformers |
|
``` |
|
|
|
--- |
|
|
|
## Usage |
|
|
|
Here’s how you can load the model and tokenizer and perform inference: |
|
|
|
```python |
|
from transformers import AutoTokenizer, AutoModelForSequenceClassification |
|
import torch |
|
|
|
# Load tokenizer and model |
|
tokenizer = AutoTokenizer.from_pretrained("syedkhalid076/RoBERTa-Sentimental-Analysis-Model") |
|
model = AutoModelForSequenceClassification.from_pretrained("syedkhalid076/RoBERTa-Sentimental-Analysis-Model") |
|
|
|
# Define input text |
|
text = "I absolutely love this product! It's fantastic." |
|
|
|
# Tokenize input |
|
inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True) |
|
|
|
# Perform inference |
|
outputs = model(**inputs) |
|
logits = outputs.logits |
|
predicted_class = torch.argmax(logits, dim=-1).item() |
|
|
|
# Print results |
|
sentiment_labels = {0: "Negative", 1: "Neutral", 2: "Positive"} |
|
print(f"Predicted sentiment: {sentiment_labels[predicted_class]}") |
|
``` |
|
|
|
--- |
|
|
|
## Dataset |
|
|
|
This model is fine-tuned on the [syedkhalid076/Sentiment-Analysis-Over-sampled](https://huggingface.co/datasets/syedkhalid076/Sentiment-Analysis-Over-sampled) dataset. The dataset has been carefully preprocessed and oversampled to ensure balanced label representation and improve model performance. |
|
|
|
--- |
|
|
|
## Performance |
|
|
|
The model was evaluated on a test set and achieved the following metrics: |
|
|
|
- **Accuracy**: 90.20% (0.9019906657776932) |
|
|
|
The evaluation strategy includes validation after each epoch and logging metrics for tracking training progress. |
|
|
|
--- |
|
|
|
## Inference Endpoints |
|
|
|
You can use the Hugging Face Inference API to deploy and test this model in production environments. |
|
|
|
--- |
|
|
|
|
|
## Author |
|
|
|
**Syed Khalid Hussain** |
|
UX Designer & Developer |
|
Specializing in crafting user-centric digital experiences. |