π‘οΈ Toxicity-RoBERTa-Base
A fine-tuned transformer model built on top of roberta-base
to detect toxic content in text β including insults, threats, hate speech, and offensive language.
The model is lightweight, accurate, and ideal for real-time moderation tasks.
π§© Use Cases
This model is designed to flag toxic messages in:
- π§΅ Social media comments and posts
- π οΈ Developer forums and Discord/Slack bots
- π§ LLM output moderation
- π§© Community Q&A sites (like Reddit, Stack Overflow)
- π¨ User-generated content platforms (blogs, review sites, games)
π Model Summary
Attribute | Details |
---|---|
Base Architecture | roberta-base |
Fine-tuned For | Toxic vs. Non-toxic classification |
Classes | 0 = Non-toxic , 1 = Toxic |
Language | English (en ) |
Data Sources | Custom dataset (multi-domain) |
Framework | π€ Transformers |
Total Parameters | ~125M |
π Performance
Metric | Result |
---|---|
Accuracy | 94% |
F1 Score | 93% |
Precision | 95% |
Recall | 91% |
βοΈ Quick Start
π‘ Install Required Libraries
pip install transformers torch
- Downloads last month
- 16
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
Model tree for raghavv2710/toxicity-roberta-base
Base model
FacebookAI/roberta-base