πŸ›‘οΈ Toxicity-RoBERTa-Base

A fine-tuned transformer model built on top of roberta-base to detect toxic content in text β€” including insults, threats, hate speech, and offensive language.
The model is lightweight, accurate, and ideal for real-time moderation tasks.


🧩 Use Cases

This model is designed to flag toxic messages in:

  • 🧡 Social media comments and posts
  • πŸ› οΈ Developer forums and Discord/Slack bots
  • 🧠 LLM output moderation
  • 🧩 Community Q&A sites (like Reddit, Stack Overflow)
  • 🚨 User-generated content platforms (blogs, review sites, games)

πŸ” Model Summary

Attribute Details
Base Architecture roberta-base
Fine-tuned For Toxic vs. Non-toxic classification
Classes 0 = Non-toxic, 1 = Toxic
Language English (en)
Data Sources Custom dataset (multi-domain)
Framework πŸ€— Transformers
Total Parameters ~125M

πŸ“Š Performance

Metric Result
Accuracy 94%
F1 Score 93%
Precision 95%
Recall 91%

βš™οΈ Quick Start

πŸ’‘ Install Required Libraries

pip install transformers torch
Downloads last month
16
Safetensors
Model size
125M params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for raghavv2710/toxicity-roberta-base

Finetuned
(1815)
this model

Space using raghavv2710/toxicity-roberta-base 1