π‘οΈ Toxicity-RoBERTa-Base
A fine-tuned transformer model built on top of roberta-base to detect toxic content in text β including insults, threats, hate speech, and offensive language.
The model is lightweight, accurate, and ideal for real-time moderation tasks.
π§© Use Cases
This model is designed to flag toxic messages in:
- π§΅ Social media comments and posts
- π οΈ Developer forums and Discord/Slack bots
- π§ LLM output moderation
- π§© Community Q&A sites (like Reddit, Stack Overflow)
- π¨ User-generated content platforms (blogs, review sites, games)
π Model Summary
| Attribute | Details |
|---|---|
| Base Architecture | roberta-base |
| Fine-tuned For | Toxic vs. Non-toxic classification |
| Classes | 0 = Non-toxic, 1 = Toxic |
| Language | English (en) |
| Data Sources | Custom dataset (multi-domain) |
| Framework | π€ Transformers |
| Total Parameters | ~125M |
π Performance
| Metric | Result |
|---|---|
| Accuracy | 94% |
| F1 Score | 93% |
| Precision | 95% |
| Recall | 91% |
βοΈ Quick Start
π‘ Install Required Libraries
pip install transformers torch
- Downloads last month
- 3
Model tree for raghavv2710/toxicity-roberta-base
Base model
FacebookAI/roberta-base