πŸ›‘οΈ Toxicity-RoBERTa-Base

A fine-tuned transformer model built on top of roberta-base to detect toxic content in text β€” including insults, threats, hate speech, and offensive language.
The model is lightweight, accurate, and ideal for real-time moderation tasks.


🧩 Use Cases

This model is designed to flag toxic messages in:

  • 🧡 Social media comments and posts
  • πŸ› οΈ Developer forums and Discord/Slack bots
  • 🧠 LLM output moderation
  • 🧩 Community Q&A sites (like Reddit, Stack Overflow)
  • 🚨 User-generated content platforms (blogs, review sites, games)

πŸ” Model Summary

Attribute Details
Base Architecture roberta-base
Fine-tuned For Toxic vs. Non-toxic classification
Classes 0 = Non-toxic, 1 = Toxic
Language English (en)
Data Sources Custom dataset (multi-domain)
Framework πŸ€— Transformers
Total Parameters ~125M

πŸ“Š Performance

Metric Result
Accuracy 94%
F1 Score 93%
Precision 95%
Recall 91%

βš™οΈ Quick Start

πŸ’‘ Install Required Libraries

pip install transformers torch
Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for raghavv2710/toxicity-roberta-base

Finetuned
(1962)
this model

Space using raghavv2710/toxicity-roberta-base 1