|
--- |
|
language: |
|
- |
|
- |
|
thumbnail: |
|
tags: |
|
- |
|
- |
|
- |
|
license: |
|
datasets: |
|
- |
|
- |
|
metrics: |
|
- |
|
- |
|
--- |
|
|
|
# Toxic language detection |
|
|
|
## Model description |
|
|
|
A toxic language detection model trained on tweets. The base model is Roberta-large. For more information, |
|
including the **training data**, **limitations and bias**, please refer to the [paper](https://arxiv.org/pdf/2102.00086.pdf) and |
|
Github [repo](https://github.com/XuhuiZhou/Toxic_Debias) for more details. |
|
|
|
#### How to use |
|
Note that LABEL_1 means toxic and LABEL_0 means non-toxic in the output. |
|
|
|
```python |
|
from transformers import pipeline |
|
classifier = pipeline("text-classification",model='Xuhui/ToxDect-roberta-large', return_all_scores=True) |
|
prediction = classifier("You are f**king stupid!", ) |
|
print(prediction) |
|
|
|
""" |
|
Output: |
|
[[{'label': 'LABEL_0', 'score': 0.002632011892274022}, {'label': 'LABEL_1', 'score': 0.9973680377006531}]] |
|
""" |
|
``` |
|
|
|
## Training procedure |
|
The random seed for this model is 22. For other details, please refer to the Github [repo](https://github.com/XuhuiZhou/Toxic_Debias) for more details. |
|
|
|
### BibTeX entry and citation info |
|
|
|
```bibtex |
|
@inproceedings{zhou-etal-2020-debiasing, |
|
title = {Challenges in Automated Debiasing for Toxic Language Detection}, |
|
author = {Zhou, Xuhui and Sap, Maarten and Swayamdipta, Swabha and Choi, Yejin and Smith, Noah A.}, |
|
booktitle = {EACL}, |
|
abbr = {EACL}, |
|
html = {https://www.aclweb.org/anthology/2021.eacl-main.274.pdf}, |
|
code = {https://github.com/XuhuiZhou/Toxic_Debias}, |
|
year = {2021}, |
|
bibtex_show = {true}, |
|
selected = {true} |
|
} |
|
``` |