File size: 485 Bytes
6f02411 f3af0bb 6f02411 2e31b3c f3af0bb |
1 2 3 4 5 6 7 8 9 10 11 |
---
license: mit
language:
- en
metrics:
- accuracy
---
# Toxic post classification using DistilBert
Use a pretrained DistilBert to train a classifier on the Toxic Comment dataset https://www.kaggle.com/c/jigsaw-toxic-comment-classification-challenge.
The goal is to classify whether a comment is toxic or not. Note that the labels from the original datasets are more fine-grained (i.e. different types of toxicity).
The model here obatains a test accuracy of 95% on a balanced split. |