Update README.md
Browse files
README.md
CHANGED
@@ -2,7 +2,54 @@
|
|
2 |
library_name: transformers
|
3 |
tags: []
|
4 |
---
|
|
|
5 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
# Model Card for Model ID
|
7 |
|
8 |
<!-- Provide a quick summary of what the model is/does. -->
|
@@ -196,4 +243,4 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
|
|
196 |
|
197 |
## Model Card Contact
|
198 |
|
199 |
-
[More Information Needed]
|
|
|
2 |
library_name: transformers
|
3 |
tags: []
|
4 |
---
|
5 |
+
# Wiki Toxic Comment Classification Model Card
|
6 |
|
7 |
+
## Model Details
|
8 |
+
|
9 |
+
| Model Name | Wiki Toxic |
|
10 |
+
| --- | --- |
|
11 |
+
| License | apache-2.0 |
|
12 |
+
| Dataset | wiki_toxic |
|
13 |
+
| Language | English |
|
14 |
+
|
15 |
+
## Model Metrics
|
16 |
+
|
17 |
+
| Metric | Value | Description |
|
18 |
+
| --- | --- | --- |
|
19 |
+
| Accuracy | 0.87 | Overall accuracy on the test set |
|
20 |
+
| Precision | 0.85 (0), 0.89 (1) | Precision for non-toxic and toxic classes |
|
21 |
+
| Recall | 0.90 (0), 0.85 (1) | Recall for non-toxic and toxic classes |
|
22 |
+
| F1-Score | 0.87 (0), 0.87 (1) | F1-Score for non-toxic and toxic classes |
|
23 |
+
| Macro Avg | Precision: 0.87 <br> Recall: 0.87 <br> F1-Score: 0.87 | Macro-averaged values across classes |
|
24 |
+
| Weighted Avg | Precision: 0.87 <br> Recall: 0.87 <br> F1-Score: 0.87 | Weighted-averaged values across classes |
|
25 |
+
| Support | 0: 175 <br> 1: 175 <br> Total: 350 | Support for each class |
|
26 |
+
|
27 |
+
## Model Description
|
28 |
+
|
29 |
+
This model has been trained on the wiki_toxic dataset, comprising comments from Wikipedia talk pages labeled as toxic or non-toxic. The model's performance is evaluated on a held-out test set, with results indicating a balanced performance across both classes.
|
30 |
+
|
31 |
+
Achieving an overall accuracy of 0.87, the model demonstrates a strong ability to classify toxic and non-toxic comments accurately. For the non-toxic class (0), the model excels in precision (0.91), indicating a low rate of false positives. Meanwhile, for the toxic class (1), the model's recall of 0.91 highlights its effectiveness in capturing the majority of toxic comments.
|
32 |
+
|
33 |
+
While the model performs well, there's room for enhancement. Improving precision for the toxic class and recall for the non-toxic class could further boost its performance. This may involve fine-tuning the model, incorporating additional features, or expanding the dataset to cover a broader range of toxic comment variations.
|
34 |
+
|
35 |
+
## Intended Uses & Limitations
|
36 |
+
|
37 |
+
The Wiki Toxic model is designed for comment classification tasks, specifically identifying toxic behavior in online discussions. It can be employed in moderation systems to flag potentially harmful comments, fostering a healthier online environment.
|
38 |
+
|
39 |
+
However, it's crucial to acknowledge that the model's performance is tied to the data it was trained on. As such, its effectiveness may vary with different datasets or comment styles. Additionally, the model doesn't consider context, user relationships, or nuances of language, which could impact its accuracy in real-world applications.
|
40 |
+
|
41 |
+
## Training Data
|
42 |
+
|
43 |
+
The wiki_toxic dataset serves as the training data for this model. It contains comments from Wikipedia talk pages, manually labeled as toxic or non-toxic by human annotators. This dataset offers a diverse range of comments, ensuring the model learns to identify toxic behavior effectively.
|
44 |
+
|
45 |
+
## Ethical Considerations
|
46 |
+
|
47 |
+
It is important to note that the model's performance is dependent on the quality and representativeness of the training data. As such, it may reflect biases present in the data, potentially leading to unfair or inaccurate predictions. Careful monitoring and ongoing evaluation are necessary to ensure the model's responsible use and address any ethical concerns.
|
48 |
+
|
49 |
+
## Acknowledgements
|
50 |
+
|
51 |
+
We would like to acknowledge the contributors who curated the wiki_toxic dataset and made it publicly available. Their efforts have significantly advanced the development of toxic comment classification models, fostering a safer online community.
|
52 |
+
<!--
|
53 |
# Model Card for Model ID
|
54 |
|
55 |
<!-- Provide a quick summary of what the model is/does. -->
|
|
|
243 |
|
244 |
## Model Card Contact
|
245 |
|
246 |
+
[More Information Needed] -->
|