dardem commited on
Commit
f032736
·
1 Parent(s): bb2f862

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -0
README.md CHANGED
@@ -1,3 +1,51 @@
1
  ---
2
  license: afl-3.0
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: afl-3.0
3
+ task_categories:
4
+ - text-classification
5
+ language:
6
+ - en
7
  ---
8
+
9
+ # ParaDetox: Detoxification with Parallel Data (English). Toxicity Task Results
10
+
11
+ This repository contains information about **Toxicity Task** markup from [English Paradetox dataset](https://huggingface.co/datasets/s-nlp/paradetox) collection pipeline.
12
+ The original paper ["ParaDetox: Detoxification with Parallel Data"](https://aclanthology.org/2022.acl-long.469/) was presented at ACL 2022 main conference.
13
+
14
+ ## ParaDetox Collection Pipeline
15
+
16
+ The ParaDetox Dataset collection was done via [Yandex.Toloka](https://toloka.yandex.com/) crowdsource platform. The collection was done in three steps:
17
+ * *Task 1:* **Generation of Paraphrases**: The first crowdsourcing task asks users to eliminate toxicity in a given sentence while keeping the content.
18
+ * *Task 2:* **Content Preservation Check**: We show users the generated paraphrases along with their original variants and ask them to indicate if they have close meanings.
19
+ * *Task 3:* **Toxicity Check**: Finally, we check if the workers succeeded in removing toxicity.
20
+
21
+ Specifically this repo contains the results of **Task 3: Toxicity Check**. Here, the samples with markup confidence >= 90 are present.
22
+ The input here is text and the label shows if the text is toxic or not.
23
+ Totally, datasets contains 26,507 samples. Among them, the minor part is toxic examples (4,009 pairs).
24
+
25
+ ## Citation
26
+
27
+ ```
28
+ @inproceedings{logacheva-etal-2022-paradetox,
29
+ title = "{P}ara{D}etox: Detoxification with Parallel Data",
30
+ author = "Logacheva, Varvara and
31
+ Dementieva, Daryna and
32
+ Ustyantsev, Sergey and
33
+ Moskovskiy, Daniil and
34
+ Dale, David and
35
+ Krotova, Irina and
36
+ Semenov, Nikita and
37
+ Panchenko, Alexander",
38
+ booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
39
+ month = may,
40
+ year = "2022",
41
+ address = "Dublin, Ireland",
42
+ publisher = "Association for Computational Linguistics",
43
+ url = "https://aclanthology.org/2022.acl-long.469",
44
+ pages = "6804--6818",
45
+ abstract = "We present a novel pipeline for the collection of parallel data for the detoxification task. We collect non-toxic paraphrases for over 10,000 English toxic sentences. We also show that this pipeline can be used to distill a large existing corpus of paraphrases to get toxic-neutral sentence pairs. We release two parallel corpora which can be used for the training of detoxification models. To the best of our knowledge, these are the first parallel datasets for this task.We describe our pipeline in detail to make it fast to set up for a new language or domain, thus contributing to faster and easier development of new parallel resources.We train several detoxification models on the collected data and compare them with several baselines and state-of-the-art unsupervised approaches. We conduct both automatic and manual evaluations. All models trained on parallel data outperform the state-of-the-art unsupervised models by a large margin. This suggests that our novel datasets can boost the performance of detoxification systems.",
46
+ }
47
+ ```
48
+
49
+ ## Contacts
50
+
51
+ For any questions, please contact: Daryna Dementieva ([email protected])