ningkko commited on
Commit
42e47b1
·
1 Parent(s): eea62fd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -18
README.md CHANGED
@@ -13,6 +13,22 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  This model is a fine-tuned version of [cardiffnlp/twitter-roberta-base-sentiment](https://huggingface.co/cardiffnlp/twitter-roberta-base-sentiment) on [COVID-CQ](https://github.com/eceveco/COVID-CQ), a dataset that contains 3-label annotated opinions (negative, neutral, and positive) of the tweet initiators regarding the use of Chloroquine or Hydroxychloroquine for the treatment or prevention of the coronavirus.
15
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  ## Model description
17
 
18
  "We developed two COVID-drug-stance RoBERTa-base models by fine-tuning a pre-trained Twitter-specific stance detection model on a stance data set called COVID-CQ. The data were divided into training-dev-test validation datasets with a 70:10:20 ratio. Model I (COVID-drug-stance-BERT) was trained on the original tweet data, and Model II (COVID-drug-stance-BERT-masked) was trained on tweets with drug names masked as “[mask]” for model generalizability on different drugs. The two models had similar performance on the COVID-19 validation set: COVID-drug-stance-BERT had an accuracy of 86.88%, and the masked model had an accuracy of 86.67%. The two models were then evaluated by predicting tweet initiators’ attitudes towards the drug mentioned in each tweet using randomly selected test sets (100 tweets) of each drug (Hydroxychloquine, Ivermectin, Molnupiravir, Remdesivir). As suggested by the evaluation in Table 2, Model I had better performance and was therefore used in this study".
@@ -28,24 +44,6 @@ This model is a fine-tuned version of [cardiffnlp/twitter-roberta-base-sentiment
28
 
29
  The model uploaded here is Model I.
30
 
31
- ## Intended uses & limitations
32
- Predict opinions (negative, neutral, and positive) of tweet initiators regarding the use of a drug for the treatment or prevention of the coronavirus.
33
-
34
- The model works better when there's only one drug name in a single tweet.
35
-
36
- ## Inference & understanding
37
- We followed COVID-CQ to use the following label representation:
38
- - 0 -> None/Neutral;
39
- - 1 -> Against;
40
- - 2 -> Favor
41
-
42
- Try these examples:
43
- - The gov's killing people by banning Ivm
44
- - Great news cheers everybody:) ivermectin proven to not work by rct lol
45
- ## Tutorial
46
-
47
- See our Github repo for [inference scripts](https://github.com/ningkko/COVID-drug/blob/main/stance_detection/inference.ipynb)
48
-
49
  ## Training and evaluation data
50
  COVID-CQ
51
 
 
13
 
14
  This model is a fine-tuned version of [cardiffnlp/twitter-roberta-base-sentiment](https://huggingface.co/cardiffnlp/twitter-roberta-base-sentiment) on [COVID-CQ](https://github.com/eceveco/COVID-CQ), a dataset that contains 3-label annotated opinions (negative, neutral, and positive) of the tweet initiators regarding the use of Chloroquine or Hydroxychloroquine for the treatment or prevention of the coronavirus.
15
 
16
+ ## Intended uses & limitations
17
+ Predict opinions (negative, neutral, and positive) of tweet initiators regarding the use of a drug for the treatment or prevention of the coronavirus. Note that having multiple drug names with different stances in a single tweet can confuse the model.
18
+
19
+ ## Inference & understanding
20
+ We followed COVID-CQ to use the following label representation:
21
+ - 0 -> None/Neutral;
22
+ - 1 -> Against;
23
+ - 2 -> Favor
24
+
25
+ Try these examples:
26
+ - The gov's killing people by banning Ivm
27
+ - Great news cheers everybody:) ivermectin proven to not work by rct lol
28
+
29
+ ## Tutorial
30
+ See our Github repo for [inference scripts](https://github.com/ningkko/COVID-drug/blob/main/stance_detection/inference.ipynb)
31
+
32
  ## Model description
33
 
34
  "We developed two COVID-drug-stance RoBERTa-base models by fine-tuning a pre-trained Twitter-specific stance detection model on a stance data set called COVID-CQ. The data were divided into training-dev-test validation datasets with a 70:10:20 ratio. Model I (COVID-drug-stance-BERT) was trained on the original tweet data, and Model II (COVID-drug-stance-BERT-masked) was trained on tweets with drug names masked as “[mask]” for model generalizability on different drugs. The two models had similar performance on the COVID-19 validation set: COVID-drug-stance-BERT had an accuracy of 86.88%, and the masked model had an accuracy of 86.67%. The two models were then evaluated by predicting tweet initiators’ attitudes towards the drug mentioned in each tweet using randomly selected test sets (100 tweets) of each drug (Hydroxychloquine, Ivermectin, Molnupiravir, Remdesivir). As suggested by the evaluation in Table 2, Model I had better performance and was therefore used in this study".
 
44
 
45
  The model uploaded here is Model I.
46
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  ## Training and evaluation data
48
  COVID-CQ
49