pszemraj commited on
Commit
60ba70f
·
1 Parent(s): c183875

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -5
README.md CHANGED
@@ -45,26 +45,36 @@ library_name: transformers
45
 
46
  # distilgpt2-HC3
47
 
 
 
 
 
 
 
 
48
  This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the "chatgpt answers" column of the `Hello-SimpleAI/HC3` dataset.
49
 
50
  It achieves the following results on the evaluation set:
51
  - Loss: 1.9983
52
  - Accuracy: 0.5441
53
 
54
- ## Model description
55
-
56
- More information needed
57
 
58
  ## Intended uses & limitations
59
 
60
- More information needed
61
 
62
  ## Training and evaluation data
63
 
64
- More information needed
65
 
 
 
 
 
 
66
  ## Training procedure
67
 
 
68
  ### Training hyperparameters
69
 
70
  The following hyperparameters were used during training:
 
45
 
46
  # distilgpt2-HC3
47
 
48
+
49
+ > what happens if you train a smaller model on a dataset of chatGPT responses?
50
+
51
+ This happens.
52
+
53
+ ## Model description
54
+
55
  This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the "chatgpt answers" column of the `Hello-SimpleAI/HC3` dataset.
56
 
57
  It achieves the following results on the evaluation set:
58
  - Loss: 1.9983
59
  - Accuracy: 0.5441
60
 
 
 
 
61
 
62
  ## Intended uses & limitations
63
 
64
+ Despite how it sounds, this model only has 80m parameters and will likely not be factually accurate most of the time.
65
 
66
  ## Training and evaluation data
67
 
68
+ Modifications made w.r.t. original dataset:
69
 
70
+ - drop all rows that did not have a chatGPT answer
71
+ - if a row (_i.e. ELI5 question, etc_) had more than one response (_from chatGPT_), randomly choose one of the responses as the answer to the question
72
+ - the "question" and chatGPT answer were combined into a single string for that row as follows: `QUESTION_TEXT <answer> CHATGPT_ANSWER_TEXT <end_answer>`
73
+ - `<answer>` and `<end_answer>` serve as added tokens to help the model learn "turns" in the conversation
74
+
75
  ## Training procedure
76
 
77
+
78
  ### Training hyperparameters
79
 
80
  The following hyperparameters were used during training: