lgfunderburk commited on
Commit
90f5a69
1 Parent(s): 673ad75

add tokenizer info

Browse files
Files changed (1) hide show
  1. README.md +8 -10
README.md CHANGED
@@ -10,24 +10,21 @@ model-index:
10
 
11
  # distilbert-truncated
12
 
13
- This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the 20 Newsgroups dataset (http://qwone.com/~jason/20Newsgroups/).
14
  It achieves the following results on the evaluation set:
15
 
16
 
17
- ## Model description
18
-
19
- More information needed
20
-
21
- ## Intended uses & limitations
22
-
23
- More information needed
24
-
25
  ## Training and evaluation data
26
 
27
- More information needed
28
 
29
  ## Training procedure
30
 
 
 
 
 
 
31
  ### Training hyperparameters
32
 
33
  The following hyperparameters were used during training:
@@ -41,6 +38,7 @@ batches_per_epoch = 636
41
  total_train_steps = 1908
42
 
43
  Model accuracy 0.8337758779525757
 
44
  Model loss 0.568471074104309
45
 
46
  ### Framework versions
 
10
 
11
  # distilbert-truncated
12
 
13
+ This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the [20 Newsgroups dataset](http://qwone.com/~jason/20Newsgroups/).
14
  It achieves the following results on the evaluation set:
15
 
16
 
 
 
 
 
 
 
 
 
17
  ## Training and evaluation data
18
 
19
+ The data was split into training and testing: model trained on 90% of the data, and had a testing data size of 10% of the original dataset.
20
 
21
  ## Training procedure
22
 
23
+ DistilBERT has a maximum input length of 512, so with this in mind the following was performed:
24
+
25
+ 1. I used the`distilbert-base-uncased` pretrained model to initialize an `AutoTokenizer`.
26
+ 2. Setting a maximum length of 256, each entry in the training, testing and validation data was truncated if it exceeded the limit and padded if it didn't reach the limit.
27
+
28
  ### Training hyperparameters
29
 
30
  The following hyperparameters were used during training:
 
38
  total_train_steps = 1908
39
 
40
  Model accuracy 0.8337758779525757
41
+
42
  Model loss 0.568471074104309
43
 
44
  ### Framework versions