rv2307 commited on
Commit
cc20ff1
·
1 Parent(s): 227d23b
Files changed (1) hide show
  1. README.md +29 -10
README.md CHANGED
@@ -1,10 +1,29 @@
1
- ---
2
- license: apache-2.0
3
- datasets:
4
- - quora
5
- language:
6
- - en
7
- metrics:
8
- - accuracy
9
- library_name: transformers
10
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ distilbert-base-uncased Quora Duplicate Pair Detection
2
+
3
+ This model is fine tuned of version of distilbert-base-uncased on quora dataset for detecting duplicate sentences or questions.
4
+
5
+ Loss :: 0.111300
6
+ Accuracy :: 0.900740
7
+ F1 Score :: 0.868633
8
+
9
+ Model Description:
10
+ DistilBERT model is a distilled form of the BERT model. The size of a BERT model was reduced by 40% via knowledge distillation during the pre-training phase while retaining 97% of its language understanding abilities and being 60% faster.
11
+
12
+ Training and evaluation data:
13
+ "quora"
14
+
15
+ Training Hyper Parameters:
16
+ learning_rate = 3e-4,
17
+ per_device_train_batch_size=32,
18
+ per_device_eval_batch_size=32,
19
+ num_train_epochs=4,
20
+ evaluation_strategy="epoch",
21
+ seed: 42,
22
+ optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
23
+
24
+ Training Results:
25
+
26
+ Epoch Training Loss Validation Loss Accuracy F1
27
+ 1 0.271500 0.264808 0.884909 0.844402
28
+ 2 0.191200 0.258109 0.896399 0.866099
29
+ 3 0.111300 0.315554 0.900740 0.868633