Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[bioformer-cased-v1.0](https://huggingface.co/bioformers/bioformer-cased-v1.0) fined-tuned on the [SQuAD1](https://rajpurkar.github.io/SQuAD-explorer) dataset for 3 epochs.
|
2 |
+
|
3 |
+
The fine-tuning process was performed on a single P100 GPUs (16GB). The hyperparameters are:
|
4 |
+
|
5 |
+
```
|
6 |
+
max_seq_length=512
|
7 |
+
per_device_train_batch_size=16
|
8 |
+
gradient_accumulation_steps=1
|
9 |
+
total train batch size (w. parallel, distributed & accumulation) = 16
|
10 |
+
learning_rate=3e-5
|
11 |
+
num_train_epochs=2
|
12 |
+
```
|
13 |
+
|
14 |
+
## Evaluation results
|
15 |
+
|
16 |
+
```
|
17 |
+
"eval_exact_match": 78.55250709555345
|
18 |
+
"eval_f1": 85.91482799690257
|
19 |
+
```
|
20 |
+
|
21 |
+
Bioformer's performance is on par with [DistilBERT](https://arxiv.org/pdf/1910.01108.pdf) (EM/F1: 77.7/85.8),
|
22 |
+
although Bioformer was only pretrained on biomedical texts.
|
23 |
+
|
24 |
+
|
25 |
+
## Speed
|
26 |
+
In our experiments, the inference speed of Bioformer is 3x as fast as BERT-base/BioBERT/PubMedBERT, and is 40% faster than DistilBERT.
|
27 |
+
|
28 |
+
|