mfarrington commited on
Commit
990661d
1 Parent(s): 128305a

End of training

Browse files
README.md CHANGED
@@ -1,66 +1,74 @@
1
  ---
2
- license: mit
3
- datasets:
4
- - mfarrington/biobert-ner-fda-recalls-dataset
5
  metrics:
 
 
 
6
  - accuracy
7
- pipeline_tag: token-classification
8
- tags:
9
- - medical
10
  ---
11
- # Model Card for Model ID
12
-
13
- Pretrained BioBERT Model for Performing Named Entity Recognition (NER) of Medical Device Names, Components and Part Numbers.
14
-
15
- ## Model Details
16
-
17
- ### Abstract
18
- *FDA Medical Device recalls are critical and time-sensitive events, requiring
19
- swift identification of impacted devices to inform the public of a recall event and
20
- ensure patient safety. The OpenFDA device recall dataset contains valuable
21
- information about ongoing device recall actions, but manually extracting relevant
22
- device information from the recall action summaries is a time-consuming task.
23
- Named Entity Recognition (NER) is a task in Natural Language Processing
24
- (NLP) that involves identifying and categorizing named entities in unstructured text.
25
- Existing NER models, including domain-specific models like BioBERT, struggle
26
- to correctly identify medical device trade names, part numbers and component terms within these
27
- summaries. To address this, we propose DeviceBERT, a medical device annotation, pre-processing and enrichment pipeline, which builds on BioBERT to identify and label medical device terminology in the device recall summaries with improved accuracy. Furthermore, we demonstrate that our approach can be applied effectively for performing entity recognition tasks where training data is limited or sparse.*
28
-
29
- ### Model Description
30
 
31
- This model was created as part of a student final project for Stanford CS224N Spring 2024.
32
- Project Title: "Optimizing BioBERT to Perform Named Entity Recognition of Medical Devices Using FDA Device Recall Summaries"
33
 
34
- DeviceBERT utilizes BioBERT (dmis-lab/biobert-base-cased-v1.2) which has been trained on PubMed and PMC and subsequently finetuned to accurately identify industry specific medical device trade names, component parts and part numbers.
35
 
36
- The model was trained utilizing a processed and annotated NER dataset created using the OpenFDA Device Recalls Dataset (https://open.fda.gov/apis/device/recall/), and further
37
- tokenized using the DistilBERT AutoTokenizer. It can be used to perform inferencing to accurately identfiy and label medical device, device component, part number and trade name related terms in a variety of downstream applications.
 
 
 
 
 
38
 
39
- - **Developed by:** Miriam Farrington for CS224N
40
- - **Model type:** Deep Learning Language Model/LLM
41
- - **Language(s) (NLP):** Python, TensorFlow
42
- - **License:** MIT
43
- - **Finetuned from model [optional]:** BioBERT (dmis-lab/biobert-base-cased-v1.2)
44
 
45
- ### Model Sources [optional]
46
 
47
- - **Repository:** https://github.com/mmfarrington/devicebert-ner-project
48
- - **Paper:** "DeviceBERT: Applied Transfer Learning With Targeted Annotations and Vocabulary Enrichment to Identify Medical Device and Component Terminology in FDA Recall Summaries"
49
 
50
- ## Uses
51
 
52
- ### Direct Use
53
 
54
- This model can be directly used to perform NER inferencing on text to identify medical device related terms, trade names, components and part numbers in a variety of downstream tasks.
55
- The model can be applied further to generate human feedback loops using inferencing to generate additional NER data for more complex downstream tasks or additional finetuning.
56
 
57
- ## Training Details
58
 
59
- ### Training Data
60
 
61
- mfarrington/biobert-ner-fda-recalls-dataset
 
 
 
 
 
 
 
62
 
 
63
 
 
 
 
 
 
 
 
 
 
 
 
 
64
 
65
 
 
66
 
 
 
 
 
 
1
  ---
2
+ base_model: dmis-lab/biobert-base-cased-v1.2
3
+ tags:
4
+ - generated_from_trainer
5
  metrics:
6
+ - precision
7
+ - recall
8
+ - f1
9
  - accuracy
10
+ model-index:
11
+ - name: devicebert-base-cased-v1.0
12
+ results: []
13
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
 
18
+ # devicebert-base-cased-v1.0
19
 
20
+ This model is a fine-tuned version of [dmis-lab/biobert-base-cased-v1.2](https://huggingface.co/dmis-lab/biobert-base-cased-v1.2) on an unknown dataset.
21
+ It achieves the following results on the evaluation set:
22
+ - Loss: nan
23
+ - Precision: 0.6816
24
+ - Recall: 0.6691
25
+ - F1: 0.6753
26
+ - Accuracy: 0.8547
27
 
28
+ ## Model description
 
 
 
 
29
 
30
+ More information needed
31
 
32
+ ## Intended uses & limitations
 
33
 
34
+ More information needed
35
 
36
+ ## Training and evaluation data
37
 
38
+ More information needed
 
39
 
40
+ ## Training procedure
41
 
42
+ ### Training hyperparameters
43
 
44
+ The following hyperparameters were used during training:
45
+ - learning_rate: 1e-05
46
+ - train_batch_size: 16
47
+ - eval_batch_size: 16
48
+ - seed: 42
49
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
+ - lr_scheduler_type: linear
51
+ - num_epochs: 10
52
 
53
+ ### Training results
54
 
55
+ | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
56
+ |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
57
+ | No log | 1.0 | 101 | nan | 0.5981 | 0.5740 | 0.5858 | 0.8131 |
58
+ | No log | 2.0 | 202 | nan | 0.6673 | 0.6197 | 0.6427 | 0.8424 |
59
+ | No log | 3.0 | 303 | nan | 0.6926 | 0.6673 | 0.6797 | 0.8498 |
60
+ | No log | 4.0 | 404 | nan | 0.686 | 0.6271 | 0.6552 | 0.8473 |
61
+ | 0.3891 | 5.0 | 505 | nan | 0.6853 | 0.6490 | 0.6667 | 0.8539 |
62
+ | 0.3891 | 6.0 | 606 | nan | 0.6857 | 0.7020 | 0.6938 | 0.8563 |
63
+ | 0.3891 | 7.0 | 707 | nan | 0.6900 | 0.6673 | 0.6784 | 0.8580 |
64
+ | 0.3891 | 8.0 | 808 | nan | 0.6795 | 0.6782 | 0.6789 | 0.8514 |
65
+ | 0.3891 | 9.0 | 909 | nan | 0.6906 | 0.6691 | 0.6797 | 0.8571 |
66
+ | 0.1315 | 10.0 | 1010 | nan | 0.6816 | 0.6691 | 0.6753 | 0.8547 |
67
 
68
 
69
+ ### Framework versions
70
 
71
+ - Transformers 4.41.2
72
+ - Pytorch 2.3.0+cu121
73
+ - Datasets 2.19.2
74
+ - Tokenizers 0.19.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fcc925315e0e239dddfa9cfee10132c820ce176fba0f68ed7368e2b7b3ef8bcf
3
  size 928741200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d5ac909e05c9af2fedd3552dc8c5b4c4357d3155f34bac3d225d04bce007404
3
  size 928741200
runs/Jun07_18-40-05_6d0acad3a460/events.out.tfevents.1717785615.6d0acad3a460.17459.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4dc79f61278c6cdde8a6c17db07e1156aebe41e73cf7c99782bb8a46aee780bf
3
- size 9453
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6fbb28ea6c9a34b9f30d89c36c94159ccd47563a6dd8cd9486ab0aed78bf260
3
+ size 10490