flaviagiammarino
commited on
Commit
•
dc082c8
1
Parent(s):
880bb40
Update README.md
Browse files
README.md
CHANGED
@@ -16,23 +16,18 @@ PubMedCLIP is a fine-tuned version of [CLIP](https://huggingface.co/docs/transfo
|
|
16 |
image–text pairs obtained from [PubMed](https://pubmed.ncbi.nlm.nih.gov/) articles.
|
17 |
|
18 |
## Model Details
|
19 |
-
PubMedCLIP was trained on the [Radiology Objects in COntext (ROCO)](https://link.springer.com/chapter/10.1007/978-3-030-01364-6_20) dataset, which provides
|
20 |
-
over 80,000 samples including diverse imaging modalities (such as ultrasound, X-Ray, MRI, etc.) from various human body regions (such as head, neck, spine, etc.)
|
21 |
-
captured from PubMed articles. The texts used for training were taken from the short captions (average length of 20 words) associated with images in the articles.
|
22 |
|
23 |
### Model Description
|
24 |
-
|
25 |
-
-
|
26 |
-
-
|
27 |
-
- **Language(s) (NLP):** {{ language | default("[More Information Needed]", true)}}
|
28 |
-
- **License:** {{ license | default("[More Information Needed]", true)}}
|
29 |
-
- **Finetuned from model [optional]:** {{ finetuned_from | default("[More Information Needed]", true)}}
|
30 |
|
31 |
-
|
|
|
32 |
|
33 |
-
|
34 |
|
35 |
-
- **Repository:**
|
36 |
- **Paper [optional]:** {{ paper | default("[More Information Needed]", true)}}
|
37 |
- **Demo [optional]:** {{ demo | default("[More Information Needed]", true)}}
|
38 |
|
|
|
16 |
image–text pairs obtained from [PubMed](https://pubmed.ncbi.nlm.nih.gov/) articles.
|
17 |
|
18 |
## Model Details
|
|
|
|
|
|
|
19 |
|
20 |
### Model Description
|
21 |
+
PubMedCLIP was trained on the [Radiology Objects in COntext (ROCO)](https://github.com/razorx89/roco-dataset) dataset, a large-scale multimodal medical imaging dataset.
|
22 |
+
The ROCO dataset includes diverse imaging modalities (such as ultrasound, X-Ray, MRI, etc.) from various human body regions (such as head, neck, spine, etc.)
|
23 |
+
captured from open-access PubMed articles. The texts used for training PubMedCLIP were taken from the short captions associated with the images in the dataset.
|
|
|
|
|
|
|
24 |
|
25 |
+
The authors of PubMedCLIP have released three different pre-trained models at this [link](https://1drv.ms/u/s!ApXgPqe9kykTgwD4Np3-f7ODAot8?e=zLVlJ2) using
|
26 |
+
ResNet-50, ResNet-50x4 and ViT32 as image encoders. This repository includes only the ViT32 variant.
|
27 |
|
28 |
+
### Model Sources
|
29 |
|
30 |
+
- **Repository:** [Official GitHub Repository](https://github.com/sarahESL/PubMedCLIP)
|
31 |
- **Paper [optional]:** {{ paper | default("[More Information Needed]", true)}}
|
32 |
- **Demo [optional]:** {{ demo | default("[More Information Needed]", true)}}
|
33 |
|