Update README.md
Browse files
README.md
CHANGED
@@ -10,6 +10,9 @@ metrics:
|
|
10 |
- accuracy
|
11 |
library_name: transformers
|
12 |
---
|
|
|
|
|
|
|
13 |
# Model Card for Euclid-convnext-xxlarge (Version on 12/05/2024)
|
14 |
|
15 |
A multimodal large language models specifically trained for strong low-level geometric perception.
|
@@ -26,8 +29,8 @@ It combines a ConvNeXt visual encoder with a Qwen-2.5 language model, connected
|
|
26 |
### Model Sources [optional]
|
27 |
|
28 |
- **Repository:** https://github.com/euclid-multimodal/Euclid
|
29 |
-
- **Paper:**
|
30 |
-
- **Demo:**
|
31 |
|
32 |
## Uses
|
33 |
|
|
|
10 |
- accuracy
|
11 |
library_name: transformers
|
12 |
---
|
13 |
+
|
14 |
+
Paper: [Euclid: Supercharging Multimodal LLMs with Synthetic High-Fidelity Visual Descriptions](https://arxiv.org/abs/2412.08737)
|
15 |
+
|
16 |
# Model Card for Euclid-convnext-xxlarge (Version on 12/05/2024)
|
17 |
|
18 |
A multimodal large language models specifically trained for strong low-level geometric perception.
|
|
|
29 |
### Model Sources [optional]
|
30 |
|
31 |
- **Repository:** https://github.com/euclid-multimodal/Euclid
|
32 |
+
- **Paper:** https://arxiv.org/abs/2412.08737
|
33 |
+
- **Demo:** https://euclid-multimodal.github.io/
|
34 |
|
35 |
## Uses
|
36 |
|