Update README.md
Browse files
README.md
CHANGED
@@ -7,6 +7,11 @@ base_model:
|
|
7 |
pipeline_tag: image-text-to-text
|
8 |
---
|
9 |
|
|
|
|
|
|
|
|
|
|
|
10 |
## Model Overview
|
11 |
|
12 |
This is a multimodal large language model fine-tuned from Qwen2.5-VL on the **R1-Onevision** dataset. The model enhances vision-language understanding and reasoning capabilities, making it suitable for various tasks such as visual reasoning, image understanding. With its robust ability to perform multimodal reasoning, R1-Onevision emerges as a powerful AI assistant capable of addressing a wide range of problem-solving challenges across different domains.
|
|
|
7 |
pipeline_tag: image-text-to-text
|
8 |
---
|
9 |
|
10 |
+
## R1-Onevision
|
11 |
+
|
12 |
+
[\[📂 GitHub\]](https://github.com/Fancy-MLLM/R1-Onevision)
|
13 |
+
[\[🤗 HF Dataset\]](https://huggingface.co/datasets/Fancy-MLLM/R1-onevision) [\[🤗 Reasoning Benchmark\]](https://huggingface.co/datasets/Fancy-MLLM/R1-OneVision-Bench) [\[🤗 HF Demo\]](https://huggingface.co/spaces/Fancy-MLLM/R1-OneVision)
|
14 |
+
|
15 |
## Model Overview
|
16 |
|
17 |
This is a multimodal large language model fine-tuned from Qwen2.5-VL on the **R1-Onevision** dataset. The model enhances vision-language understanding and reasoning capabilities, making it suitable for various tasks such as visual reasoning, image understanding. With its robust ability to perform multimodal reasoning, R1-Onevision emerges as a powerful AI assistant capable of addressing a wide range of problem-solving challenges across different domains.
|