Update README.md
Browse files
README.md
CHANGED
@@ -10,6 +10,7 @@ pipeline_tag: image-text-to-text
|
|
10 |
## R1-Onevision
|
11 |
|
12 |
[\[📂 GitHub\]](https://github.com/Fancy-MLLM/R1-Onevision)[\[📝 Report\]](https://yangyi-vai.notion.site/r1-onevision?pvs=4)
|
|
|
13 |
[\[🤗 HF Dataset\]](https://huggingface.co/datasets/Fancy-MLLM/R1-onevision) [\[🤗 Reasoning Benchmark\]](https://huggingface.co/datasets/Fancy-MLLM/R1-OneVision-Bench) [\[🤗 HF Demo\]](https://huggingface.co/spaces/Fancy-MLLM/R1-OneVision)
|
14 |
|
15 |
## Model Overview
|
@@ -102,6 +103,7 @@ print(output_text)
|
|
102 |
We are also focused on integrating Chinese multimodal reasoning CoT data into the training process. By adding this language-specific dataset, we aim to improve the model’s capability to perform reasoning tasks in Chinese, expanding its multilingual and multimodal reasoning proficiency.
|
103 |
|
104 |
4. **Release of the 3B Model**
|
|
|
105 |
|
106 |
We are working on the release of a smaller, more efficient 3B model, which is designed to provide a balance between performance and resource efficiency. This model aims to deliver strong multimodal reasoning capabilities while being more accessible and optimized for environments with limited computational resources, offering a more compact alternative to the current 7B model.
|
107 |
|
|
|
10 |
## R1-Onevision
|
11 |
|
12 |
[\[📂 GitHub\]](https://github.com/Fancy-MLLM/R1-Onevision)[\[📝 Report\]](https://yangyi-vai.notion.site/r1-onevision?pvs=4)
|
13 |
+
[\[🤗 HF Demo\]](https://huggingface.co/spaces/Fancy-MLLM/R1-Onevision)
|
14 |
[\[🤗 HF Dataset\]](https://huggingface.co/datasets/Fancy-MLLM/R1-onevision) [\[🤗 Reasoning Benchmark\]](https://huggingface.co/datasets/Fancy-MLLM/R1-OneVision-Bench) [\[🤗 HF Demo\]](https://huggingface.co/spaces/Fancy-MLLM/R1-OneVision)
|
15 |
|
16 |
## Model Overview
|
|
|
103 |
We are also focused on integrating Chinese multimodal reasoning CoT data into the training process. By adding this language-specific dataset, we aim to improve the model’s capability to perform reasoning tasks in Chinese, expanding its multilingual and multimodal reasoning proficiency.
|
104 |
|
105 |
4. **Release of the 3B Model**
|
106 |
+
|
107 |
|
108 |
We are working on the release of a smaller, more efficient 3B model, which is designed to provide a balance between performance and resource efficiency. This model aims to deliver strong multimodal reasoning capabilities while being more accessible and optimized for environments with limited computational resources, offering a more compact alternative to the current 7B model.
|
109 |
|