Fancy-MLLM commited on
Commit
fb867be
·
verified ·
1 Parent(s): de3f7c6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -0
README.md CHANGED
@@ -7,6 +7,11 @@ base_model:
7
  pipeline_tag: image-text-to-text
8
  ---
9
 
 
 
 
 
 
10
  ## Model Overview
11
 
12
  This is a multimodal large language model fine-tuned from Qwen2.5-VL on the **R1-Onevision** dataset. The model enhances vision-language understanding and reasoning capabilities, making it suitable for various tasks such as visual reasoning, image understanding. With its robust ability to perform multimodal reasoning, R1-Onevision emerges as a powerful AI assistant capable of addressing a wide range of problem-solving challenges across different domains.
 
7
  pipeline_tag: image-text-to-text
8
  ---
9
 
10
+ ## R1-Onevision
11
+
12
+ [\[📂 GitHub\]](https://github.com/Fancy-MLLM/R1-Onevision)
13
+ [\[🤗 HF Dataset\]](https://huggingface.co/datasets/Fancy-MLLM/R1-onevision) [\[🤗 Reasoning Benchmark\]](https://huggingface.co/datasets/Fancy-MLLM/R1-OneVision-Bench) [\[🤗 HF Demo\]](https://huggingface.co/spaces/Fancy-MLLM/R1-OneVision)
14
+
15
  ## Model Overview
16
 
17
  This is a multimodal large language model fine-tuned from Qwen2.5-VL on the **R1-Onevision** dataset. The model enhances vision-language understanding and reasoning capabilities, making it suitable for various tasks such as visual reasoning, image understanding. With its robust ability to perform multimodal reasoning, R1-Onevision emerges as a powerful AI assistant capable of addressing a wide range of problem-solving challenges across different domains.