JustinLin610
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -18,13 +18,14 @@ We're excited to unveil **Qwen2-VL**, the latest iteration of our Qwen-VL model,
|
|
18 |
|
19 |
#### Key Enhancements:
|
20 |
|
21 |
-
* **
|
22 |
|
23 |
-
* **
|
24 |
|
25 |
-
* **
|
|
|
|
|
26 |
|
27 |
-
* **Expanded Multilingual Support**: We've broadened our language capabilities to better serve a diverse global user base, making Qwen2-VL more accessible and effective across different linguistic contexts.
|
28 |
|
29 |
#### Model Architecture Updates:
|
30 |
|
|
|
18 |
|
19 |
#### Key Enhancements:
|
20 |
|
21 |
+
* **SoTA understanding of images of various resolution & ratio**: Qwen2-VL achieves state-of-the-art performance on visual understanding benchmarks, including MathVista, DocVQA, RealworldQA, MMBench, etc.
|
22 |
|
23 |
+
* **Understanding videos of 20min+**: with the online streaming capabilities, Qwen2-VL can understand long videos by high-quality video-based question answering, dialog, content creation, etc.
|
24 |
|
25 |
+
* **Agent that can operate your mobiles, robots, ...**: with the abilities of complex reasoning and decision making, Qwen2-VL can be integrated with devices like mobile phones, robots, etc., for automatic operation based on visual environment and text instructions.
|
26 |
+
|
27 |
+
* **Multilingual Support**: to serve global users, besides English and Chinese, Qwen2-VL now supports the understanding of texts in different languages inside images, including most European languages, Japanese, Korean, Arabic, Vietnamese, etc.
|
28 |
|
|
|
29 |
|
30 |
#### Model Architecture Updates:
|
31 |
|