FanqingM commited on
Commit
4832d7f
1 Parent(s): e02d8b6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -0
README.md CHANGED
@@ -13,6 +13,16 @@
13
  ## Introduction
14
  MMIU encompasses 7 types of multi-image relationships, 52 tasks, 77K images, and 11K meticulously curated multiple-choice questions, making it the most extensive benchmark of its kind. Our evaluation of 24 popular MLLMs, including both open-source and proprietary models, reveals significant challenges in multi-image comprehension, particularly in tasks involving spatial understanding. Even the most advanced models, such as GPT-4o, achieve only 55.7% accuracy on MMIU. Through multi-faceted analytical experiments, we identify key performance gaps and limitations, providing valuable insights for future model and data improvements. We aim for MMIU to advance the frontier of LVLM research and development, moving us toward achieving sophisticated multimodal multi-image user interactions.
15
 
 
 
 
 
 
 
 
 
 
 
16
  ## Data Structure
17
  ### Data Fields
18
 
@@ -28,6 +38,7 @@ Each field of annotation is as follows:
28
  * `output`: The correct option for the question
29
 
30
 
 
31
  ### Example
32
  ```
33
  {
 
13
  ## Introduction
14
  MMIU encompasses 7 types of multi-image relationships, 52 tasks, 77K images, and 11K meticulously curated multiple-choice questions, making it the most extensive benchmark of its kind. Our evaluation of 24 popular MLLMs, including both open-source and proprietary models, reveals significant challenges in multi-image comprehension, particularly in tasks involving spatial understanding. Even the most advanced models, such as GPT-4o, achieve only 55.7% accuracy on MMIU. Through multi-faceted analytical experiments, we identify key performance gaps and limitations, providing valuable insights for future model and data improvements. We aim for MMIU to advance the frontier of LVLM research and development, moving us toward achieving sophisticated multimodal multi-image user interactions.
15
 
16
+ ## Data Viewer Configuration
17
+
18
+ The Data Viewer is configured to display the contents of `all.parquet`.
19
+
20
+ ```python
21
+ {
22
+ "task": "table-qa",
23
+ "path": "all.parquet"
24
+ }
25
+
26
  ## Data Structure
27
  ### Data Fields
28
 
 
38
  * `output`: The correct option for the question
39
 
40
 
41
+
42
  ### Example
43
  ```
44
  {