Search is not available for this dataset
bytes
stringlengths
45.8k
18.4M
path
stringclasses
1 value
index
int64
41
23.6k
question
stringlengths
11
235
multi-choice options
sequencelengths
5
6
answer
stringclasses
5 values
category
stringclasses
9 values
l2-category
stringclasses
37 values
"/9j/4AAQSkZJRgABAQAAAQABAAD/2wBDAAgGBgcGBQgHBwcJCQgKDBQNDAsLDBkSEw8UHRofHh0aHBwgJC4nICIsIxwcKDcpLDA(...TRUNCATED)
dummy/path
23,204
Where is the yellow crane in the picture?
["(A) In the upper right area of the picture","(B) In the upper left area of the picture","(C) In th(...TRUNCATED)
A
Perception/Remote Sensing
position
"/9j/4AAQSkZJRgABAQAAAQABAAD/2wBDAAgGBgcGBQgHBwcJCQgKDBQNDAsLDBkSEw8UHRofHh0aHBwgJC4nICIsIxwcKDcpLDA(...TRUNCATED)
dummy/path
22,882
Where is the tennis court located in this picture?
["(A) In the middle right area of this picture","(B) In the top left corner of this picture","(C) In(...TRUNCATED)
B
Perception/Remote Sensing
position
"/9j/4AAQSkZJRgABAQAAAQABAAD/2wBDAAgGBgcGBQgHBwcJCQgKDBQNDAsLDBkSEw8UHRofHh0aHBwgJC4nICIsIxwcKDcpLDA(...TRUNCATED)
dummy/path
22,819
Where is the white tower in the picture?
["(A) To the left above the picture, amidst the green trees.","(B) To the left corner of the picture(...TRUNCATED)
D
Perception/Remote Sensing
position
"/9j/4AAQSkZJRgABAQAAAQABAAD/2wBDAAgGBgcGBQgHBwcJCQgKDBQNDAsLDBkSEw8UHRofHh0aHBwgJC4nICIsIxwcKDcpLDA(...TRUNCATED)
dummy/path
22,837
Where is the gray billboard with gorillas in the picture?
["(A) In the upper left area of the picture","(B) In the upper right area of the picture","(C) In th(...TRUNCATED)
D
Perception/Remote Sensing
position
"/9j/4AAQSkZJRgABAQAAAQABAAD/2wBDAAgGBgcGBQgHBwcJCQgKDBQNDAsLDBkSEw8UHRofHh0aHBwgJC4nICIsIxwcKDcpLDA(...TRUNCATED)
dummy/path
23,426
Where is the structure of blue, yellow, and green alternating colors in the picture?
["(A) In the lower right corner of the picture.","(B) At the bottom left of the picture","(C) Inside(...TRUNCATED)
C
Perception/Remote Sensing
position
"/9j/4AAQSkZJRgABAQAAAQABAAD/2wBDAAgGBgcGBQgHBwcJCQgKDBQNDAsLDBkSEw8UHRofHh0aHBwgJC4nICIsIxwcKDcpLDA(...TRUNCATED)
dummy/path
23,199
Where are the red and white cell towers in the picture?
["(A) In the upper left area of the picture","(B) In the upper right area of the picture","(C) In th(...TRUNCATED)
C
Perception/Remote Sensing
position
"/9j/4AAQSkZJRgABAQAAAQABAAD/2wBDAAgGBgcGBQgHBwcJCQgKDBQNDAsLDBkSEw8UHRofHh0aHBwgJC4nICIsIxwcKDcpLDA(...TRUNCATED)
dummy/path
22,759
Where is the largest vacant lot in the picture?
["(A) In the upper left area of the picture.","(B) In the upper right area of the picture.","(C) In (...TRUNCATED)
C
Perception/Remote Sensing
position
"/9j/4AAQSkZJRgABAQAAAQABAAD/2wBDAAgGBgcGBQgHBwcJCQgKDBQNDAsLDBkSEw8UHRofHh0aHBwgJC4nICIsIxwcKDcpLDA(...TRUNCATED)
dummy/path
23,356
Where is the circular mini amusement park with amusement facilities located in the picture?
["(A) In the left area of the picture","(B) In the right area of the picture","(C) In the low area o(...TRUNCATED)
D
Perception/Remote Sensing
position
"/9j/4AAQSkZJRgABAQAAAQABAAD/2wBDAAgGBgcGBQgHBwcJCQgKDBQNDAsLDBkSEw8UHRofHh0aHBwgJC4nICIsIxwcKDcpLDA(...TRUNCATED)
dummy/path
23,575
Where is the white umbrella object in the picture?
["(A) In the lower right area of the picture","(B) In the right left area of the picture","(C) In th(...TRUNCATED)
B
Perception/Remote Sensing
position
"/9j/4AAQSkZJRgABAQAAAQABAAD/2wBDAAgGBgcGBQgHBwcJCQgKDBQNDAsLDBkSEw8UHRofHh0aHBwgJC4nICIsIxwcKDcpLDA(...TRUNCATED)
dummy/path
22,761
Where the river is in the picture?
["(A) In the right area of the picture.","(B) In the upper right area of the picture.","(C) In the u(...TRUNCATED)
D
Perception/Remote Sensing
position

  • 2024.11.14 🌟 MME-RealWorld now has a lite version (50 samples per task, or all if fewer than 50) for inference acceleration, which is also supported by VLMEvalKit and Lmms-eval.
  • 2024.09.03 🌟 MME-RealWorld is now supported in the VLMEvalKit and Lmms-eval repository, enabling one-click evaluation—give it a try!"
  • 2024.08.20 🌟 We are very proud to launch MME-RealWorld, which contains 13K high-quality images, annotated by 32 volunteers, resulting in 29K question-answer pairs that cover 43 subtasks across 5 real-world scenarios. As far as we know, MME-RealWorld is the largest manually annotated benchmark to date, featuring the highest resolution and a targeted focus on real-world applications.

Paper: arxiv.org/abs/2408.13257

Code: https://github.com/yfzhang114/MME-RealWorld

Project page: https://mme-realworld.github.io/

Results of representative models on the MME-RealWorld-lite subset:

image/png

MME-RealWorld Data Card

Dataset details

Existing Multimodal Large Language Model benchmarks present several common barriers that make it difficult to measure the significant challenges that models face in the real world, including:

  1. small data scale leads to a large performance variance;
  2. reliance on model-based annotations results in restricted data quality;
  3. insufficient task difficulty, especially caused by the limited image resolution.

We present MME-RealWord, a benchmark meticulously designed to address real-world applications with practical relevance. Featuring 13,366 high-resolution images averaging 2,000 × 1,500 pixels, MME-RealWord poses substantial recognition challenges. Our dataset encompasses 29,429 annotations across 43 tasks, all expertly curated by a team of 25 crowdsource workers and 7 MLLM experts. The main advantages of MME-RealWorld compared to existing MLLM benchmarks as follows:

  1. Data Scale: with the efforts of a total of 32 volunteers, we have manually annotated 29,429 QA pairs focused on real-world scenarios, making this the largest fully human-annotated benchmark known to date.

  2. Data Quality: 1) Resolution: Many image details, such as a scoreboard in a sports event, carry critical information. These details can only be properly interpreted with high- resolution images, which are essential for providing meaningful assistance to humans. To the best of our knowledge, MME-RealWorld features the highest average image resolution among existing competitors. 2) Annotation: All annotations are manually completed, with a professional team cross-checking the results to ensure data quality.

  3. Task Difficulty and Real-World Utility.: We can see that even the most advanced models have not surpassed 60% accuracy. Additionally, many real-world tasks are significantly more difficult than those in traditional benchmarks. For example, in video monitoring, a model needs to count the presence of 133 vehicles, or in remote sensing, it must identify and count small objects on a map with an average resolution exceeding 5000×5000.

  4. MME-RealWord-CN.: Existing Chinese benchmark is usually translated from its English version. This has two limitations: 1) Question-image mismatch. The image may relate to an English scenario, which is not intuitively connected to a Chinese question. 2) Translation mismatch [58]. The machine translation is not always precise and perfect enough. We collect additional images that focus on Chinese scenarios, asking Chinese volunteers for annotation. This results in 5,917 QA pairs.

image/png

Downloads last month
371

Collection including yifanzhang114/MME-RealWorld-lite-lmms-eval