MMMU
multimodal
Massive Multi-discipline Multimodal Understanding (MMMU) evaluates multimodal understanding across 30 disciplines.
Published: 2023
Scale: 0-100
Top Score: 82.9
MMMU Leaderboard
Rank | Model | Provider | Score | Parameters | Released | Type |
---|---|---|---|---|---|---|
1 | o3 | OpenAI | 82.9 | 2025-04-16 | Multimodal | |
2 | o4-mini | OpenAI | 81.6 | 2025-04-16 | Multimodal | |
3 | Gemini 2.5 Flash | 79.7 | 2025-05-20 | Multimodal | ||
4 | Gemini 2.5 Pro | 79.6 | 2025-05-06 | Multimodal | ||
5 | o1 | OpenAI | 78.2 | 2024-09-12 | Multimodal | |
6 | Claude 3.7 Sonnet | Anthropic | 75 | 2025-02-24 | Multimodal | |
7 | GPT-4.5 | OpenAI | 74.4 | 2025-02-27 | Multimodal | |
8 | Claude Opus 4 | Anthropic | 73.7 | 2025-05-22 | Multimodal | |
9 | Gemini 2.0 Pro | 72.7 | 2025-02-05 | Multimodal | ||
10 | Claude Sonnet 4 | Anthropic | 72.6 | 2025-05-22 | Multimodal |
About MMMU
Description
Massive Multi-discipline Multimodal Understanding (MMMU) evaluates multimodal understanding across 30 disciplines.
Methodology
MMMU evaluates models on a scale of 0 to 100. Higher scores indicate better performance. For detailed information about the methodology, please refer to the original paper.
Publication
This benchmark was published in 2023.Read the full paper