MMMU

multimodal

Massive Multi-discipline Multimodal Understanding (MMMU) evaluates multimodal understanding across 30 disciplines.

Published: 2023
Scale: 0-100
Top Score: 82.9

MMMU Leaderboard

RankModelProviderScoreParametersReleasedType
1o3OpenAI
82.9
2025-04-16Multimodal
2o4-miniOpenAI
81.6
2025-04-16Multimodal
3Gemini 2.5 FlashGoogle
79.7
2025-05-20Multimodal
4Gemini 2.5 ProGoogle
79.6
2025-05-06Multimodal
5o1OpenAI
78.2
2024-09-12Multimodal
6Claude 3.7 SonnetAnthropic
75
2025-02-24Multimodal
7GPT-4.5OpenAI
74.4
2025-02-27Multimodal
8Claude Opus 4Anthropic
73.7
2025-05-22Multimodal
9Gemini 2.0 ProGoogle
72.7
2025-02-05Multimodal
10Claude Sonnet 4Anthropic
72.6
2025-05-22Multimodal

About MMMU

Description

Massive Multi-discipline Multimodal Understanding (MMMU) evaluates multimodal understanding across 30 disciplines.

Methodology

MMMU evaluates models on a scale of 0 to 100. Higher scores indicate better performance. For detailed information about the methodology, please refer to the original paper.

Publication

This benchmark was published in 2023.Read the full paper

Related Benchmarks