Benchmark snapshot

MiMo benchmark highlights

These panels summarize public benchmark signals from Xiaomi MiMo source pages. They are not a replacement for the original tables; they are a faster reading layer on top of them.

Reasoning

MiMo-V2-Flash is publicly framed as a reasoning-forward model family, with strong results on math and knowledge-heavy tasks.

BenchmarkContextValue
MMLU-ProMiMo-V2-Flash84.9
GPQA-DiamondMiMo-V2-Flash83.7
AIME 2025MiMo-V2-Flash94.1

Code and agents

Public benchmark tables emphasize coding, SWE-Bench, and terminal-style agent tasks as part of the MiMo-V2-Flash story.

BenchmarkContextValue
LiveCodeBench-v6MiMo-V2-Flash80.6
SWE-Bench VerifiedMiMo-V2-Flash73.4
Terminal-Bench 2.0MiMo-V2-Flash38.5

Earlier MiMo-7B line

The earlier MiMo release remains useful for understanding the public lineage and the reasoning-first positioning of the wider model family.

BenchmarkContextValue
MATH500MiMo-7B-RL95.8
AIME 2024MiMo-7B-RL68.2
LiveCodeBench v6MiMo-7B-RL49.3
Source: MiMo README