MathNet: a Global Multimodal Benchmark for Mathematical Reasoning and Retrieval

Researchers released MathNet, a 30,676-problem multimodal benchmark spanning 47 countries and 17 languages to evaluate how well LLMs and embedding systems handle Olympiad-level math reasoning and retrieval. The dataset covers two decades of competition problems and includes expert-curated pairs for testing mathematical equivalence detection.
Modelwire context
ExplainerThe retrieval dimension is the part worth pausing on: MathNet doesn't just test whether a model can solve problems, it tests whether embedding systems can detect mathematical equivalence across different notations and languages, which is a distinct and underexplored failure mode for RAG pipelines used in educational or scientific tools.
The benchmark wave we've been tracking since mid-April continues here. CoopEval (covered April 16) tested LLM agents on behavioral tasks; QuantCode-Bench tested code generation under domain constraints; MADE tested classification under label uncertainty. What those share with MathNet is a push toward benchmarks that stress-test a specific, narrow capability rather than general performance, which is a methodological correction to the broad leaderboard culture. The related coverage doesn't surface a direct competitor to MathNet specifically, but the pattern is consistent: researchers are building evaluation infrastructure faster than models are being evaluated on it.
Watch whether any of the major embedding model providers (Cohere, OpenAI, Voyage) publish retrieval scores on MathNet's equivalence-detection split within the next two quarters. If they don't, that silence is itself informative about how seriously the field takes math-specific retrieval as a benchmark axis.
Coverage we drew on
This analysis is generated by Modelwire’s editorial layer from our archive and the summary above. It is not a substitute for the original reporting. How we write it.
MentionsMathNet
Modelwire summarizes — we don’t republish. The full article lives on arxiv.org. If you’re a publisher and want a different summarization policy for your work, see our takedown page.