HMMT Feb 26
HMMT February 2026 is a math competition benchmark based on problems from the Harvard-MIT Mathematics Tournament, testing advanced mathematical problem-solving and reasoning.
Progress Over Time
Interactive timeline showing model performance evolution on HMMT Feb 26
State-of-the-art frontier
Open
Proprietary
HMMT Feb 26 Leaderboard
7 models
| Context | Cost | License | ||||
|---|---|---|---|---|---|---|
| 1 | DeepSeek | 1.6T | 1.0M | $1.74 / $3.48 | ||
| 2 | DeepSeek | 284B | 1.0M | $0.14 / $0.28 | ||
| 3 | Moonshot AI | 1.0T | 262K | $0.95 / $4.00 | ||
| 4 | Alibaba Cloud / Qwen Team | — | 1.0M | $0.50 / $3.00 | ||
| 5 | Alibaba Cloud / Qwen Team | 28B | 262K | $0.60 / $3.60 | ||
| 6 | Alibaba Cloud / Qwen Team | 35B | — | — | ||
| 7 | Zhipu AI | 754B | 200K | $1.40 / $4.40 |
Notice missing or incorrect data?
FAQ
Common questions about HMMT Feb 26
HMMT February 2026 is a math competition benchmark based on problems from the Harvard-MIT Mathematics Tournament, testing advanced mathematical problem-solving and reasoning.
The HMMT Feb 26 leaderboard ranks 7 AI models based on their performance on this benchmark. Currently, DeepSeek-V4-Pro-Max by DeepSeek leads with a score of 0.952. The average score across all models is 0.887.
The highest HMMT Feb 26 score is 0.952, achieved by DeepSeek-V4-Pro-Max from DeepSeek.
7 models have been evaluated on the HMMT Feb 26 benchmark, with 0 verified results and 7 self-reported results.
HMMT Feb 26 is categorized under math and reasoning. The benchmark evaluates text models.