In what year will AI achieve a score of 85% or higher on the SimpleBench benchmark?
19
1.4kṀ4220
2033
March 26, 2032
25%
2025
43%
2026
23%
2027
7%
2028
7%
2029
7%
2030
8%
2031
7%
2032

Background:

SimpleBench is a 200‑item multiple‑choice test designed to probe everyday human reasoning that still eludes frontier LLMs, including spatio‑temporal reasoning, social intelligence and linguistic “trick” questions. Unlike most other benchmarks, humans still outperform AI on the SimpleBench.

State of play:

• Human reference accuracy: 83.7 %

• 2025 AI accuracy (Gemini 2.5 Pro): 62.4 %

Why this milestone matters

  • Everyday reasoning: Passing SimpleBench would indicate that LLMs can handle commonsense scenarios that remain brittle today.

  • Benchmark head‑room: Unlike MMLU, SimpleBench has not yet been “solved”, so it is a useful yard‑stick for progress to compare AI to humans.

Resolution Criteria:

This market resolves to the year in which a fully automated AI system first achieves ≥ 85% average accuracy on SimpleBench (ALL metric), subject to all of the following:

  • Verification – The claim must be confirmed by either

    1. a peer‑reviewed paper on arXiv

    2. a public leaderboard entry on SimpleBench Official Website or another credible source.

  • Compute resources – Unlimited.

Fine Print:

If the resolution criteria are unsatisfied by Jan 1, 2033 the market resolves to “Not Applicable.”

Get
Ṁ1,000
to start trading!
© Manifold Markets, Inc.TermsPrivacy