
Background
ARC‑AGI was introduced in 2019 as a grid‑based reasoning benchmark (“v1”) designed to test whether AI systems can infer novel rules from a few examples rather than rely on pattern memorization. Open‑source solvers plateaued near 53 % accuracy, while a high‑compute run of OpenAI’s o3‑preview model achieved roughly 75–88 %, indicating that v1 was largely saturated.
To raise the bar, the ARC Prize Foundation unveiled the harder, human‑validated “ARC‑AGI‑2” (v2) on 24 March 2025 and opened a Kaggle contest capped at about US $0.42 of compute per task. The headline rule remains: the first fully open‑source system to reach ≥ 85 % on the private v2 set wins the $1 million Grand Prize.
Resolution Criteria
The market resolves YES if before January 1, 2026 the ARC Prize Foundation publicly announces and awards any portion of the $1 million Grand Prize to one or more teams.
Primary rule: The winning submission must achieve ≥ 85 % accuracy on ARC‑AGI‑2 (or an officially designated successor) during an official competition period.
Future changes: If ARC publishes a new test or alters the accuracy threshold, the operative condition remains “the first public, binding commitment to pay out—or the actual payout of—the prize labelled the ARC Grand Prize.”
This is super interesting:
https://iliao2345.github.io/blog_posts/arc_agi_without_pretraining/arc_agi_without_pretraining.html
@MalachiteEagle Does rhat mean you only get 1 million $ by passing 85 on v2? If so that’s crap!
@AndrewMcKnight it is not an accident. I thought it would be better if people predicted it a year ahead of time.