The Agentica SDK by Symbolica achieves an unverified competition score of 36.08% on the publicly available dataset of 25 ARC-AGI-3 games [1], passing 113 out of 182 playable levels and completing 7 out of 25 games [2]. This uses the same harness we previously published for solving the 3 public pre-release ARC-AGI-3 games.
Our implementation outperforms CoT baselines of 0.2% (Opus 4.6 Max) and 0.3% (GPT 5.4 High), while maintaining a far lower cost: Agentica's 36.08% for $1,005 vs. Opus 4.6's 0.25% for $8,900.
Check out the code on GitHub symbolica-ai/ARC-AGI-3-AgentsGallery - Games Won
Score Breakdown - All Games
Chat with Agentica
We've sandboxed the SDK and let it run any persistent task, including solving ARC puzzles.
References
[1] ARC Prize Foundation. ARC-AGI-3: A New Challenge for Frontier Agentic Intelligence. Arc Prize Foundation.
[2] ARC Prize. ARC-AGI-3. ARC Prize.
Appendix
A note on scoring
Human baseline scores available via the ARC-AGI-3 API state that the game cn04 has 6 levels in total. This does not match the number of levels in the corresponding game available via the API.