Agentic AI / Generative AI

NVIDIA Kaggle Grandmasters Win Artificial General Intelligence Competition

The ARC-AGI-2 leaderboard measures how well AI systems perform abstract reasoning.

NVIDIA researchers on Friday won a key Kaggle competition many in the field treat as a real-time pulse check on humanity’s progress toward artificial general intelligence (AGI). 

Ivan Sorokin and Jean-Francois Puget, two members of the Kaggle Grandmasters of NVIDIA (KGMoN), came in first on the Kaggle ARC Prize 2025 public leaderboard with a 27.64% score by building a solution evaluated on the same dataset behind the ARC-AGI-2 benchmark

After the competition, the team, which called itself NVARC, elevated its benchmark performance to 29.72%, all while delivering strong cost efficiency: just 20 cents per task. Its fine-tuned 4B model variant outperformed far larger, more expensive models on the same benchmark, showcasing not only state-of-the-art results, but a breakthrough in scalable, economical AGI-style reasoning.

The ARC-AGI benchmark measures how well AI systems perform abstract reasoning and then generalize from very few examples using grid-based visual puzzles. ARC-AGI-2 is a harder, updated version that removes overlap with public training data. It is explicitly designed to resist shortcuts and brute-force memorization, making it a sharper test of true systematic abstraction.

The ARC-AGI benchmark has become one of the most closely watched indicators of real progress toward general reasoning in AI. Unlike typical machine learning benchmarks, ARC-AGI tasks can’t be solved through scale, memorization, or pattern scraping. Each puzzle is a tiny grid with only a handful of examples, forcing systems to infer abstract rules—and apply them to a brand-new test case. Scores on the more-difficult ARC-AGI-2 are widely viewed as a proxy for how well an AI system can learn from almost nothing.

That’s why the Kaggle ARC Prize 2025 leaderboard matters: It’s the most open, reproducible arena where researchers test AGI-style reasoning under strict compute and time limits.

The winning NVIDIA NVARC solution wasn’t powered by giant models or brute-force search. Instead, it leaned on three ideas any developer can appreciate: synthetic data, test-time training, and disciplined engineering.

Heavyweight LLM reasoning methods—chain-of-thought, tool use, even RL-style agents—couldn’t fit within Kaggle’s tight runtime. So NVARC flipped the strategy: Move all complex reasoning offline into a synthetic data pipeline, and train smaller models capable of running fast during evaluation.

Using staged puzzle generation, concept decomposition, and progressively stronger open-weight models, the team built a diverse synthetic corpus of ARC-style tasks. Final models only needed to recognize and adapt patterns, rather than execute full program-search logic. Test-time training learns each puzzle’s specifics from its tiny example set—a technique that has become essential for leading ARC-AGI performance.

The result was a compact, cost-efficient ensemble that outperformed much larger systems and set a new bar on ARC-AGI-2, showing how synthetic data and adaptive learning can push reasoning forward.

​To successfully build out these winning solutions, the team leveraged the NVIDIA NeMo suite of tools, including NeMo RL for scalable reinforcement learning, and NeMo Skills for streamlining SDG pipelines.

Learn more about the technical details from NVARC’s writeup on Kaggle and watch this interview with ARC.

Discuss (0)

Tags