Skip to main content

One training example is all you need for LLM reasoning?

Created on May 30|Last edited on May 30
A recent study shows that training a large language model with reinforcement learning on just one carefully chosen math problem can yield surprising and massive improvements in performance. This unconventional approach challenges the standard practice of relying on thousands of training examples. The research focuses on Qwen2.5-Math, a 1.5 billion parameter model, and shows a dramatic improvement: accuracy on a benchmark set of 500 math problems jumped from 36.0 percent to 73.6 percent after reinforcement learning using only a single problem.

Astonishing Gains from a Single Example

The pattern repeats across six math benchmarks, with average accuracy increasing from 17.6 percent to 35.7 percent. That performance matches or surpasses what you’d expect from training on over a thousand examples. Adding just one more training example yields even stronger improvements. These effects are consistent across different models and reinforcement learning algorithms, showing that the result isn’t limited to one architecture or training setup.

Why It Might Work

The paper offers some clues and hypotheses but doesn’t fully explain why one-shot reinforcement learning unlocks such large performance gains. It presents evidence, discusses likely mechanisms, and rules out common assumptions like memorization or regularization-based “grokking,” but leaves the ultimate explanation open for future research.
One idea is the latent capability hypothesis. It suggests that the base language model already contains buried reasoning abilities, but doesn’t reliably access them. The RL signal, even from one example, helps trigger this latent ability and makes it more consistent in generating correct reasoning steps.
Immediate feedback on a single correct example somehow generalizes far beyond that problem. The model doesn’t just memorize a solution format—it recalibrates its reasoning approach and becomes better at solving many unrelated tasks, even across different domains like moving from geometry to algebra.
The researchers also make clear that the gains aren’t just from learning to format answers correctly. Real improvements in reasoning occur. For example, the model begins issuing commands like “recheck” or “recalculate,” showing more reflective behavior.
Importantly, this is not “grokking” in the traditional sense. In grokking, a model suddenly generalizes after a long period of memorization due to regularization effects. Here, the test accuracy continues to improve even after the model has fully memorized the training example—what the authors call post-saturation generalization.
The paper acknowledges limitations. It doesn’t provide a full theoretical explanation and leaves open questions about what’s happening at the level of neural circuits, attention, or internal representations. It’s a compelling empirical result that calls for deeper investigation.

Generalization Beyond the Example

The model doesn’t just learn to solve one kind of problem. Training on a single geometry example can lead to better algebra performance. It starts generalizing not only within a domain but across different types of reasoning tasks. The training even seems to enhance metacognition, making the model more likely to reflect on and verify its steps. That reflective capacity may be part of what’s being unlocked.

What This Means for LLM Training

This work raises a significant challenge to assumptions about scale and data requirements. Rather than collecting massive datasets, it might be more effective to pick the right few examples and use them in conjunction with RL. Even randomly chosen examples can offer big improvements, but handpicked ones make an even stronger impact. The ability to amplify reasoning with minimal supervision suggests a new path forward in model fine-tuning.

The Bigger Picture and Caution

These results are striking, but a single paper is never enough to overturn established practice. The effect needs to be reproduced across models, tasks, and implementations. Still, the core idea—that a model’s reasoning ability can be rapidly unlocked with minimal but smartly applied feedback—offers a fresh angle on how we train and improve language models. Whether this approach can scale or generalize to other areas remains to be seen, but it opens a compelling new direction for future research.