Zulip Chat Archive

Stream: Machine Learning for Theorem Proving

Topic: Numina-Lean-Agent solves 12 problems on Putnam2025


Junqi Liu (Jan 20 2026 at 14:28):

We are thrilled to announce Numina-Lean-Agent, an open-source agentic framework that achieves state of the art performance in formal theorem proving and autoformalization.

Github: https://github.com/project-numina/numina-lean-agent

Paper: https://github.com/project-numina/numina-lean-agent/blob/main/NuminaLeanAgent.pdf

Demo:: https://demo.projectnumina.ai/

Unlike proprietary systems that rely on massive, domain-specific fine-tuning (such as Aristotle, Axiom Prover, or Seed Prover 1.5), Numina-Lean-Agent is built entirely on generic, accessible foundation models—specifically Claude, Gemini, and GPT-5.2. By engineering a superior agentic reasoning loop rather than training proprietary weights, we have unlocked a new paradigm in automated reasoning.

Key Breakthroughs

  1. Solving the Putnam 2025 Competition (12/12) We have achieved a perfect score on the formal Putnam 2025 benchmark, solving 12 out of 12 problems.
  • This matches the performance of domain-specific systems, surpassing top tier human competitors.

  • It demonstrates that general-purpose reasoning models, when guided by a rigorous agentic framework, can perform at the highest levels of mathematical competition.

  1. Research Formalization Beyond competitions, we are targeting the frontier of modern research. Using an interactive backend with Numina-Lean-Agent, we successfully formalized the recent paper "Effective Brascamp-Lieb inequalities" (arXiv:2511.11091).
  • High Autonomy: this was achieved in collaboration with one of the paper's authors, Weikun He. The whole formalization lasted two weeks, with more than 90% of the lean codes written by Numina-Lean-Agent.

  • Historic milestone: the first time a fully accessible AI system has formalized a cutting-edge, research-level mathematics paper in collaboration with a human expert.

  • Crucially, the original manuscript is to be reviewed and validated. Our formalization provides a machine-checked guarantee of correctness before the traditional review process is complete. This implies a transformation in mathematical publishing: a future where research papers are validated by autoformalization with minimal human effort, shifting the verification bottleneck from months of human review to some amounts of compute.

Community & Vision

We are deeply grateful for the foundational contributions of the entire Mathlib community. Without the rigorous groundwork and high-quality data curated by human maintainers, generic models such as Claude or Gemini would never be able to reach the level of formal reasoning we see today.

Building on this foundation, our goal is to close the loop. We believe AI can now contribute meaningfully to the progress of mathematical research via formalization—not by replacing mathematicians, but by empowering them. Ultimately, the future of mathematical research lies not in black-box solvers, but in transparent, collaborative tools that amplify human insight.

Join the Frontier

At Numina, we are committed to being the transformation engine of AI for mathematics. We are currently scaling our efforts to explore the absolute limits of AI in formalizing cutting-edge research.

  • We are hiring: We are hiring mathlib experts who are interested in open source AI to join us to build a research level autoformalization system. 

  • Open Science: All our autoformalization results are open-sourced and will be refined for integration into Mathlib.

If you are ready to help build the future of formal mathematics, contact us at contact@projectnumina.ai.

Numina Team

Patrick Massot (Jan 20 2026 at 17:38):

What does open source mean in your first sentence if you rely on Claude, Gemini or GPT?

Kelly Davis (Jan 20 2026 at 17:46):

Patrick Massot said:

What does open source mean in your first sentence if you rely on Claude, Gemini or GPT?

I'm not an author of numina-lean-agent, but normally it means the repo's code and only the repo's code is under an open source license, in this case the MIT License.

Eric Wieser (Jan 20 2026 at 19:03):

It's great to see this being achieved with open code (even if not with open models)!

Eric Wieser (Jan 20 2026 at 19:05):

I just tried out the demo on a toy problem that I have been planning to contribute to mathlib for a while, and it seems rather eager to give up and use an axiom:

Final proof

(You can import the full trace from kimina-export.json)

Eric Wieser (Jan 20 2026 at 19:06):

(I'll admit it's highly likely that this theorem is stated incorrectly)

Joseph Myers (Jan 20 2026 at 20:50):

Eric Wieser said:

(I'll admit it's highly likely that this theorem is stated incorrectly)

Yes. As previously discussed, the Hausdorff measure normalization we have disagrees with Lebesgue measure in Euclidean spaces (by a factor of 2n2^n divided by the volume of the unit nn-ball, or something like that; in 1-dimensional space these factors cancel each other out so the measures do agree there). I think we should define Euclidean-scaled Hausdorff measure in mathlib. The simple way of doing so involves dividing by the Hausdorff measure of a unit cuboid in Euclidean space, and all you need to know to do useful things with it is that said measure of the cuboid is not zero or infinite. If you want a definition that works for fractional-dimensional Hausdorff measure, however, and thus want to divide by a more explicit expression involving the gamma function (to extend the volume of the unit ball to fractional dimensions), then you need to prove the isodiametric inequality to get the explicit scaling factor involving the volume of the unit ball.

The fractional-dimensional version should not be needed for any kind of elementary geometry, which might involve talking about lengths of curves or surface areas of surfaces that aren't flat, but only for non-fractal curves and surfaces.

Joseph Myers (Jan 20 2026 at 21:00):

Regarding Numina-Lean-Agent: section 3.1 talks about different budgets for different problems, and section 3.2 says "For problem A5, we adopt a novel subagent mechanism [...]". Were these choices made entirely autonomously by the AI (given only the formal and informal problem statements with associated labels A1, ..., B6), or were they made by humans before presenting the problems to the AI, or were they made by humans based on an iterative process seeing what the AI did on the problems before being given such parameters of budget or subagent mechanism? There's a big different between possibilities such as:

  1. The AI and its parameters were all frozen before the Putnam problems were given to human competitors, and all subsequent choices about budget, subagent mechanism, etc. were made entirely autonomously by that AI version that predated the Putnam 2025 exams.
  2. Humans chose those parameters after seeing the Putnam problems, but the AI was run just the once with the first choice of parameters given by the humans and solved the problems when run.
  3. There were multiple AI attempts with different parameters and the ones described in sections 3.1 and 3.2 resulted from an iterative process based on examining the results of prior attempts.

Option 1 in that list would be a genuine AI solution to the problem, whereas option 3 seems much more like a human/AI collaboration (where the humans might or might not also have been working using prior knowledge of informal solutions) to solve the problems.

David Renshaw (Jan 20 2026 at 22:06):

Prompted by this announcement, I've been playing with lean-lsp-mcp today, and it's very fun! Here are some nice golfs it found: 1, 2, 3, 4. And here it filled a sorry: 5

Junqi Liu (Jan 21 2026 at 06:17):

Joseph Myers said:

Regarding Numina-Lean-Agent: section 3.1 talks about different budgets for different problems, and section 3.2 says "For problem A5, we adopt a novel subagent mechanism [...]". Were these choices made entirely autonomously by the AI (given only the formal and informal problem statements with associated labels A1, ..., B6), or were they made by humans before presenting the problems to the AI, or were they made by humans based on an iterative process seeing what the AI did on the problems before being given such parameters of budget or subagent mechanism? There's a big different between possibilities such as:

  1. The AI and its parameters were all frozen before the Putnam problems were given to human competitors, and all subsequent choices about budget, subagent mechanism, etc. were made entirely autonomously by that AI version that predated the Putnam 2025 exams.
  2. Humans chose those parameters after seeing the Putnam problems, but the AI was run just the once with the first choice of parameters given by the humans and solved the problems when run.
  3. There were multiple AI attempts with different parameters and the ones described in sections 3.1 and 3.2 resulted from an iterative process based on examining the results of prior attempts.

Option 1 in that list would be a genuine AI solution to the problem, whereas option 3 seems much more like a human/AI collaboration (where the humans might or might not also have been working using prior knowledge of informal solutions) to solve the problems.

For the budgets, we calculated the cost budgets after we finished the proof. For the subagent mechanism, we were not involved in the proof process; the agent decided on its own to separate the subgoal and create a new lean file. The prompts are in our repo.

Junqi Liu (Jan 21 2026 at 06:18):

David Renshaw said:

Prompted by this announcement, I've been playing with lean-lsp-mcp today, and it's very fun! Here are some nice golfs it found: 1, 2, 3, 4. And here it filled a sorry: 5

The golf feature of Numina-Lean-Agent will come soon!

Bas Spitters (Jan 30 2026 at 20:25):

@Junqi Liu Looks exciting. The installation instructions https://github.com/project-numina/numina-lean-agent/ are not great.
I understand one needs a Gemini API key, but there's no instructions on how to use it. Also, is it really necessary to use two models ?

Junqi Liu (Feb 03 2026 at 06:58):

Bas Spitters said:

Junqi Liu Looks exciting. The installation instructions https://github.com/project-numina/numina-lean-agent/ are not great.
I understand one needs a Gemini API key, but there's no instructions on how to use it. Also, is it really necessary to use two models ?

Set as environment variable

TongKe Xue (Feb 05 2026 at 04:15):

As far as Putnam/IMO is concerned, is Numina on the same level as DeepMind/OpenAI/Harmonic/AxiomMath ?


Last updated: Feb 28 2026 at 14:05 UTC