Zulip Chat Archive

Stream: Machine Learning for Theorem Proving

Topic: ABEL


Fabian Glöckle (Oct 14 2024 at 16:19):

@Jannis Limperg and I made a REPL based on Aesop that doesn't take "states = current goal stack" but allows to disentangle independent goals, allowing for methods like Hypertree proof search (https://arxiv.org/abs/2205.11491).

We are planning to open source the code soon. Paper: https://openreview.net/pdf?id=kk3mSjVCUO

Mert Ünsal (Oct 15 2024 at 00:41):

Hey Fabian - I saw your paper, very interesting results! I saw the Github repository but wasn’t able to easily hack it into something I can use.

Please let us know when you release this!

Fabian Glöckle (Oct 15 2024 at 06:53):

Which GitHub?

Fabian Glöckle (Oct 15 2024 at 06:54):

Aesoprepl is not public yet (on it!), you'll need 200 lines on top of Aesop.

Mert Ünsal (Oct 15 2024 at 09:23):

I was talking about github repo of ABEL but not even sure if it has the code for this - apparently not.

Let us know when we can use Aesoprepl :)

Fabian Glöckle (Oct 15 2024 at 09:29):

oh no, i just realized the name conflict to https://github.com/GAIR-NLP/abel -- it's unrelated

Mert Ünsal (Oct 15 2024 at 12:39):

I was actually refering to this: https://github.com/facebookresearch/Evariste which is the Github repository for your paper, right?

Fabian Glöckle (Oct 16 2024 at 11:18):

no it's the GitHub repo for facebook's 2022 paper https://arxiv.org/abs/2205.11491, but it's no longer used/maintained

Notification Bot (Oct 16 2024 at 11:28):

8 messages were moved here from #Machine Learning for Theorem Proving > Interacting with Lean 4 by Jason Rute.

Fabian Glöckle (Oct 16 2024 at 12:05):

To give a bit more context: this recreates HTPS (https://arxiv.org/abs/2205.11491) with a modern stack: Lean4, twice larger mathlib, Aesop, larger language models, faster language model inference code.

We view reinforcement learning as an inference technique: throw in all your theorems, let it run for some hours, get back as many proofs as it found. (Others would call this "life-long learning". :p)

The results exceed the reinforcement learnning baselines from before 2022 with a fraction of the compute. We're slightly below DeepSeekProver and InternLM2-StepProver without adding any additional data and likely also with a fraction of their compute (they don't release theirs).

Huajian Xin (Oct 16 2024 at 13:17):

Excellent work! I'm really looking forward to the open-sourcing of AesopRepl and the online RL algorithms!

Albert Jiang (Oct 16 2024 at 20:51):

Fabulous

Albert Jiang (Oct 16 2024 at 20:51):

Fabulous

Charlie Meyer (Apr 14 2025 at 22:15):

Any updates on open-sourcing AesopREPL?


Last updated: May 02 2025 at 03:31 UTC