Zulip Chat Archive
Stream: Machine Learning for Theorem Proving
Topic: ABEL
Fabian Glöckle (Oct 14 2024 at 16:19):
@Jannis Limperg and I made a REPL based on Aesop that doesn't take "states = current goal stack" but allows to disentangle independent goals, allowing for methods like Hypertree proof search (https://arxiv.org/abs/2205.11491).
We are planning to open source the code soon. Paper: https://openreview.net/pdf?id=kk3mSjVCUO
Mert Ünsal (Oct 15 2024 at 00:41):
Hey Fabian - I saw your paper, very interesting results! I saw the Github repository but wasn’t able to easily hack it into something I can use.
Please let us know when you release this!
Fabian Glöckle (Oct 15 2024 at 06:53):
Which GitHub?
Fabian Glöckle (Oct 15 2024 at 06:54):
Aesoprepl is not public yet (on it!), you'll need 200 lines on top of Aesop.
Mert Ünsal (Oct 15 2024 at 09:23):
I was talking about github repo of ABEL but not even sure if it has the code for this - apparently not.
Let us know when we can use Aesoprepl :)
Fabian Glöckle (Oct 15 2024 at 09:29):
oh no, i just realized the name conflict to https://github.com/GAIR-NLP/abel -- it's unrelated
Mert Ünsal (Oct 15 2024 at 12:39):
I was actually refering to this: https://github.com/facebookresearch/Evariste which is the Github repository for your paper, right?
Fabian Glöckle (Oct 16 2024 at 11:18):
no it's the GitHub repo for facebook's 2022 paper https://arxiv.org/abs/2205.11491, but it's no longer used/maintained
Notification Bot (Oct 16 2024 at 11:28):
8 messages were moved here from #Machine Learning for Theorem Proving > Interacting with Lean 4 by Jason Rute.
Fabian Glöckle (Oct 16 2024 at 12:05):
To give a bit more context: this recreates HTPS (https://arxiv.org/abs/2205.11491) with a modern stack: Lean4, twice larger mathlib, Aesop, larger language models, faster language model inference code.
We view reinforcement learning as an inference technique: throw in all your theorems, let it run for some hours, get back as many proofs as it found. (Others would call this "life-long learning". :p)
The results exceed the reinforcement learnning baselines from before 2022 with a fraction of the compute. We're slightly below DeepSeekProver and InternLM2-StepProver without adding any additional data and likely also with a fraction of their compute (they don't release theirs).
Huajian Xin (Oct 16 2024 at 13:17):
Excellent work! I'm really looking forward to the open-sourcing of AesopRepl and the online RL algorithms!
Albert Jiang (Oct 16 2024 at 20:51):
Fabulous
Albert Jiang (Oct 16 2024 at 20:51):
Fabulous
Charlie Meyer (Apr 14 2025 at 22:15):
Any updates on open-sourcing AesopREPL?
Last updated: May 02 2025 at 03:31 UTC