Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
datasets:
|
4 |
+
- kfdong/STP_Lean_SFT
|
5 |
+
base_model:
|
6 |
+
- deepseek-ai/DeepSeek-Prover-V1.5-SFT
|
7 |
+
---
|
8 |
+
|
9 |
+
This is the final Self-play Theorem Prover model as described in the paper [https://arxiv.org/abs/2502.00212](https://arxiv.org/abs/2502.00212). The training and evalution code is avaliable [here](https://github.com/kfdong/STP/tree/main).
|
10 |
+
|
11 |
+
|
12 |
+
```tex
|
13 |
+
@article{dong2025beyond,
|
14 |
+
title={Beyond Limited Data: Self-play LLM Theorem Provers with Iterative Conjecturing and Proving},
|
15 |
+
author={Dong, Kefan and Ma, Tengyu},
|
16 |
+
journal={arXiv preprint arXiv:2502.00212},
|
17 |
+
year={2025}
|
18 |
+
}
|
19 |
+
```
|
20 |
+
|
21 |
+
## 1. Evaluation Results
|
22 |
+
|
23 |
+
The table below compares the pass@3200 performance of STP (our model) and DeepSeek-Prover-V1.5 on miniF2F-test and ProofNet-test.
|
24 |
+
|
25 |
+
<div align="center">
|
26 |
+
|
27 |
+
| | miniF2F-test | ProofNet-test |
|
28 |
+
|--------|------------------|------------------|
|
29 |
+
| **DeepSeek-Prover-V1.5-SFT** | 53.3% ± 0.5% | 21.0% ± 0.9% |
|
30 |
+
| **DeepSeek-Prover-V1.5-RL** | 54.9% ± 0.7% | 22.0% ± 0.5% |
|
31 |
+
| **STP** | **65.0% ± 0.5%** | **23.9% ± 0.6%** |
|
32 |
+
|
33 |
+
</div>
|
34 |
+
|
35 |
+
## 2. Dataset
|
36 |
+
|
37 |
+
We also release the dataset [here](https://huggingface.co/datasets/kfdong/STP_Lean_0320), which contains:
|
38 |
+
- Extracted examples from mathlib4,
|
39 |
+
- Generated correct proofs of statements in LeanWorkbook,
|
40 |
+
- Generated correct proofs of conjectures proposed by our model during self-play training.
|
41 |
+
|
42 |
+
Our final model is finetuned from DeepSeek-Prover-V1.5-SFT with this dataset for 1 epoch.
|