ATLAS: Autoformalizing Theorems through Lifting, Augmentation, and Synthesis of Data

📅 2025-02-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Automatic formalization is hindered by the scarcity of parallel corpora pairing natural-language mathematics with formal proofs. Method: This paper proposes ATLAS, an iterative data generation framework introducing the novel “lift–enhance–synthesize” paradigm—comprising large-model iterative self-distillation, formal-grammar-guided augmentation, cross-granularity theorem alignment, and multi-round feedback-based quality control. Contribution/Results: ATLAS constructs the first high-quality, undergraduate-level parallel corpus of 300K theorem statements. Fine-tuning on this data yields the ATLAS theorem translator, achieving 80.59% pass@8 on ProofNet—a 56.6-percentage-point improvement over prior baselines—and establishing new state-of-the-art results on both miniF2F and the newly introduced graduate-level MathQual benchmark. This work advances both the data foundation and modeling capabilities for automatic formalization.

Technology Category

Application Category

📝 Abstract
Autoformalization, the process of automatically translating natural language mathematics into machine-verifiable formal language, has demonstrated advancements with the progress of large language models (LLMs). However, a key obstacle to further advancements is the scarcity of paired datasets that align natural language with formal language. To address this challenge, we introduce ATLAS (Autoformalizing Theorems through Lifting, Augmentation, and Synthesis of Data), an iterative data generation framework designed to produce large-scale, high-quality parallel theorem statements. With the proposed ATLAS running for 10 iterations, we construct an undergraduate-level dataset comprising 300k theorem statements and develop the ATLAS translator, achieving accuracies of 80.59% (pass@8) and 92.99% (pass@128) on ProofNet, significantly outperforming the base model (23.99% and 47.17%) and InternLM2-Math-Plus-7B (50.94% and 80.32%). Furthermore, the ATLAS translator also achieves state-of-the-art performance on both the high-school-level miniF2F dataset and the graduate-level MathQual dataset introduced in this work. The datasets, model, and code will be released to the public soon.
Problem

Research questions and friction points this paper is trying to address.

Autoformalization of natural language mathematics
Scarcity of paired natural-formal language datasets
Development of iterative data generation framework
Innovation

Methods, ideas, or system contributions that make the work stand out.

Iterative data generation framework
High-quality parallel theorem statements
State-of-the-art autoformalization translator
🔎 Similar Papers
No similar papers found.
X
Xiaoyang Liu
School of Mathematical Sciences, Shanghai Jiao Tong University
K
Kangjie Bao
School of Mathematical Sciences, Shanghai Jiao Tong University
Jiashuo Zhang
Jiashuo Zhang
Peking University
Software EngeneeringLLM4SESmart Contract
Yunqi Liu
Yunqi Liu
School of Mathematical Sciences, Shanghai Jiao Tong University
Y
Yu Chen
School of Mathematical Sciences, Shanghai Jiao Tong University
Y
Yuntian Liu
School of Mathematical Sciences, Shanghai Jiao Tong University
Y
Yang Jiao
SPEIT, Shanghai Jiao Tong University; JoinTech Co., Ltd
T
Tao Luo
School of Mathematical Sciences, Shanghai Jiao Tong University; Institute of Natural Sciences, MOE-LSC, Shanghai Jiao Tong University; CMA-Shanghai, Shanghai Artificial Intelligence Laboratory