File Download
Supplementary
-
Citations:
- Appears in Collections:
postgraduate thesis: Mathematical reasoning in large language models : bridging natural language and formal mathematics
| Title | Mathematical reasoning in large language models : bridging natural language and formal mathematics |
|---|---|
| Authors | |
| Advisors | Advisor(s):Huang, Z |
| Issue Date | 2025 |
| Publisher | The University of Hong Kong (Pokfulam, Hong Kong) |
| Citation | Lu, J. [陸劍橋]. (2025). Mathematical reasoning in large language models : bridging natural language and formal mathematics. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR. |
| Abstract | Large Language Models (LLMs) increasingly confront the challenge of processing and reasoning about mathematical content in both natural language and formal systems. This complexity arises from the dual nature of mathematical reasoning: understanding informal mathematical text and producing verifiable formal proofs. The challenge is further amplified by the need to bridge these two domains effectively while maintaining mathematical rigor.
In this thesis, we investigate mathematical reasoning in LLMs with a focus on three interconnected areas: natural language mathematical reasoning, autoformalization, and automated theorem proving. In the natural language reasoning domain, we introduce SELF, a novel framework enabling autonomous improvement of LLMs through language-based feedback, and AutoPSV, an automated verification system that evaluates reasoning steps without manual annotation. For autoformalization, we develop Process-Driven Autoformalization (PDA), which leverages compiler feedback to enhance translation quality, and FormalAlign, an automated framework for evaluating semantic alignment between informal and formal mathematics. In automated theorem proving, we propose POETRY, a recursive theorem-proving approach that significantly extends proof length capabilities through level-based search strategies.
Our research establishes a comprehensive framework for enhancing mathematical reasoning capabilities in LLMs through three interconnected domains. Experimental results demonstrate substantial improvements across multiple stages: in natural language reasoning, SELF achieves continuous performance improvement without human intervention, while AutoPSV improves verification accuracy by 15\% through automated process supervision. In bridging natural and formal mathematics, PDA increases compiler success rates by 25\% with process-driven training, and FormalAlign outperforms GPT-4 by 11.58\% in semantic alignment evaluation. Finally, in formal mathematics, POETRY extends the maximum provable theorem length from 10 to 26 steps while improving the average proving success rate by 5.1\%.
These quantitative advances demonstrate the effectiveness of our systematic approach in leveraging both natural language and formal mathematical capabilities to enhance LLMs' mathematical reasoning. |
| Degree | Doctor of Philosophy |
| Subject | Natural language processing (Computer science) Artificial intelligence |
| Dept/Program | Computer Science |
| Persistent Identifier | http://hdl.handle.net/10722/355601 |
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.advisor | Huang, Z | - |
| dc.contributor.author | Lu, Jianqiao | - |
| dc.contributor.author | 陸劍橋 | - |
| dc.date.accessioned | 2025-04-23T01:31:20Z | - |
| dc.date.available | 2025-04-23T01:31:20Z | - |
| dc.date.issued | 2025 | - |
| dc.identifier.citation | Lu, J. [陸劍橋]. (2025). Mathematical reasoning in large language models : bridging natural language and formal mathematics. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR. | - |
| dc.identifier.uri | http://hdl.handle.net/10722/355601 | - |
| dc.description.abstract | Large Language Models (LLMs) increasingly confront the challenge of processing and reasoning about mathematical content in both natural language and formal systems. This complexity arises from the dual nature of mathematical reasoning: understanding informal mathematical text and producing verifiable formal proofs. The challenge is further amplified by the need to bridge these two domains effectively while maintaining mathematical rigor. In this thesis, we investigate mathematical reasoning in LLMs with a focus on three interconnected areas: natural language mathematical reasoning, autoformalization, and automated theorem proving. In the natural language reasoning domain, we introduce SELF, a novel framework enabling autonomous improvement of LLMs through language-based feedback, and AutoPSV, an automated verification system that evaluates reasoning steps without manual annotation. For autoformalization, we develop Process-Driven Autoformalization (PDA), which leverages compiler feedback to enhance translation quality, and FormalAlign, an automated framework for evaluating semantic alignment between informal and formal mathematics. In automated theorem proving, we propose POETRY, a recursive theorem-proving approach that significantly extends proof length capabilities through level-based search strategies. Our research establishes a comprehensive framework for enhancing mathematical reasoning capabilities in LLMs through three interconnected domains. Experimental results demonstrate substantial improvements across multiple stages: in natural language reasoning, SELF achieves continuous performance improvement without human intervention, while AutoPSV improves verification accuracy by 15\% through automated process supervision. In bridging natural and formal mathematics, PDA increases compiler success rates by 25\% with process-driven training, and FormalAlign outperforms GPT-4 by 11.58\% in semantic alignment evaluation. Finally, in formal mathematics, POETRY extends the maximum provable theorem length from 10 to 26 steps while improving the average proving success rate by 5.1\%. These quantitative advances demonstrate the effectiveness of our systematic approach in leveraging both natural language and formal mathematical capabilities to enhance LLMs' mathematical reasoning. | - |
| dc.language | eng | - |
| dc.publisher | The University of Hong Kong (Pokfulam, Hong Kong) | - |
| dc.relation.ispartof | HKU Theses Online (HKUTO) | - |
| dc.rights | The author retains all proprietary rights, (such as patent rights) and the right to use in future works. | - |
| dc.rights | This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. | - |
| dc.subject.lcsh | Natural language processing (Computer science) | - |
| dc.subject.lcsh | Artificial intelligence | - |
| dc.title | Mathematical reasoning in large language models : bridging natural language and formal mathematics | - |
| dc.type | PG_Thesis | - |
| dc.description.thesisname | Doctor of Philosophy | - |
| dc.description.thesislevel | Doctoral | - |
| dc.description.thesisdiscipline | Computer Science | - |
| dc.description.nature | published_or_final_version | - |
| dc.date.hkucongregation | 2025 | - |
| dc.identifier.mmsid | 991044954590003414 | - |
