🤖 AI Summary
Frequent tax code revisions impose high maintenance costs and error-prone updates on tax compliance software. Method: This paper proposes an LLM-based automated code diff generation and secure integration framework. It employs rule-guided prompt engineering and structured parsing of IRS official publications to systematically model and address six domain-specific challenges in tax-law-to-code translation—including legal ambiguity, implicit logical dependencies, and cross-year provision coupling. Contribution/Results: We introduce the first LLM-driven code diff paradigm tailored for tax software. Evaluated on ChatGPT and Llama, our approach demonstrates feasibility in accurately parsing statutory text and generating executable, semantically faithful code changes. Experiments show significant improvements in maintenance efficiency and correctness, establishing a reusable technical foundation for sustainable evolution of tax compliance systems.
📝 Abstract
As the US tax law evolves to adapt to ever-changing politico-economic realities, tax preparation software plays a significant role in helping taxpayers navigate these complexities. The dynamic nature of tax regulations poses a significant challenge to accurately and timely maintaining tax software artifacts. The state-of-the-art in maintaining tax prep software is time-consuming and error-prone as it involves manual code analysis combined with an expert interpretation of tax law amendments. We posit that the rigor and formality of tax amendment language, as expressed in IRS publications, makes it amenable to automatic translation to executable specifications (code). Our research efforts focus on identifying, understanding, and tackling technical challenges in leveraging Large Language Models (LLMs), such as ChatGPT and Llama, to faithfully extract code differentials from IRS publications and automatically integrate them with the prior version of the code to automate tax prep software maintenance.