🤖 AI Summary
This work addresses the Spanish-language table question answering (TQA) task by proposing a multi-stage natural language-to-code reasoning framework. Methodologically, it extends the Multi-Step Reasoning and Translation (MRT) framework with a stepwise inference pipeline: table structure understanding, key column identification, natural language instruction generation, Python code synthesis, and safe execution—augmented by fine-grained prompt optimization and robust exception handling. The core contribution lies in decomposing complex table semantic parsing into verifiable, intermediate reasoning steps, thereby significantly improving code generation accuracy and interpretability of open-source large language models (LLMs) for low-resource languages. Evaluated on the IberLEF-2025 PRESTA benchmark, the framework achieves 85% exact-match accuracy, outperforming existing end-to-end approaches.
📝 Abstract
This paper presents our approach for the IberLEF 2025 Task PRESTA: Preguntas y Respuestas sobre Tablas en Español (Questions and Answers about Tables in Spanish). Our solution obtains answers to the questions by implementing Python code generation with LLMs that is used to filter and process the table. This solution evolves from the MRT implementation for the Semeval 2025 related task. The process consists of multiple steps: analyzing and understanding the content of the table, selecting the useful columns, generating instructions in natural language, translating these instructions to code, running it, and handling potential errors or exceptions. These steps use open-source LLMs and fine-grained optimized prompts for each step. With this approach, we achieved an accuracy score of 85% in the task.