MRT at IberLEF-2025 PRESTA Task: Maximizing Recovery from Tables with Multiple Steps

📅 2025-07-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the Spanish-language table question answering (TQA) task by proposing a multi-stage natural language-to-code reasoning framework. Methodologically, it extends the Multi-Step Reasoning and Translation (MRT) framework with a stepwise inference pipeline: table structure understanding, key column identification, natural language instruction generation, Python code synthesis, and safe execution—augmented by fine-grained prompt optimization and robust exception handling. The core contribution lies in decomposing complex table semantic parsing into verifiable, intermediate reasoning steps, thereby significantly improving code generation accuracy and interpretability of open-source large language models (LLMs) for low-resource languages. Evaluated on the IberLEF-2025 PRESTA benchmark, the framework achieves 85% exact-match accuracy, outperforming existing end-to-end approaches.

Technology Category

Application Category

📝 Abstract
This paper presents our approach for the IberLEF 2025 Task PRESTA: Preguntas y Respuestas sobre Tablas en Español (Questions and Answers about Tables in Spanish). Our solution obtains answers to the questions by implementing Python code generation with LLMs that is used to filter and process the table. This solution evolves from the MRT implementation for the Semeval 2025 related task. The process consists of multiple steps: analyzing and understanding the content of the table, selecting the useful columns, generating instructions in natural language, translating these instructions to code, running it, and handling potential errors or exceptions. These steps use open-source LLMs and fine-grained optimized prompts for each step. With this approach, we achieved an accuracy score of 85% in the task.
Problem

Research questions and friction points this paper is trying to address.

Generate Python code from tables for Q&A
Filter and process Spanish table data
Optimize multi-step LLM table analysis
Innovation

Methods, ideas, or system contributions that make the work stand out.

Python code generation with LLMs
Fine-grained optimized prompts
Multi-step table processing
🔎 Similar Papers
No similar papers found.
Maximiliano Hormazábal Lagos
Maximiliano Hormazábal Lagos
PhD Student
Computer VisionNatural Language ProcessingDocument Image AnalysisVision Language Models
Á
Álvaro Bueno Sáez
Fundación Centro Tecnolóxico de Telecomunicacións de Galicia (GRADIANT), Vigo, Spain
H
Héctor Cerezo-Costas
Fundación Centro Tecnolóxico de Telecomunicacións de Galicia (GRADIANT), Vigo, Spain
P
Pedro Alonso Doval
Fundación Centro Tecnolóxico de Telecomunicacións de Galicia (GRADIANT), Vigo, Spain
J
Jorge Alcalde Vesteiro
Fundación Centro Tecnolóxico de Telecomunicacións de Galicia (GRADIANT), Vigo, Spain