Question Answering with Texts and Tables through Deep Reinforcement Learning

📅 2024-07-05
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenge of joint reasoning over textual and tabular data in open-domain multi-hop question answering, this paper proposes an end-to-end cross-modal reasoning framework based on deep reinforcement learning. Methodologically, it introduces Proximal Policy Optimization (PPO)—the first application of policy gradient methods to text-table joint QA—integrating a multimodal encoder (BERT + TabTransformer) with a differentiable table operation module to dynamically plan reading order and operational actions, eliminating the need for predefined alignments or intermediate supervision. Its key innovations include joint optimization of cross-modal reasoning paths and Monte Carlo policy evaluation. The framework achieves state-of-the-art performance on WikiTableQuestions and HybridQA, improving accuracy by 3.2% and 4.7%, respectively, demonstrating substantial gains in complex, cross-table, multi-hop reasoning capability.

Technology Category

Application Category

Problem

Research questions and friction points this paper is trying to address.

Multi-hop QA with texts and tables
Sequential model selection challenge
Reinforcement learning for optimal tool choice
Innovation

Methods, ideas, or system contributions that make the work stand out.

Deep Reinforcement Learning
Multi-hop Answer Generation
Open Table-and-Text Dataset
🔎 Similar Papers
No similar papers found.
M
Marcos M. José
Escola Politécnica, Universidade de São Paulo, Sao Paulo, Brazil
F
Flávio Nakasato Cação
Escola Politécnica, Universidade de São Paulo, Sao Paulo, Brazil
M
Maria F. Ribeiro
R
Rafael M. Cheang
Escola Politécnica, Universidade de São Paulo, Sao Paulo, Brazil
P
Paulo Pirozelli
Escola Politécnica, Universidade de São Paulo, Sao Paulo, Brazil; Instituto Mauá de Tecnologia
F
Fabio G. Cozman
Escola Politécnica, Universidade de São Paulo, Sao Paulo, Brazil