Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Accurate and Regret-aware Numerical Problem Solver for Tabular Question Answering

About

Question answering on free-form tables (a.k.a. TableQA) is a challenging task because of the flexible structure and complex schema of tables. Recent studies use Large Language Models (LLMs) for this task, exploiting their capability in understanding the questions and tabular data, which are typically given in natural language and contain many textual fields, respectively. While this approach has shown promising results, it overlooks the challenges brought by numerical values which are common in tabular data, and LLMs are known to struggle with such values. We aim to address this issue, and we propose a model named TabLaP that uses LLMs as a planner rather than an answer generator. This approach exploits LLMs' capability in multi-step reasoning while leaving the actual numerical calculations to a Python interpreter for accurate calculation. Recognizing the inaccurate nature of LLMs, we further make a first attempt to quantify the trustworthiness of the answers produced by TabLaP, such that users can use TabLaP in a regret-aware manner. Experimental results on two benchmark datasets show that TabLaP is substantially more accurate than the state-of-the-art models, improving the answer accuracy by 5.7% and 5.8% on the two datasets, respectively.

Yuxiang Wang, Jianzhong Qi, Junhao Gan• 2024

Related benchmarks

TaskDatasetResultRank
Table Question AnsweringWTQ
Accuracy76.6
101
Table Question AnsweringWikiTQ (test)
Accuracy76.6
92
Table Question AnsweringWikiTableQuestions (test)
Accuracy76.6
86
Table Question AnsweringTableBench
EM61.2
40
Table Question AnsweringPenguins in a Table
EM93.1
40
TableQATabFact small (test)
Accuracy90.4
8
Table Question AnsweringFTQ
Accuracy44.1
7
Showing 7 of 7 rows

Other info

Code

Follow for update