TableMaster: A Recipe to Advance Table Understanding with Language Models
About
Tables serve as a fundamental format for representing structured relational data. While current language models (LMs) excel at many text-based tasks, they still face challenges in table understanding due to the complex characteristics of tabular data, such as their structured nature. In this paper, we aim to enhance LMs for improved table understanding. We identify four key challenges: 1) difficulty in locating target data, 2) deficiency in table semantics, 3) numerical inaccuracies in textual reasoning, and 4) semantic inflexibility in symbolic reasoning. To address these issues, we propose TableMaster, a recipe and comprehensive framework that integrates multiple solutions to overcome these obstacles. TableMaster first extracts relevant table content and verbalizes it with enriched semantic context. Additionally, we introduce adaptive reasoning, a flexible approach that dynamically adjusts between textual and symbolic reasoning, tailoring the reasoning process to each query. Extensive analyses and experiments demonstrate our findings and the effectiveness of TableMaster. On the WikiTQ dataset, TableMaster achieves an accuracy of 78.13% using GPT-4o-mini, surpassing existing baselines. We hope this work will serve as a practical step toward more robust and reliable table understanding.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Table Fact Verification | TabFact (test) | Accuracy91.16 | 136 | |
| Table Question Answering | WikiTQ (test) | Accuracy78.13 | 130 | |
| Table Question Answering | HiTab | Accuracy74.2 | 121 | |
| Table Question Answering | WikiTQ | Accuracy78.13 | 118 | |
| Table Fact Verification | TabFact | Accuracy0.9012 | 104 | |
| Financial Question Answering | FinQA | Accuracy66.4 | 16 | |
| Hierarchical Table Question Answering | HiTab | Accuracy69.3 | 11 | |
| Long-form Question Answering | FeTaQA (test) | BLEU28.94 | 7 |