[2602.23286] SPARTA: Scalable and Principled Benchmark of Tree-Structured Multi-hop QA over Text and Tables
Summary
The paper presents SPARTA, a novel framework for generating scalable benchmarks for tree-structured multi-hop question answering (QA) over text and tables, addressing limitations in existing benchmarks.
Why It Matters
SPARTA addresses the critical need for robust benchmarks in multi-hop QA tasks, which are essential for advancing AI's ability to reason across complex data. By automating benchmark creation, it reduces human effort while improving the quality of QA systems, thus enhancing AI's practical applications in real-world scenarios.
Key Takeaways
- SPARTA automates the creation of large-scale QA benchmarks, significantly reducing annotation time.
- The framework ensures high-quality question generation through novel techniques like provenance-based refinement.
- State-of-the-art models show significant performance drops on SPARTA, highlighting weaknesses in current cross-modal reasoning capabilities.
Computer Science > Computation and Language arXiv:2602.23286 (cs) [Submitted on 26 Feb 2026] Title:SPARTA: Scalable and Principled Benchmark of Tree-Structured Multi-hop QA over Text and Tables Authors:Sungho Park, Jueun Kim, Wook-Shin Han View a PDF of the paper titled SPARTA: Scalable and Principled Benchmark of Tree-Structured Multi-hop QA over Text and Tables, by Sungho Park and 1 other authors View PDF HTML (experimental) Abstract:Real-world Table-Text question answering (QA) tasks require models that can reason across long text and source tables, traversing multiple hops and executing complex operations such as aggregation. Yet existing benchmarks are small, manually curated - and therefore error-prone - and contain shallow questions that seldom demand more than two hops or invoke aggregations, grouping, or other advanced analytical operations expressible in natural-language queries. We present SPARTA, an end-to-end construction framework that automatically generates large-scale Table-Text QA benchmarks with lightweight human validation, requiring only one quarter of the annotation time of HybridQA. The framework first constructs a reference fact database by enriching each source table with grounding tables whose tuples are atomic facts automatically extracted from the accompanying unstructured passages, then synthesizes nested queries whose number of nested predicates matches the desired hop count. To ensure that every SQL statement is executable and that its verbal...