Efficient multi-agent collaboration with tool use for online planning in complex table question answering

  • Complex table question answering (TQA) aims to answer questions that require complex reasoning, such as multi-step or multi-category reasoning, over data represented in tabular form. Previous approaches demonstrate notable performance by leveraging either closed-source large language models (LLMs) or fine-tuned open-weight LLMs. However, fine-tuning LLMs requires high-quality training data, which is costly to obtain. The use of closed-source LLMs poses accessibility challenges and leads to reproducibility issues. In this paper, we propose Multi Agent Collaboration with Tool use (MACT), a framework that requires neither fine-tuning nor closed-source models. In MACT, a planning agent and a coding agent that also make use of tools collaborate for TQA. MACT outperforms previous SoTA systems on three out of four benchmarks and performs comparably to the larger and more expensive closed-source model GPT-4 on two benchmarks, even when using only open-weight models without any fine-tuning. OurComplex table question answering (TQA) aims to answer questions that require complex reasoning, such as multi-step or multi-category reasoning, over data represented in tabular form. Previous approaches demonstrate notable performance by leveraging either closed-source large language models (LLMs) or fine-tuned open-weight LLMs. However, fine-tuning LLMs requires high-quality training data, which is costly to obtain. The use of closed-source LLMs poses accessibility challenges and leads to reproducibility issues. In this paper, we propose Multi Agent Collaboration with Tool use (MACT), a framework that requires neither fine-tuning nor closed-source models. In MACT, a planning agent and a coding agent that also make use of tools collaborate for TQA. MACT outperforms previous SoTA systems on three out of four benchmarks and performs comparably to the larger and more expensive closed-source model GPT-4 on two benchmarks, even when using only open-weight models without any fine-tuning. Our extensive analyses prove the effectiveness of MACT’s multi-agent collaboration in TQA. We release our code publicly.show moreshow less

Download full text files

Export metadata

Statistics

Number of document requests

Additional Services

Share in Twitter Search Google Scholar
Metadaten
Author:Wei Zhou, Mohsen Mesgar, Annemarie FriedrichORCiDGND, Heike Adel
URN:urn:nbn:de:bvb:384-opus4-1263934
Frontdoor URLhttps://opus.bibliothek.uni-augsburg.de/opus4/126393
ISBN:979-8-89176-195-7OPAC
Parent Title (English):Findings of the Association for Computational Linguistics: NAACL 2025 - Annual Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics, 29 April - 4 May 2025, Albuquerque, NM, USA
Publisher:Association for Computational Linguistics (ACL)
Place of publication:Stroudsburg, PA
Editor:Luis Chiruzzo, Alan Ritter, Lu Wang
Type:Conference Proceeding
Language:English
Date of Publication (online):2025/11/18
Year of first Publication:2025
Publishing Institution:Universität Augsburg
Release Date:2025/11/26
First Page:945
Last Page:968
DOI:https://doi.org/10.18653/v1/2025.findings-naacl.54
Institutes:Fakultät für Angewandte Informatik
Fakultät für Angewandte Informatik / Institut für Informatik
Fakultät für Angewandte Informatik / Institut für Informatik / Lehrstuhl für Computerlinguistik
Dewey Decimal Classification:0 Informatik, Informationswissenschaft, allgemeine Werke / 00 Informatik, Wissen, Systeme / 004 Datenverarbeitung; Informatik
Licence (German):CC-BY 4.0: Creative Commons: Namensnennung