Transactions of the Japanese Society for Artificial Intelligence
Online ISSN : 1346-8030
Print ISSN : 1346-0714
ISSN-L : 1346-0714
Original Paper
Step-Wise Selection from Multiple Reasoning Chains to Enhance LLMs for Multi-hop QA
Tao XuYuri MurayamaKiyoshi Izumi
Author information
JOURNAL OPEN ACCESS FULL-TEXT HTML

2025 Volume 40 Issue 5 Pages A-P21_1-13

Details
Abstract

Multi-hop question answering refers to tasks that involve gathering clues from multiple information sourcesiteratively to arrive at a final answer. A recent research direction focuses on using large language models (LLMs)combined with chain-of-thought and retrieval-augmented generation to address multi-hop QA tasks. These methodshave significantly improved LLM performance. However, hallucination by LLMs negatively impacts the task accuracy.The Self-consistency approach reduces hallucination by having LLMs generate multiple reasoning chains andchoosing the majority answer. This method requires the majority reasoning chains lead to the correct answer, whichmay not be true for the complex tasks such as multi-hop QA. Since LLMs may make mistake at each reasoning step,the likelihood of a fully correct reasoning chain is lower than that of a single correct step. Therefore, we propose amethod that shifts the focus from selecting reasoning chains to selecting steps. This method involves three phases:(1) Have the LLM break down a question into step-by-step sub-questions, and use RAG to derive the final answer.Repeat this process to generate multiple reasoning chains; (2) Use a verifier to select most reliable steps from all reasoningchains; (3) Based on the selected steps, let another independent LLM deduce the final answer. We conductedexperiments on four multi-hop QA datasets, and the results show that our method outperforms strong baselines.

Content from these authors
© JSAI (The Japanese Society for Artificial Intelligence)
Next article
feedback
Top