人工知能学会全国大会論文集
Online ISSN : 2758-7347
38th (2024)
セッションID: 3Xin2-111
会議情報

Pyramid of Thought: A Novel Approach for Enhancing Chain-of-Thought Reasoning in Large Language Models
*Lei SUNYoudi LI
著者情報
会議録・要旨集 フリー

詳細
抄録

Chain-of-Thought (CoT) prompting shows promise in unleashing the latent reasoning abilities of large language models (LLMs). While CoT prompts guide LLMs through a step-wise reasoning chain, its traditional format encounters challenges with truly complex problems. In response, we introduce the Pyramid of Thought (PoT), a pioneering prompting strategy inspired by the Pyramid Principle. Unlike CoT's linear chain, Our PoT establishes a layered structure of 'facts'—mutually exclusive or exhaustive subsets relevant to the problem. This hierarchical architecture guides LLMs through multiple reasoning steps, offering focused clarity to tackle complexity. Through extensive experiments, PoT demonstrates significant accuracy improvement of 6% compared to CoT.

著者関連情報
© 2024 The Japanese Society for Artificial Intelligence
前の記事 次の記事
feedback
Top