Learning to Plan for Retrieval-Augmented Large Language Models from Knowledge Graphs

AmazUtah_NLP at SemEval-2024 Task 9: A MultiChoice Question Answering System for Commonsense Defying Reasoning


View a PDF of the paper titled Learning to Plan for Retrieval-Augmented Large Language Models from Knowledge Graphs, by Junjie Wang and 12 other authors

View PDF
HTML (experimental)

Abstract:Improving the performance of large language models (LLMs) in complex question-answering (QA) scenarios has always been a research focal point. Recent studies have attempted to enhance LLMs’ performance by combining step-wise planning with external retrieval. While effective for advanced models like GPT-3.5, smaller LLMs face challenges in decomposing complex questions, necessitating supervised fine-tuning. Previous work has relied on manual annotation and knowledge distillation from teacher LLMs, which are time-consuming and not accurate enough. In this paper, we introduce a novel framework for enhancing LLMs’ planning capabilities by using planning data derived from knowledge graphs (KGs). LLMs fine-tuned with this data have improved planning capabilities, better equipping them to handle complex QA tasks that involve retrieval. Evaluations on multiple datasets, including our newly proposed benchmark, highlight the effectiveness of our framework and the benefits of KG-derived planning data.

Submission history

From: Junjie Wang [view email]
[v1]
Thu, 20 Jun 2024 13:07:38 UTC (1,527 KB)
[v2]
Wed, 9 Oct 2024 05:56:07 UTC (1,529 KB)
[v3]
Wed, 23 Oct 2024 09:42:59 UTC (1,529 KB)



Source link
lol

By stp2y

Leave a Reply

Your email address will not be published. Required fields are marked *

No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.