View a PDF of the paper titled LogicAsker: Evaluating and Improving the Logical Reasoning Ability of Large Language Models, by Yuxuan Wan and 7 other authors
Abstract:We introduce LogicAsker, a novel approach for evaluating and enhancing the logical reasoning capabilities of large language models (LLMs) such as ChatGPT and GPT-4. Despite LLMs’ prowess in tasks like writing assistance, code generation, and machine translation, assessing their ability to reason has been challenging. Traditional evaluations often prioritize accuracy on downstream tasks over direct assessments of reasoning processes. LogicAsker addresses this gap by employing a set of atomic reasoning skills grounded in propositional and predicate logic to systematically examine and improve the reasoning prowess of LLMs. Our methodology reveals significant gaps in LLMs’ learning of logical rules, with identified reasoning failures ranging from 29% to 90% across different models. Moreover, we leverage these findings to construct targeted demonstration examples and fine-tune data, notably enhancing logical reasoning in models like GPT-4o by up to 5%. To our knowledge, this is the first effort to utilize test case outcomes to effectively refine LLMs’ formal reasoning capabilities. We make our code, data, and results publicly available (this https URL) to facilitate further research and replication of our findings.
Submission history
From: Wenxuan Wang [view email]
[v1]
Mon, 1 Jan 2024 13:53:53 UTC (2,328 KB)
[v2]
Wed, 2 Oct 2024 16:30:34 UTC (9,072 KB)
Source link
lol