Reducing Reasoning Costs: The Path of Optimization for Chain of Thought via Sparse Attention Mechanism

AmazUtah_NLP at SemEval-2024 Task 9: A MultiChoice Question Answering System for Commonsense Defying Reasoning


View a PDF of the paper titled Reducing Reasoning Costs: The Path of Optimization for Chain of Thought via Sparse Attention Mechanism, by Libo Wang

View PDF

Abstract:In order to address the chain of thought in the large language model inference cost surge, this research proposes to use a sparse attention mechanism that only focuses on a few relevant tokens. The researcher constructed a new attention mechanism and used GiantRabbit trained with custom GPTs as an experimental tool. The experiment tested and compared the reasoning time, correctness score and chain of thought length of this model and o1 Preview in solving the linear algebra test questions of MIT OpenCourseWare. The results show that GiantRabbit’s reasoning time and chain of thought length are significantly lower than o1 Preview. It verifies the feasibility of sparse attention mechanism for optimizing chain of thought reasoning. Detailed architectural details and experimental process have been uploaded to Github, the link is:this https URL.

Submission history

From: Libo Wang [view email]
[v1]
Thu, 14 Nov 2024 00:59:13 UTC (625 KB)
[v2]
Fri, 15 Nov 2024 21:28:27 UTC (625 KB)
[v3]
Sun, 1 Dec 2024 13:08:57 UTC (786 KB)
[v4]
Wed, 11 Dec 2024 18:50:30 UTC (1,015 KB)
[v5]
Thu, 23 Jan 2025 16:09:32 UTC (1,949 KB)



Source link
lol

By stp2y

Leave a Reply

Your email address will not be published. Required fields are marked *

No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.