Select High-Level Features: Efficient Experts from a Hierarchical Classification Network

AmazUtah_NLP at SemEval-2024 Task 9: A MultiChoice Question Answering System for Commonsense Defying Reasoning


View a PDF of the paper titled Select High-Level Features: Efficient Experts from a Hierarchical Classification Network, by Andr’e Kelm and 7 other authors

View PDF
HTML (experimental)

Abstract:This study introduces a novel expert generation method that dynamically reduces task and computational complexity without compromising predictive performance. It is based on a new hierarchical classification network topology that combines sequential processing of generic low-level features with parallelism and nesting of high-level features. This structure allows for the innovative extraction technique: the ability to select only high-level features of task-relevant categories. In certain cases, it is possible to skip almost all unneeded high-level features, which can significantly reduce the inference cost and is highly beneficial in resource-constrained conditions. We believe this method paves the way for future network designs that are lightweight and adaptable, making them suitable for a wide range of applications, from compact edge devices to large-scale clouds. In terms of dynamic inference our methodology can achieve an exclusion of up to 88.7,% of parameters and 73.4,% fewer giga-multiply accumulate (GMAC) operations, analysis against comparative baselines showing an average reduction of 47.6,% in parameters and 5.8,% in GMACs across the cases we evaluated.

Submission history

From: Andre Kelm [view email]
[v1]
Fri, 8 Mar 2024 00:02:42 UTC (95 KB)
[v2]
Wed, 20 Nov 2024 08:42:04 UTC (139 KB)



Source link
lol

By stp2y

Leave a Reply

Your email address will not be published. Required fields are marked *

No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.