View a PDF of the paper titled Improving In-Context Learning with Small Language Model Ensembles, by M. Mehdi Mojarradi and Lingyi Yang and Robert McCraith and Adam Mahdi
Abstract:Large language models (LLMs) have shown impressive capabilities across various tasks, but their performance on domain-specific tasks remains limited. While methods like retrieval augmented generation and fine-tuning can help to address this, they require significant resources. In-context learning (ICL) is a cheap and efficient alternative but cannot match the accuracies of advanced methods. We present Ensemble SuperICL, a novel approach that enhances ICL by leveraging the expertise of multiple fine-tuned small language models (SLMs). Ensemble SuperICL achieves state of the art (SoTA) results on several natural language understanding benchmarks. Additionally, we test it on a medical-domain labelling task and showcase its practicality by using off-the-shelf SLMs fine-tuned on a general language task, achieving superior accuracy in large-scale data labelling compared to all baselines. Finally, we conduct an ablation study and sensitivity analyses to elucidate the underlying mechanism of Ensemble SuperICL. Our research contributes to the growing demand for efficient domain specialisation methods in LLMs, offering a cheap and effective method for practitioners.
Submission history
From: Lingyi Yang [view email]
[v1]
Tue, 29 Oct 2024 09:02:37 UTC (430 KB)
[v2]
Fri, 20 Dec 2024 12:22:37 UTC (430 KB)
Source link
lol