View a PDF of the paper titled Acquiring Bidirectionality via Large and Small Language Models, by Takumi Goto and 2 other authors
Abstract:Using token representation from bidirectional language models (LMs) such as BERT is still a widely used approach for token-classification tasks. Even though there exist much larger unidirectional LMs such as Llama-2, they are rarely used to replace the token representation of bidirectional LMs. In this work, we hypothesize that their lack of bidirectionality is keeping them behind. To that end, we propose to newly train a small backward LM and concatenate its representations to those of existing LM for downstream tasks. Through experiments in named entity recognition, we demonstrate that introducing backward model improves the benchmark performance more than 10 points. Furthermore, we show that the proposed method is especially effective for rare domains and in few-shot learning settings.
Submission history
From: Takumi Goto [view email]
[v1]
Mon, 19 Aug 2024 01:54:37 UTC (90 KB)
[v2]
Tue, 10 Dec 2024 04:29:24 UTC (52 KB)
Source link
lol