Evaluating LLMs for Quotation Attribution in Literary Texts: A Case Study of LLaMa3

AmazUtah_NLP at SemEval-2024 Task 9: A MultiChoice Question Answering System for Commonsense Defying Reasoning


View a PDF of the paper titled Evaluating LLMs for Quotation Attribution in Literary Texts: A Case Study of LLaMa3, by Gaspard Michel and Elena V. Epure and Romain Hennequin and Christophe Cerisara

View PDF
HTML (experimental)

Abstract:Large Language Models (LLMs) have shown promising results in a variety of literary tasks, often using complex memorized details of narration and fictional characters. In this work, we evaluate the ability of Llama-3 at attributing utterances of direct-speech to their speaker in novels. The LLM shows impressive results on a corpus of 28 novels, surpassing published results with ChatGPT and encoder-based baselines by a large margin. We then validate these results by assessing the impact of book memorization and annotation contamination. We found that these types of memorization do not explain the large performance gain, making Llama-3 the new state-of-the-art for quotation attribution in English literature. We release publicly our code and data.

Submission history

From: Gaspard Michel [view email]
[v1]
Mon, 17 Jun 2024 09:56:46 UTC (235 KB)
[v2]
Thu, 23 Jan 2025 15:44:07 UTC (8,517 KB)



Source link
lol

By stp2y

Leave a Reply

Your email address will not be published. Required fields are marked *

No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.