View a PDF of the paper titled CorPipe at CRAC 2024: Predicting Zero Mentions from Raw Text, by Milan Straka
Abstract:We present CorPipe 24, the winning entry to the CRAC 2024 Shared Task on Multilingual Coreference Resolution. In this third iteration of the shared task, a novel objective is to also predict empty nodes needed for zero coreference mentions (while the empty nodes were given on input in previous years). This way, coreference resolution can be performed on raw text. We evaluate two model variants: a~two-stage approach (where the empty nodes are predicted first using a pretrained encoder model and then processed together with sentence words by another pretrained model) and a single-stage approach (where a single pretrained encoder model generates empty nodes, coreference mentions, and coreference links jointly). In both settings, CorPipe surpasses other participants by a large margin of 3.9 and 2.8 percent points, respectively. The source code and the trained model are available at this https URL.
Submission history
From: Milan Straka [view email]
[v1]
Thu, 3 Oct 2024 17:58:55 UTC (267 KB)
[v2]
Sat, 9 Nov 2024 23:24:14 UTC (267 KB)
Source link
lol