LOLA — An Open-Source Massively Multilingual Large Language Model

AmazUtah_NLP at SemEval-2024 Task 9: A MultiChoice Question Answering System for Commonsense Defying Reasoning


View a PDF of the paper titled LOLA — An Open-Source Massively Multilingual Large Language Model, by Nikit Srivastava and 7 other authors

View PDF
HTML (experimental)

Abstract:This paper presents LOLA, a massively multilingual large language model trained on more than 160 languages using a sparse Mixture-of-Experts Transformer architecture. Our architectural and implementation choices address the challenge of harnessing linguistic diversity while maintaining efficiency and avoiding the common pitfalls of multilinguality. Our analysis of the evaluation results shows competitive performance in natural language generation and understanding tasks. Additionally, we demonstrate how the learned expert-routing mechanism exploits implicit phylogenetic linguistic patterns to potentially alleviate the curse of multilinguality. We provide an in-depth look at the training process, an analysis of the datasets, and a balanced exploration of the model’s strengths and limitations. As an open-source model, LOLA promotes reproducibility and serves as a robust foundation for future research. Our findings enable the development of compute-efficient multilingual models with strong, scalable performance across languages.

Submission history

From: Nikit Srivastava [view email]
[v1]
Tue, 17 Sep 2024 15:23:08 UTC (718 KB)
[v2]
Wed, 18 Sep 2024 13:55:04 UTC (718 KB)
[v3]
Thu, 19 Sep 2024 15:50:01 UTC (718 KB)
[v4]
Mon, 9 Dec 2024 13:21:45 UTC (907 KB)



Source link
lol

By stp2y

Leave a Reply

Your email address will not be published. Required fields are marked *

No widgets found. Go to Widget page and add the widget in Offcanvas Sidebar Widget Area.