View a PDF of the paper titled Breaking Language Barriers in Multilingual Mathematical Reasoning: Insights and Observations, by Nuo Chen and 5 other authors
Abstract:Existing research predominantly focuses on developing powerful language learning models (LLMs) for mathematical reasoning within monolingual languages, with few explorations in preserving efficacy in a multilingual context. To bridge this gap, this paper pioneers exploring and training powerful Multilingual Math Reasoning (xMR) LLMs. Firstly, by utilizing translation, we construct the first multilingual math reasoning instruction dataset, MGSM8KInstruct, encompassing ten distinct languages, thus addressing the issue of training data scarcity in xMR tasks. Based on the collected dataset, we propose different training strategies to build powerful xMR LLMs, named MathOctopus, notably outperform conventional open-source LLMs and exhibit superiority over ChatGPT in few-shot scenarios. Notably, MathOctopus-13B reaches 47.6% accuracy which exceeds ChatGPT 46.3% on MGSM testset. Beyond remarkable results, we unearth several pivotal observations and insights from extensive experiments: (1) When extending the rejection sampling strategy to the multilingual context, it proves effective for model performances, albeit limited. (2) Employing parallel corpora for math Supervised Fine-Tuning (SFT) across multiple languages not only significantly enhances model performance multilingually but also elevates their monolingual performance. This indicates that crafting multilingual corpora can be regarded as a vital strategy for enhancing model performance in a specific language, especially in mathematical reasoning tasks. For instance, MathOctopus-7B improves its counterparts that trained on English from 42.2% to 50.8% on GSM8K testset. Codes are available at this https URL.
Submission history
From: Nuo Chen [view email]
[v1]
Tue, 31 Oct 2023 08:09:20 UTC (512 KB)
[v2]
Wed, 1 Nov 2023 06:56:14 UTC (739 KB)
[v3]
Tue, 7 Nov 2023 12:13:02 UTC (739 KB)
[v4]
Tue, 28 Nov 2023 05:25:14 UTC (739 KB)
[v5]
Wed, 16 Oct 2024 04:26:04 UTC (8,560 KB)
Source link
lol