Papers
arxiv:2602.11961

Scaling Model and Data for Multilingual Machine Translation with Open Large Language Models

Published on Feb 12
Authors:
,
,
,
,

Abstract

Open large language models demonstrate improved multilingual capabilities through continual pretraining and instruction fine-tuning, achieving state-of-the-art performance in multilingual machine translation across 46 languages.

AI-generated summary

Open large language models (LLMs) have demonstrated improving multilingual capabilities in recent years. In this paper, we present a study of open LLMs for multilingual machine translation (MT) across a range of languages, and investigate the effects of model scaling and data scaling when adapting open LLMs to multilingual MT through continual pretraining and instruction finetuning. Based on the Gemma3 model family, we develop MiLMMT-46, which achieves top-tier multilingual translation performance across 46 languages. Extensive experiments show that MiLMMT-46 consistently outperforms recent state-of-the-art (SOTA) models, including Seed-X, HY-MT-1.5, and TranslateGemma, and achieves competitive performance with strong proprietary systems such as Google Translate and Gemini 3 Pro.

Community

Sign up or log in to comment

Models citing this paper 6

Browse 6 models citing this paper

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2602.11961 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2602.11961 in a Space README.md to link it from this page.

Collections including this paper 1