All Posts
ENApril 4, 2026 6 min read

Debunking the Curse of Multilinguality: New Evidence on Language Model Pretraining

The field of multilingual natural language processing has long grappled with a seemingly fundamental trade-off: the more languages you include in pretraining, the worse your model performs on each individual language. This phenomenon, dubbed the "curse of multilinguality," has shaped how researchers and practitioners approach multilingual model development. However, a new study titled "Revisiting Multilingual Data Mixtures in Language Model Pretraining" by Foroutan et al. from EPFL challenges this conventional wisdom with compelling evidence that reframes our understanding of multilingual pretraining entirely.

The Conventional Wisdom Under Scrutiny

The curse of multilinguality has been a persistent concern in the field, suggesting that model capacity is fundamentally limited and that adding more languages necessarily degrades performance across the board. This belief has led to careful language selection strategies and concerns about the scalability of truly multilingual models. Previous studies, however, have been constrained by either small model sizes (45M to 85M parameters) or limited language coverage (typically fewer than 25 languages).

The EPFL researchers approached this problem with unprecedented scale and rigor, training 1.1B and 3B parameter models on corpora containing up to 400 languages across 100-225 billion tokens. This scale allows for more definitive conclusions about the fundamental nature of multilingual interference effects.

Token Sufficiency: The Real Bottleneck

The study's most significant finding challenges the core assumption underlying the curse of multilinguality. The researchers demonstrate that performance degradation is not an inevitable consequence of adding more languages, but rather stems from insufficient tokens per language. When each language receives adequate representation in the training corpus, models can successfully learn from hundreds of languages without meaningful performance degradation.

This finding has profound implications for how we conceptualize multilingual model capacity. The bottleneck appears to be data collection and curation rather than fundamental architectural limitations. The researchers show that once a language crosses a threshold of token sufficiency, the marginal cost of including additional languages approaches zero.

The experimental evidence is particularly compelling when examining their 400-language experiments. Rather than observing catastrophic performance degradation as predicted by the curse of multilinguality, they found that models maintained stable performance across languages when token distributions were appropriately balanced. This suggests that previous observations of the curse may have been artifacts of inadequate data rather than fundamental capacity constraints.

English as a Universal Pivot Language

Another striking finding concerns the role of pivot languages in multilingual transfer learning. Conventional wisdom suggested that selecting pivot languages from within specific language families would maximize transfer learning benefits. For instance, using Russian as a pivot for Slavic languages or Arabic for Semitic languages seemed intuitive given shared linguistic features and evolutionary relationships.

The study reveals that English consistently outperforms family-specific pivots across all language families tested. This counterintuitive result has several important implications. First, it suggests that the quality and diversity of training data may matter more than linguistic similarity for effective transfer learning. English benefits from having the highest quality and most diverse web content, which appears to provide richer representational foundations for multilingual generalization.

Second, this finding simplifies multilingual pretraining strategies significantly. Rather than requiring complex language family analysis and family-specific pivot selection, practitioners can leverage English's robust data ecosystem to improve performance across linguistically diverse languages. This has practical advantages for model development pipelines and reduces the complexity of multilingual data curation.

The Failure of Curriculum Learning

The researchers also investigated whether curriculum learning, gradually introducing languages during training, could mitigate negative interference effects. This approach has theoretical appeal, as it allows models to establish strong representations in high-resource languages before adapting to lower-resource ones.

However, their experiments reveal that curriculum learning provides no measurable benefits over simultaneous multilingual training. Models trained with staged language introduction performed no better than those exposed to all languages from the beginning. This finding suggests that the interference effects addressed by curriculum learning may not be as problematic as previously thought, at least when token sufficiency requirements are met.

My Analysis: Rethinking Multilingual Model Development

These findings collectively point to a fundamental shift in how we should approach multilingual model development. The traditional framing of multilingual pretraining as a zero-sum game between language coverage and individual language performance appears to be largely incorrect, at least at the scales investigated.

The token sufficiency threshold concept deserves particular attention. While the paper doesn't provide precise thresholds for different languages, the principle suggests that multilingual model development should prioritize achieving minimum viable token counts across target languages rather than optimizing language selection. This reframes resource allocation decisions from "which languages to include" to "how to efficiently collect sufficient data for all target languages."

The English pivot finding also has broader implications for cross-lingual transfer research. The superiority of English may not solely reflect its data abundance but could also indicate that linguistic diversity within a single high-resource language provides better transfer foundations than linguistic similarity across language families. English's extensive domain coverage, register variation, and stylistic diversity might create more robust multilingual representations than family-specific pivots with more limited data diversity.

Limitations and Future Directions

Despite the study's significant contributions, several limitations warrant consideration. The experiments focus on models up to 3B parameters, and scaling behaviors may differ at larger scales where capacity constraints become more pronounced. Additionally, the evaluation primarily relies on language modeling perplexity and downstream task performance, which may not capture all aspects of multilingual competence.

The token sufficiency thresholds also require more precise characterization. Different languages likely have different minimum requirements based on factors like morphological complexity, orthographic systems, and domain coverage needs. Understanding these language-specific requirements would enable more efficient multilingual corpus design.

Furthermore, the study focuses on web-scraped data, which may have different characteristics than carefully curated multilingual corpora. The generalizability of these findings to other data sources and domains remains an open question.

Implications for the Field

These results suggest that the multilingual NLP community has been operating under unnecessarily pessimistic assumptions about multilingual model scalability. Rather than viewing multilingual coverage as fundamentally limited by model capacity, we can approach it as a data engineering challenge. This perspective shift opens new possibilities for truly comprehensive multilingual models covering hundreds of languages without significant performance trade-offs.

The findings also highlight the importance of data quality and diversity over linguistic similarity in multilingual transfer learning. This could influence how we approach low-resource language support, suggesting that leveraging high-quality English data may be more effective than seeking linguistically similar pivot languages with limited data.

Moving forward, the field would benefit from more precise characterization of token sufficiency requirements across different languages and linguistic families. Additionally, investigating these phenomena at larger model scales and with different architectural approaches would strengthen our understanding of multilingual scaling laws. The ultimate goal should be developing principled approaches to multilingual data curation that maximize language coverage while ensuring adequate representation for each included language.

Debunking the Curse of Multilinguality: New Evidence on Language Model Pretraining | kualia.ai