mRoBERTa Model Card

mRoBERTa is a new multilingual foundational model based on the RoBERTa architecture. It has been pretrained from scratch using 35 European languages and code. The pretraining corpus consists of 12.8TB of high-quality data. This is significantly larger compared to previous state-of-the-art encoder-only foundational models like XLM-RoBERTa-base and XLM-RoBERTa-large, whose training datasets included less multilingual data, amounting to 2.5TB.

Technical Description

Technical details of the mRoBERTa model.

Description Value
Model Parameters 283M
Tokenizer Type SPM
Vocabulary size 256,000
Precision bfloat16
Context length 512

Training Hyperparemeters

Hyperparameter Value
Pretraining Objective Masked Language Modeling
Learning Rate 7E-05
Learning Rate Scheduler Cosine
Warmup 10k
Optimizer AdamW
Optimizer Hyperparameters AdamW (β1=0.9,β2=0.98,ε =1e-06 )
Optimizer Decay 1E-02
Global Batch Size 8192
Dropout 1E-01
Attention Dropout 1E-01
Activation Function GeLU

Data

Pretraining Corpus

The training corpus consists of 35 European languages and 92 programming languages, amounting to a total of 12.8TB of high-quality data.

This highly multilingual corpus is predominantly composed of data from Colossal OSCAR, which contributes a significant 66.06% of the total tokens. Following this, Starcoder provides 11.91%, and Spanish Crawling adds 3.34%. The next largest sources are French PD at 3.12% and Proof Pile at 1.98%. Other notable contributions include Macocu, Pile of Law, and Eurlex, each contributing around 1.5% to 1.3%. These major sources collectively form the bulk of the corpus, ensuring a rich and diverse dataset for training the language model. The remaining 10% comes from smaller sources in various languages.

The final pretraining language distribution split by language can be seen in the following picture: drawing

Further details about the pretraining corpus can be found here, as it is the same as the Salamandra foundational model.

Multilingual Evaluation and Performance

Evaluation is done using multilingual benchmarks in order to assess the multilingual capabilities of the models.

The following multilingual benchmarks have been considered:

Benchmark Description Languages Source
XTREME Benchmark for the evaluation of the cross-lingual generalization ability of pre-trained multilingual models bg,ca,de,el,en,es,et,eu,fi,fr,hu,it,lt,nl,pl,pt,ro,ru,uk LINK
CLUB Human-Annotated Catalan Benchmark ca LINK
Basque Custom Benchmark A set of NER, POS and TC evaluation tasks to assess the performace in Basque language. eu LINK
Galician Custom Benchmark NER and POS evaluation tasks to assess the performace in Galician language. gl LINK LINK

The following base foundational models have been considered:

Multilingual Foundational Model Number of Parameters Vocab Size Description
BERTa 126M 52K BERTa is a Catalan-specific language model pretrained with Catalan-only data.
BERTinho 109M 30K BERTinho is monolingual BERT model for Galician language.
mBERT 178M 120K Multilingual BERT model pretrained on the top 104 languages with the largest Wikipedia.
mRoBERTa 283M 256K RoBERTa base model pretrained with 35 European languages and a larger vocabulary size.
roberta-base-bne 125M 50K RoBERTa base model pretrained with 570GB of data from web crawlings performed by the National Library of Spain from 2009 to 2019.
RoBERTa-ca 125M 50K RoBERTa-ca is a Catalan-specific language model obtained by using vocabulary adaptation from mRoBERTa.
xlm-roberta-base 279M 250K Foundational RoBERTa model pretrained with CommonCrawl data containing 100 languages.
xlm-roberta-large 561M 250K Foundational RoBERTa model pretrained with CommonCrawl data containing 100 languages.

RESULTS

This section presents results across various multilingual benchmarks, with the maximum values highlighted in bold and the second-highest values underlined.

XTREME Benchmark

The Cross-lingual TRansfer Evaluation of Multilingual Encoders (XTREME) benchmark is a benchmark for the evaluation of the cross-lingual generalization ability of pre-trained multilingual models. It includes nine tasks that collectively require reasoning about different levels of syntax and semantics. The languages in XTREME are selected to maximize language diversity, coverage in existing tasks, availability of training data, and overlap with the languages present during pre-training of the models.

🔵 Sentence Classification

🔵 XNLI

Metric used: Accuracy.

langsmBERT (178M)xlm-roberta-base (279M)xlm-roberta-large (561M)mRoBERTa (283M)
bg69.3478.2682.1077.56
de71.5476.7581.6277.01
el66.5176.3781.4676.35
en82.2084.4587.9885.69
es74.8178.1883.6579.66
fr74.2578.2482.7179.16
ru68.5676.2179.1074.73
🔵 PAWS-X

Metric used: Accuracy.

langsmBERT (178M)xlm-roberta-base (279M)xlm-roberta-large (561M)mRoBERTa (283M)
de85.6586.9585.0587.35
en93.5093.9091.4594.75
es87.7589.3087.6588.60
fr86.6088.5587.3089.20

🟣 Structured Prediction: POS

🟣 POS (UDPOS)

Metric used: F1.

langsmBERT (178M)xlm-roberta-base (279M)xlm-roberta-large (561M)mRoBERTa (283M)
bg85.1488.6289.0688.19
de85.7188.4188.6588.58
el80.9287.1286.5587.03
en95.4395.7996.0795.85
es85.8588.1089.3187.45
et79.6886.2287.3686.25
eu60.1868.8371.8569.22
fi79.7285.9086.5484.23
fr81.2086.3488.2487.00
hu78.3983.0583.8482.96
it87.8688.9190.0189.11
lt78.5983.8684.9181.12
nl88.5989.1689.7089.31
pl80.3484.6185.7784.23
pt85.7787.5388.5687.18
ro76.5183.9986.4782.74
ru85.3688.7589.8389.09
uk80.6384.7985.8485.19
🟣 NER (PANX)

Metric used: F1.

langsmBERT (178M)xlm-roberta-base (279M)xlm-roberta-large (561M)mRoBERTa (283M)
bg78.3876.5281.9778.66
de78.8973.9278.5978.17
el74.0973.0775.4974.81
en84.6982.7084.5083.56
es72.3272.8373.4678.30
et77.5572.5678.3773.92
eu66.5258.3460.0156.74
fi78.1174.9878.4676.42
fr79.4577.0080.1676.94
hu77.3975.4880.1073.31
it81.3376.6880.6080.04
lt75.4873.7676.4172.71
nl82.4079.8082.9281.42
pl80.5777.1580.5580.26
pt79.6676.6080.9776.13
ro74.7371.7981.4266.85
ru65.4263.9370.6867.53
uk71.7166.7874.1271.69

⚪️ Sentence Retrieval

⚪️ BUCC2018

Metric used: F1.

langsmBERT (178M)xlm-roberta-base (279M)xlm-roberta-large (561M)mRoBERTa (283M)
de63.2666.8375.2386.09
fr62.6265.7969.2979.21
ru54.9770.1275.5782.93
⚪️ Tatoeba

Metric used: Accuracy.

langsmBERT (178M)xlm-roberta-base (279M)xlm-roberta-large (561M)mRoBERTa (283M)
bg48.8066.9071.6077.60
ca59.8057.3062.2080.20
de75.4088.4088.8095.60
el29.8051.6061.8072.30
es64.1071.0075.7089.70
et28.1044.2052.2061.80
eu25.5026.1035.8053.40
fi39.0063.9071.6063.90
fr64.3072.5073.7081.30
hu36.9058.7065.4062.40
it57.3064.7068.3080.30
lt31.1054.8059.6049.30
nl63.7076.8080.8086.60
pl50.1065.2075.9079.00
pt68.4076.6082.2088.80
ro51.5068.8075.7069.00
ru59.4069.8074.1081.60
uk52.6057.3069.1077.50

⚫ Question Answering

⚫ XQUAD

Metric used: F1.

langsmBERT (178M)xlm-roberta-base (279M)xlm-roberta-large (561M)mRoBERTa (283M)
de73.5574.8180.1573.92
el63.7473.3480.8673.56
en84.8484.2288.1382.70
es75.0676.4482.2177.07
ru72.0274.7380.1172.85
⚫ MLQA

Metric used: F1.

langsmBERT (178M)xlm-roberta-base (279M)xlm-roberta-large (561M)mRoBERTa (283M)
de57.6862.2068.7863.25
en80.1680.2783.5279.81
es64.9066.9772.9368.14
⚫ TyDiQA

Metric used: F1.

langsmBERT (178M)xlm-roberta-base (279M)xlm-roberta-large (561M)mRoBERTa (283M)
en68.2659.5771.3361.50
fi55.7051.9170.6252.32
ru53.7150.7564.4850.66

CLUB Benchmark

The Catalan Language Understanding Benchmark consists of 6 tasks: Named Entity Recognition (NER), Part-of-Speech Tagging (POS), Semantic Textual Similarity (STS), Text Classification (TC), Textual Entailment (TE), and Question Answering (QA). This benchmark evaluates the model's capabilities in the Catalan language.

This comparison also includes RoBERTa-ca, a model derived from mRoBERTa by applying vocabulary adaptation and performing continual pre-training on a 95GB Catalan-only corpus. For further details, visit here.

tasksroberta-base-bne (125M)berta (126M)mBERT (178M)xlm-roberta-base (279M)xlm-roberta-large (561M)roberta-ca (125M)mRoBERTa (283M)
ner (F1)87.5989.4785.8987.5089.4789.7088.33
pos (F1)98.6498.8998.7898.9199.0399.0098.98
sts (Person)74.2781.3977.0575.1183.4982.9979.52
tc (Acc.)73.8673.1672.0073.0574.1072.8172.41
te (Acc.)72.2780.1175.8678.2786.6382.1482.38
viquiquad (F1)82.5686.7487.4286.8190.3587.3187.86
xquad (F1)60.5667.3867.7268.5676.0870.5369.40

Galician Benchmark

To evaluate performance in Galician, the models are tested on two tasks highlighted in Bertinho's paper:

tasksbertinho (109M)roberta-base-bne (125M)mBERT (178M)xlm-roberta-base (279M)xlm-roberta-large (561M)mRoBERTa (283M)
ner-dataset-SLI NERC (F1)86.2786.8086.2285.9988.1087.75
pos-dataset-UD_GL_CTG (F1)97.5897.2797.5797.7797.9597.75

BasqueGLUE Benchmark

To assess the model performance in the Basque language, the BasqueGLUE Benchmark is used as the baseline. BasqueGLUE has been elaborated from previously existing datasets and following similar criteria to those used for the construction of GLUE and SuperGLUE. Some of the tasks has been slightly adapted to easily assess all the models (e.g, FMTODeu_slot is originally described as "Slot filling" task, but it is evaluated as NERC, since it follows the BIO annotation scheme).

tasksroberta-base-bne (125M)mBERT (178M)xlm-roberta-base (279M)xlm-roberta-large (561M)mRoBERTa (283M)
NERC - NERCid (F1)71.5379.9881.7483.9680.86
NERC - NERCood (F1)61.4776.9576.2380.2576.97
NERC - FMTODeu_slot (F1)72.7073.6573.8077.0977.32
Sentiment Analysis - BEC2016eu (Acc.)67.1367.0569.8967.9069.20
Topic Classification - BHTCv2 (Acc.)66.7270.1772.0175.7872.55
Intent Classification - FMTODeu_intent (Acc.)78.3878.0182.1583.3583.07
Stance Detection - VaxxStance (Acc.)58.0166.6761.2266.0365.71

Additional information

Author

The Language Technologies Lab from Barcelona Supercomputing Center.

Contact

For further information, please send an email to [email protected].

Copyright

Copyright(c) 2025 by Language Technologies Lab, Barcelona Supercomputing Center.

Funding

This work has been promoted and financed by the Ministerio para la Transformación Digital y de la Función Pública and Plan de Recuperación, Transformación y Resiliencia - Funded by EU – NextGenerationEU within the framework of the project Modelos del Lenguaje.

Acknowledgements

This project has benefited from the contributions of numerous teams and institutions through data contributions.

In Catalonia, many institutions have been involved in the project. Our thanks to Òmnium Cultural, Parlament de Catalunya, Institut d'Estudis Aranesos, Racó Català, Vilaweb, ACN, Nació Digital, El món and Aquí Berguedà.

At national level, we are especially grateful to our ILENIA project partners: CENID, HiTZ and CiTIUS for their participation. We also extend our genuine gratitude to the Spanish Senate and Congress, Fundación Dialnet, Fundación Elcano and the ‘Instituto Universitario de Sistemas Inteligentes y Aplicaciones Numéricas en Ingeniería (SIANI)’ of the University of Las Palmas de Gran Canaria.

At the international level, we thank the Welsh government, DFKI, Occiglot project, especially Malte Ostendorff, and The Common Crawl Foundation, especially Pedro Ortiz, for their collaboration.

Their valuable efforts have been instrumental in the development of this work.

Disclaimer

Be aware that the model may contain biases or other unintended distortions. When third parties deploy systems or provide services based on this model, or use the model themselves, they bear the responsibility for mitigating any associated risks and ensuring compliance with applicable regulations, including those governing the use of Artificial Intelligence.

The Barcelona Supercomputing Center, as the owner and creator of the model, shall not be held liable for any outcomes resulting from third-party use.

License

Apache License, Version 2.0

Downloads last month
172
Safetensors
Model size
283M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support