mmBERT: ModernBERT goes Multilingual

This blog post introduces mmBERT, a state-of-the-art massively multilingual encoder model trained on 3T+ tokens of text in over 1800 languages. It shows significant performance and speed improvements over previous multilingual models, being the first to improve upon XLM-R, while also developing new strategies for effectively learning low-resource languages. mmBERT builds upon ModernBERT for a blazingly fast architecture, and adds novel components to enable efficient multilingual learning.

If you are interested in trying out the models yourself, some example boilerplate is available at the end of this blogpost!

 

To finish reading, please visit source site