mmBERT: How Johns Hopkins Built a 1,833-Language AI That Outperforms XLM-R mmBERT: How Johns Hopkins Built a 1,833-Language AI That Outperforms XLM-R Imagine trying to build an AI system that truly understands human language not just in English, but in over 7,000 languages s... AI research annealed language learning cross-lingual digital inclusion encoder-only FlashAttention Gemma tokenizer GLUE inverse masking language model mmBERT ModernBERT MTEB multilingual NLP XLM-R XTREME