Search⌘ K

Language-Specific BERT

Learn how to apply monolingual BERT models for specific languages, including French FlauBERT, Spanish BETO, and more. Understand tokenization, embeddings, and fine-tuning to enhance language-specific NLP performance.

The M-BERT model is used in many different languages. However, instead of having a single M-BERT model for many languages, can we train a monolingual BERT for a specific target language? We can, and that is precisely what we will learn in this lesson.

M-BERT for various languages

Below are several interesting and popular monolingual BERT models for various languages:

  • FlauBERT for French

  • BETO for Spanish

  • BERTje for Dutch

  • German BERT

  • Chinese BERT

  • Japanese BERT

  • FinBERT for Finnish

  • UmBERTo for Italian

  • BERTimbay for Portuguese

  • RuBERT for Russian ...

FlauBERT for French