Search⌘ K
AI Features

German BERT

Explore how to use the German BERT model trained on German Wikipedia and legal texts for NLP tasks such as classification and named entity recognition. Learn to load and apply the pre-trained model and tokenizer with Hugging Face's AutoModel and AutoTokenizer classes to process German language data effectively.

German BERT was developed by an organization called deepset.ai. They trained the BERT model from scratch using German text. The pre-trained German BERT model is open-sourced and free to use. German BERT is trained using recent German Wikipedia text, news articles, and data from OpenLegalData on a single Cloud TPU v2 for a period of 9 days.

German BERT is evaluated on many downstream tasks, ...