Chinese BERT

Learn about the Chinese BERT model, its training dataset, and pre-training, as well as explore the language technology platform along with its uses.

Along with M-BERT, Google Research has also open sourced the Chinese BERT model. The configuration of the Chinese BERT model is the same as the vanilla BERT-base model. It consists of 12 encoder layers, 12 attention heads, and 768 hidden units with 110 million parameters. The pre-trained Chinese BERT model can be downloaded from GitHub.

We can use the pre-trained Chinese BERT model with the transformers library, as shown here:

Get hands-on with 1200+ tech skills courses.