Chinese BERT
Learn about the Chinese BERT model, its training dataset, and pre-training, as well as explore the language technology platform along with its uses.
We'll cover the following
Along with M-BERT, Google Research has also open sourced the Chinese BERT model. The configuration of the Chinese BERT model is the same as the vanilla BERT-base model. It consists of 12 encoder layers, 12 attention heads, and 768 hidden units with 110 million parameters. The pre-trained Chinese BERT model can be downloaded from GitHub.
We can use the pre-trained Chinese BERT model with the transformers
library, as shown here:
Get hands-on with 1400+ tech skills courses.