FinBERT for Finnish
Learn about the FinBERT model, its configurations, and its performance.
We'll cover the following
FinBERT is the pre-trained BERT model for the Finnish language. FinBERT outperforms M-BERT on many downstream Finnish NLP tasks. We learned that M-BERT is trained using the Wikipedia text of 104 languages, but it comprises only 3% Finnish text. FinBERT is trained using the Finnish text from news articles, online discussions, and internet crawling. It consists of about 50K WordPiece vocabulary items, covering many Finnish words compared to M-BERT. This makes FinBERT better than M-BERT.
Configurations of FinBERT
The architecture of FinBERT is similar to the BERT-base model and has two configurations.
FinBERT-cased for cased text.
FinBERT-uncased for uncased text.
FinBERT is pre-trained on MLM and NSP tasks with a WWM scheme. We can download the pre-trained FinBERT model from GitHub. We can also use the pre-trained FinBERT model with the transformers
library, as shown here:
Get hands-on with 1400+ tech skills courses.