Pre-Training Strategies for the BERT Model
Learn about the different pre-training strategies used to train the BERT model.
Now that we've learned how to feed the input to BERT by converting it into embeddings and also how to tokenize the input using a WordPiece tokenizer, let's learn how to pre-train the BERT model.
Pre-training strategies
The BERT model is pre-trained on the following two tasks:
Masked language modeling
Next sentence prediction
Let's understand how the two ...