Summary: Different BERT Variants
Let’s summarize what we have learned so far.
We'll cover the following
Key highlights
Summarized below are the main highlights of what we've learned in this chapter.
We learned how the ALBERT works. We learned that ALBERT is a lite version of BERT, and it uses two interesting parameter reduction techniques, called cross-layer parameter sharing and factorized embedding parameterization. We also learned about the SOP task used in ALBERT. We learned that SOP is a binary classification task where the goal of the model is to classify whether the given sentence pair is swapped or not.
We looked into the RoBERTa model. We learned that the RoBERTa is a variant of BERT and it uses only the MLM task for training. Unlike BERT, it uses dynamic masking instead of static masking, and it is trained with a large batch size. It uses BBPE as a tokenizer, and it has a vocabulary size of 50,000.
Get hands-on with 1400+ tech skills courses.