Learn Before
What is BERT?
BERT's Contributions
- This paper demonstrates the importance of bidirectional pre-training for language representations
- Masked language model enables bidirectional pre-training, where previous models used unidirectional language models for pre-training
- It shows that pre-trained representations reduce the need for many heavily-engineered task-specific architectures
- BERT is the first fine-tuning based representation model that achieves state-of-the-art performance on a large suite of sentence-level and token-level tasks
- BERT advances the state of the art for eleven NLP tasks
0
1
4 years ago
Tags
Data Science
Related
BERT’s Innovations
BERT's Contributions
BERT Experiments
Input representation
BERT&GPT and Fine Tuning