Learn Before
  • What is BERT?

BERT's Contributions

  • This paper demonstrates the importance of bidirectional pre-training for language representations
  • Masked language model enables bidirectional pre-training, where previous models used unidirectional language models for pre-training
  • It shows that pre-trained representations reduce the need for many heavily-engineered task-specific architectures
  • BERT is the first fine-tuning based representation model that achieves state-of-the-art performance on a large suite of sentence-level and token-level tasks
  • BERT advances the state of the art for eleven NLP tasks

0

1

4 years ago

Tags

Data Science

Related
  • BERT’s Innovations

  • BERT's Contributions

  • BERT Experiments

  • Input representation

  • BERT&GPT and Fine Tuning