Concept

What is BERT?

B\textbf{B}idirectional E\textbf{E}ncoder R\textbf{R}epresentations from T\textbf{T}ransformers

  • A new language representation model that is designed to pre-train deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers
  • Beneficial because it is a generalist approach that allows for customization to specific tasks (question answering, language inference, etc) by adding just one additional layer, while other approaches might build entirely different models for different language-processing tasks

0

2

Updated 2026-04-17

Tags

Data Science

Foundations of Large Language Models

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related