Learn Before
Classification

Architectural Approaches to Self-Supervised Pre-training

Self-supervised pre-training can be examined through the lens of different neural network designs, specifically decoder-only, encoder-only, and encoder-decoder architectures. The focus of such analysis is often narrowed to the Transformer architecture, as it underpins the vast majority of modern pre-trained models in the field of Natural Language Processing.

Image 0

0

1

Updated 2026-04-15

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences