Learn Before
Comparison

Divergent Pre-training Paradigms in NLP and Computer Vision

A key distinction in the historical development of pre-training lies in the different approaches adopted by the fields of computer vision and Natural Language Processing. The standard paradigm in computer vision involved supervised pre-training, where models were trained on large, manually labeled datasets such as ImageNet. In contrast, the breakthrough in modern NLP was driven by large-scale, self-supervised learning, which leverages vast quantities of unlabeled text.

0

1

Updated 2026-04-14

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences