Concept

Token Alteration as an Input Corruption Method

Token alteration is a method for corrupting input sequences in denoising autoencoder training where some tokens are replaced with different, often incorrect, tokens from the vocabulary. This forces the model to learn robust representations that are not dependent on the exact original tokens.

0

1

Updated 2025-10-06

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences