Learn Before
Concept

Arbitrary Distribution Shift

When data distributions shift between training and testing in arbitrary, unconstrained ways, learning a robust classifier is fundamentally impossible. For instance, in a binary classification task like distinguishing cats from dogs, if the input distribution pS(x)=pT(x)p_S(\mathbf{x}) = p_T(\mathbf{x}) remains exactly the same but all labels are deterministically flipped such that pS(yx)=1pT(yx)p_S(y \mid \mathbf{x}) = 1 - p_T(y \mid \mathbf{x}), an algorithm cannot distinguish this pathological scenario from one where the distribution never changed at all.

0

1

Updated 2026-05-03

Contributors are:

Who are from:

Tags

D2L

Dive into Deep Learning @ D2L