Concept

Distance Metric Inductive Bias

In nonparametric methods such as the kk-nearest neighbor algorithm, a distance function dd (or equivalently, a vector-valued basis function ϕ(x)\phi(\mathbf{x})) must be specified to measure similarity between data points. This choice of distance metric is critical because it encodes the model's inductive bias. Even if any metric allows a model like 11-nearest neighbor to achieve zero training error, different distance functions represent different underlying assumptions about the data patterns. Consequently, with finite data, these varying inductive biases will yield different predictors, and their generalization performance will depend on how compatible the chosen metric is with the true data distribution.

0

1

Updated 2026-05-07

Contributors are:

Who are from:

Tags

D2L

Dive into Deep Learning @ D2L