Concept

Using Bayes' theorem in LDA

According to Bayes' theorem, we have

P(Y=kX=x)=P(Y=k)P(X=xY=k)l=1KP(Y=l)P(X=xY=l)=πkfk(x)l=1Kπlfl(x).P\left(Y=k | X=x\right)=\frac{P\left(Y=k\right) P\left(X=x | Y=k\right)}{\sum_{l=1}^{K} P\left(Y=l\right) P\left(X=x | Y=l\right)}\\ =\frac{\pi_{k} f_{k}(x)}{\sum_{l=1}^{K} \pi_{l} f_{l}(x)}.

In LDA with the assumption of Gaussian distribution in every class, we have the Gaussian density

fk(x)=1(2πk)p/2Σ1/2exp(12(xμk)TΣ1(xμk))f_k(x)=\frac{1}{(2 \pi_k)^{p / 2}|\boldsymbol{\Sigma}|^{1 / 2}} \exp \left(-\frac{1}{2}(x-\mu_k)^{T} \boldsymbol{\Sigma}^{-1}(x-\mu_k)\right)

for a pp-dimensional random vector XX with the distribution N(μk,Σ)N(\mu_k, \mathbf{\Sigma}). All we need to do is to plug fk(x)f_k(x) into Bayes' theorem above and find out the kk which maximize P(Y=kX=x)P\left(Y=k | X=x\right). LDA is a Bayes classifier with the assumption of Gaussian distribution in every class.

0

3

Updated 2020-03-07

Tags

Data Science