Learn Before
Concept

Deep Belief Networks (DBNs)

  • DBNs were one of the first nonconvolutional models to successfully admit training of deep architectures
  • They are generative models with several layers of latent/hidden variables in which the latent variables are typically binary, while the visible units may be binary or real.
  • Every unit in each layer is connected to every unit in each neighboring layer.
  • The probability distribution represented by the DBN is given by P(h(l),h(lāˆ’1))āˆexp(b(l)⊤h(l)+b(lāˆ’1)⊤h(lāˆ’1)+h(lāˆ’1)⊤W(l)h(l))P(h^{(l)}, h^{(l-1)}) \propto exp(b^{{(l)}^\top}h^{(l)} + b^{(l-1)\top}h^{(l-1)} + h^{(l-1)\top}W^{(l)}h^{(l)}) P(hi(k)=1∣h(k+1))=σ(bi(k)+W:,i(k+1)⊤h(k+1))āˆ€i,āˆ€k∈1,...,lāˆ’2P(h_{i}^{(k)}=1 | h^{(k+1)}) = \sigma (b_{i}^{(k)}+W_{:,i}^{(k+1)\top}h^{(k+1)})\forall i, \forall k \in 1, ..., l-2 P(vi=1∣h(1))=σ(bi(0)+W:,i(1)⊤h(1))āˆ€iP(v_i=1 | h^{(1)}) = \sigma (b_i^{(0)}+W_{:,i}^{(1)\top}h^{(1)})\forall i
  • In the case of real-valued visible units, we substitute v∼N(v;b(0)+W(1)⊤h(1),Ī²āˆ’1)\mathbf{v}\sim\mathcal{N}(v;b^{(0)}+W^{(1)\top}h^{(1)},\beta^{-1})
  • with β\beta diagonal for tractability
Image 0

0

1

Updated 2021-07-22

References


Tags

Data Science