🧐Variational Autoencoder (VAE)

Maximize Likelihood and KL divergence

Maximum Likelihood = Minimize KL Divergence

Lower Bound

How to generatePθ(xz)P_{\theta}(x|z)👍

VAE lower Bound of logPθ(x)\log P_{\theta}(x)

ELBO

Maximizing the likelihood of the observed xx:

  • P(z)P(\bold{z}): a normal distribution

  • P(xz)=N(x;μ(z),σ(z))P(x|\bold{z}) = N(x; \mu(\bold{z}), \sigma(\bold{z})), μ(z),σ(z)\mu(\bold{z}), \sigma(\bold{z}) is unknown and going to be estimated.

Loss: L=ilogP(x(i))L = \sum_i\log P(x^{(i)})

It is straightforward to figure out the following:

Then the lower bound LbL_b could be derived as follows:

  • q(zx)q(\bold{z}|x) is also a normal distribution, which is estimated by a neural network. q(zx)=N(z;μ(x),σ(x))q(\bold{z}|x)=N(\bold{z}; \mu'(x), \sigma'(x)). In other words, the mean and variance of z\bold{z} are given by two functions μ()\mu'(\cdot) and σ()\sigma'(\cdot), which will be estimated by the ouput of a neural network.

  • maximizing this lower bound needs to minimize KL[q(zx)P(z)]KL[q(\bold{z}|x) || P(\bold{z})] and maximize the second term, which will be connected with neural networks.

Minimize the first term KL[q(zx)P(z)]KL[q(\bold{z}|x) || P(\bold{z})]

Maximize the second term

VAE and GMM

Problem in VAE

Last updated