Autoencoder
Autoencoder
Autoencoders
• Autoencoders (AE) are a specific type of feedforward neural
network where the input is the same as the output.
Latent representation, h
https://www.edureka.co/blog/autoencoders-tutorial/
Denoising Autoencoders
• Keeping the code layer small forced the autoencoder to learn
an intelligent representation of the data.
https://ift6266h17.files.wordpress.com/2017/03/14_autoencoders.pdf
Sparse Autoencoder
• Sparse autoencoder learning algorithm automatically learns
features from unlabeled data.
• This method works even if the code size is large, since only a
small subset of the nodes will be active at any time.
• a j (2) denotes the activation of hidden unit j at hidden layer (i.e.
2n d) in the autoencoder.
Sparsity parameter
• The average activation of hidden unit j (averaged over the training
set).
• The result is a column vector with one row per hidden neuron.
Penalty Term
• To satisfy this constraint, the hidden unit’s activations must
mostly be near 0.
• To achieve this, we will add an extra penalty term to our
optimization objective that penalizes ˆρj deviating significantly
from ρ.
The third term which penalizes the absolute value of the vector
of activations a in layer h for sample i.
Cost function: KL Divergence
• Now compute:
Gradient Calculation
• To Compute ρˆi allow a forward pass on all the training
examples first i.e. the average activations on the training set,
before computing backpropagation on any example.
• When decoding from the latent state, we'll randomly sample from
each latent state distribution to generate a vector as input for our
decoder model.
varitational inference
• Suppose that there exists some hidden variable z which
generates an observation x.
• We can only see x, but we would like to infer the
characteristics of z i.e. compute p(z|x).