Autoencoders are deep learning structures that learn a compressed latent space of the input data. In a sense, autoencoders generalize ๐๏ธ Principle Component Analysis with a ๐ธ๏ธ Multilayer Perceptron or ๐๏ธ Convolutional Neural Network; the hidden layer of the neural network is analogous to
Using the neural network, we still optimize weights to minimize
- Bottleneck, where the hidden layer has dimension
. - Input noise, making a de-noising autoencoder.
or regularization penalties. - Forcing orthogonality or independence.
Stacked Training
One potentially-more efficient method of training an autoencoder is the stacking method, which trains layers one at a time.
- For each hidden layer
, take the values from , then process it and feed it to . - Optimize
to reconstruct in ; in other words, weโre training to learn to reconstruct intermediate features in the neural network.