Autoencoder
Short Definition
Full Definition
Autoencoders are fundamental building blocks of unsupervised deep learning, designed to learn efficient data representations without labeled examples. The architecture forces the network to learn the most important features of the data by compressing it through a bottleneck layer (the latent space) that has fewer dimensions than the input. The encoder transforms high-dimensional input into a low-dimensional latent representation, and the decoder attempts to reconstruct the original input from this representation. The network is trained by minimizing the reconstruction error between input and output. Because the bottleneck is smaller than the input, the autoencoder must learn to capture only the most essential information, effectively performing dimensionality reduction. Different variants serve different purposes. Sparse autoencoders add a sparsity constraint to learn more meaningful features. Denoising autoencoders learn to remove noise from corrupted inputs. Variational autoencoders (VAEs) learn a probabilistic latent space suitable for generating new data. Contractive autoencoders learn representations robust to small input changes. Autoencoders find applications in anomaly detection (detecting inputs that reconstruct poorly), data denoising, feature extraction, image compression, and as components of larger generative models. They serve as an important conceptual bridge between simple neural networks and more complex generative models.
Technical Explanation
The autoencoder minimizes reconstruction loss: L = ||x – decode(encode(x))||^2. The encoder maps input to latent space: z = f_encoder(x) where z has dimension d much less than input dimension D. The decoder reconstructs: x_hat = f_decoder(z). Variational autoencoders add a KL divergence term: L_VAE = E[||x – x_hat||^2] + beta * KL(q(z|x) || p(z)), where q(z|x) is the encoder distribution and p(z) is typically N(0,I). The reparameterization trick enables backpropagation: z = mu + sigma * epsilon, where epsilon ~ N(0,I). Sparse autoencoders add: L_sparse = L_recon + lambda * sum(KL(rho || rho_hat)), where rho is the desired sparsity level.
Use Cases
Advantages
Disadvantages
Schema Type
Featured Snippet Candidate
Difficulty Level