Autoencoders for Representation Learning
1. What is the primary objective of an autoencoder?
2. Which component of an autoencoder compresses the input into a lower-dimensional representation?
3. What is the bottleneck layer in an autoencoder?
4. In a denoising autoencoder, what is used as input during training?
5. Autoencoders are increasingly described in modern deep learning literature as which type of learning, since the supervision signal is derived from the input itself?
6. Which loss function measures the pixel-wise squared difference between the input and the reconstruction?
7. What is the size of each image in the Fashion-MNIST dataset?
8. What advantage does a 2-D latent space provide in autoencoders?
9. What happens if the bottleneck dimension is too small in an autoencoder?
10. What is the main advantage of using autoencoders over supervised learning methods for feature extraction?