Today, we’re gonna learn how to create a virtual agent that discovers how to interact with the world. The technique we’re going to use is called Q-Learning, and it’s super cool.
Today, we will see how autoencoders can help us visualize the data in some very cool ways. For that, we will work on images, using the Convolutional Autoencoder architecture (CAE).
Autoencoders (AE) are a family of neural networks for which the input is the same as the output*. They work by compressing the input into a latent-spacerepresentation, and then reconstructing the output from this representation.
Have you ever found yourself eating, with no free hands to change the volume of your movie? Or the brightness of the screen? We’ll see how to use state-of-the-art artificial intelligence techniques to solve this problem by sending commands your computer with eye movements!