Deep Q-Networks (DQN)
Deep Q-Networks (DQN) represent the fusion of Reinforcement Learning and Deep Neural Networks. While standard Q-Learning uses a table to store values, DQN uses a Neural Network to approximate the Q-value function.
This advancement allowed RL agents to handle environments with high-dimensional state spaces, such as raw pixels from a video game screen.
1. Why Deep Learning for Q-Learning?
In a complex environment, the number of possible states is astronomical.
- Atari 2600: A pixel screen with 128 colors has more possible states than there are atoms in the universe.
- The Solution: Instead of a table, we use a Neural Network () that takes a State as input and outputs the predicted Q-values for all possible actions.
2. The Two "Secret Ingredients" of DQN
Standard neural networks struggle with RL because the data is highly correlated (sequential frames in a game are nearly identical). To fix this, DQN introduced two revolutionary concepts:
A. Experience Replay
Instead of learning from the current experience immediately, the agent saves its experiences