Update units/en/unit2/introduction.mdx

Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
This commit is contained in:
Thomas Simonini
2022-12-03 11:11:41 +01:00
committed by GitHub
parent 5a4cb88b1e
commit 10618b77ba

View File

@@ -15,7 +15,12 @@ We'll also **implement our first RL agent from scratch**: a Q-Learning agent an
<img src="https://huggingface.co/datasets/huggingface-deep-rl-course/course-images/resolve/main/en/unit3/envs.gif" alt="Environments"/>
We'll learn about the value-based methods and the difference between Monte Carlo and Temporal Difference Learning. And then, **we'll study and code our first RL algorithm**: Q-Learning, and implement our first RL Agent.
Concretely, we'll:
* learn about value-based methods
* learn about the differences between Monte Carlo and Temporal Difference Learning
* study and implement our first RL algorithm: Q-Learning
* implement our first RL agent
This unit is **fundamental if you want to be able to work on Deep Q-Learning**: the first Deep RL algorithm that was able to play Atari games and beat the human level on some of them (breakout, space invaders…).