mirror of
https://github.com/huggingface/deep-rl-class.git
synced 2026-04-13 16:49:44 +08:00
Update units/en/unit2/introduction.mdx
Co-authored-by: Sayak Paul <spsayakpaul@gmail.com>
This commit is contained in:
@@ -15,7 +15,12 @@ We'll also **implement our first RL agent from scratch**: a Q-Learning agent an
|
||||
<img src="https://huggingface.co/datasets/huggingface-deep-rl-course/course-images/resolve/main/en/unit3/envs.gif" alt="Environments"/>
|
||||
|
||||
|
||||
We'll learn about the value-based methods and the difference between Monte Carlo and Temporal Difference Learning. And then, **we'll study and code our first RL algorithm**: Q-Learning, and implement our first RL Agent.
|
||||
Concretely, we'll:
|
||||
|
||||
* learn about value-based methods
|
||||
* learn about the differences between Monte Carlo and Temporal Difference Learning
|
||||
* study and implement our first RL algorithm: Q-Learning
|
||||
* implement our first RL agent
|
||||
|
||||
This unit is **fundamental if you want to be able to work on Deep Q-Learning**: the first Deep RL algorithm that was able to play Atari games and beat the human level on some of them (breakout, space invaders…).
|
||||
|
||||
|
||||
Reference in New Issue
Block a user