From 3ff8c4e3c220a9e5f2ddf5e5cb8e3e8dce7b6c90 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?C=C3=A9sar=20Rodr=C3=ADguez?= Date: Sun, 5 Feb 2023 19:57:05 +0100 Subject: [PATCH] typo --- units/en/unit2/hands-on.mdx | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/units/en/unit2/hands-on.mdx b/units/en/unit2/hands-on.mdx index 5e4c164..473047b 100644 --- a/units/en/unit2/hands-on.mdx +++ b/units/en/unit2/hands-on.mdx @@ -369,7 +369,7 @@ def epsilon_greedy_policy(Qtable, state, epsilon): ``` ## Define the hyperparameters ⚙️ -The exploration related hyperparamters are some of the most important ones. +The exploration related hyperparameters are some of the most important ones. - We need to make sure that our agent **explores enough of the state space** to learn a good value approximation. To do that, we need to have progressive decay of the epsilon. - If you decrease epsilon too fast (too high decay_rate), **you take the risk that your agent will be stuck**, since your agent didn't explore enough of the state space and hence can't solve the problem.