From eed8941ff53988c455b4a60fdf42de765c229b8b Mon Sep 17 00:00:00 2001 From: PierreCounathe Date: Sun, 27 Aug 2023 23:07:54 +0200 Subject: [PATCH] Change action-value wording in glossary to match previous wording --- units/en/unit2/glossary.mdx | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/units/en/unit2/glossary.mdx b/units/en/unit2/glossary.mdx index b44d40f..f76ea52 100644 --- a/units/en/unit2/glossary.mdx +++ b/units/en/unit2/glossary.mdx @@ -11,7 +11,7 @@ This is a community-created glossary. Contributions are welcomed! ### Among the value-based methods, we can find two main strategies - **The state-value function.** For each state, the state-value function is the expected return if the agent starts in that state and follows the policy until the end. -- **The action-value function.** In contrast to the state-value function, the action-value calculates for each state and action pair the expected return if the agent starts in that state and takes an action. Then it follows the policy forever after. +- **The action-value function.** In contrast to the state-value function, the action-value calculates for each state and action pair the expected return if the agent starts in that state, takes that action, and then follows the policy forever after. ### Epsilon-greedy strategy: