From d65108046ac4de9ada43a8bcdc49480ae222f321 Mon Sep 17 00:00:00 2001 From: Vinay Kumar <12234745+imflash217@users.noreply.github.com> Date: Thu, 12 Jan 2023 14:35:45 -0500 Subject: [PATCH] Minor typo --- units/en/unit3/quiz.mdx | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/units/en/unit3/quiz.mdx b/units/en/unit3/quiz.mdx index 841ee4d..13f0295 100644 --- a/units/en/unit3/quiz.mdx +++ b/units/en/unit3/quiz.mdx @@ -76,8 +76,8 @@ For instance, in pong, our agent **will be unable to know the ball direction if **1. Make more efficient use of the experiences during the training** -Usually, in online reinforcement learning, the agent interacts in the environment, gets experiences (state, action, reward, and next state), learns from them (updates the neural network), and discards them. This is not efficient -But with experience replay, **we create a replay buffer that saves experience samples that we can reuse during the training**. +Usually, in online reinforcement learning, the agent interacts in the environment, gets experiences (state, action, reward, and next state), learns from them (updates the neural network), and discards them. This is not efficient. +But, with experience replay, **we create a replay buffer that saves experience samples that we can reuse during the training**. **2. Avoid forgetting previous experiences and reduce the correlation between experiences**