diff --git a/unit7/unit7.ipynb b/unit7/unit7.ipynb index 7301397..ad1f131 100644 --- a/unit7/unit7.ipynb +++ b/unit7/unit7.ipynb @@ -7,7 +7,7 @@ "provenance": [], "collapsed_sections": [], "private_outputs": true, - "authorship_tag": "ABX9TyNPB+iXGKgIWKts27HKZacW", + "authorship_tag": "ABX9TyNhsp4tLJxwa+BsFeBThLb7", "include_colab_link": true }, "kernelspec": { @@ -232,7 +232,7 @@ "source": [ "## Step 4: Create the PPO Model 🤖\n", "\n", - "PPO is one of the SOTA (state of the art) Deep Reinforcement Learning algorithms. If you don't know how it works, you can check this blogpost and the paper\n", + "PPO is one of the SOTA (state of the art) Deep Reinforcement Learning algorithms. \n", "\n", "In this case, because we have a vector as input, we'll use an MLP (multi-layer perceptron) as policy.\n", "\n",