Update colab (PPO link)

This commit is contained in:
Thomas Simonini
2022-07-16 17:04:37 +02:00
parent 70d7bef114
commit cc635a0713

View File

@@ -7,7 +7,7 @@
"provenance": [],
"collapsed_sections": [],
"private_outputs": true,
"authorship_tag": "ABX9TyNPB+iXGKgIWKts27HKZacW",
"authorship_tag": "ABX9TyNhsp4tLJxwa+BsFeBThLb7",
"include_colab_link": true
},
"kernelspec": {
@@ -232,7 +232,7 @@
"source": [
"## Step 4: Create the PPO Model 🤖\n",
"\n",
"PPO is one of the SOTA (state of the art) Deep Reinforcement Learning algorithms. If you don't know how it works, you can check this blogpost and the paper\n",
"PPO is one of the SOTA (state of the art) Deep Reinforcement Learning algorithms. \n",
"\n",
"In this case, because we have a vector as input, we'll use an MLP (multi-layer perceptron) as policy.\n",
"\n",