fix gym library link

This commit is contained in:
rram12
2022-09-18 19:45:13 +02:00
parent cd8615a52a
commit e17c505d42
5 changed files with 692 additions and 692 deletions

View File

@@ -3,8 +3,8 @@
{
"cell_type": "markdown",
"metadata": {
"id": "view-in-github",
"colab_type": "text"
"colab_type": "text",
"id": "view-in-github"
},
"source": [
"<a href=\"https://colab.research.google.com/github/huggingface/deep-rl-class/blob/main/unit1/unit1.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>"
@@ -24,7 +24,7 @@
"❓ If you have questions, please post them on #study-group-unit1 discord channel 👉 https://discord.gg/aYka4Yhff9\n",
"\n",
"\n",
"🎮 Environment: [LunarLander-v2](https://www.gymlibrary.ml/environments/box2d/lunar_lander/)\n",
"🎮 Environment: [LunarLander-v2](https://www.gymlibrary.dev/environments/box2d/lunar_lander/)\n",
"\n",
"📚 RL-Library: [Stable-Baselines3](https://stable-baselines3.readthedocs.io/en/master/)\n",
"\n",
@@ -395,8 +395,8 @@
},
"source": [
"### Step 4: Create the LunarLander environment 🌛 and understand how it works\n",
"#### [The environment 🎮](https://www.gymlibrary.ml/environments/box2d/lunar_lander/)\n",
"In this first tutorial, were going to train our agent, a [Lunar Lander](https://www.gymlibrary.ml/environments/box2d/lunar_lander/), **to land correctly on the moon**. To do that, the agent needs to learn **to adapt its speed and position(horizontal, vertical, and angular) to land correctly.**\n",
"#### [The environment 🎮](https://www.gymlibrary.dev/environments/box2d/lunar_lander/)\n",
"In this first tutorial, were going to train our agent, a [Lunar Lander](https://www.gymlibrary.dev/environments/box2d/lunar_lander/), **to land correctly on the moon**. To do that, the agent needs to learn **to adapt its speed and position(horizontal, vertical, and angular) to land correctly.**\n",
"\n",
"\n",
"---\n",
@@ -404,7 +404,7 @@
"\n",
"💡 A good habit when you start to use an environment is to check its documentation \n",
"\n",
"👉 https://www.gymlibrary.ml/environments/box2d/lunar_lander/\n",
"👉 https://www.gymlibrary.dev/environments/box2d/lunar_lander/\n",
"\n",
"---\n"
]
@@ -1068,23 +1068,23 @@
},
{
"cell_type": "markdown",
"source": [
"If you have a runtime bug `RuntimeError: imageio.ffmpeg.download() has been deprecated. Use 'pip install imageio-ffmpeg' instead.'` here please install this package and click on restart button after the installation"
],
"metadata": {
"id": "XAq_Fosen9Z7"
}
},
"source": [
"If you have a runtime bug `RuntimeError: imageio.ffmpeg.download() has been deprecated. Use 'pip install imageio-ffmpeg' instead.'` here please install this package and click on restart button after the installation"
]
},
{
"cell_type": "code",
"source": [
"!pip install imageio==2.4.1"
],
"execution_count": null,
"metadata": {
"id": "3wmUjb5LoCiK"
},
"execution_count": null,
"outputs": []
"outputs": [],
"source": [
"!pip install imageio==2.4.1"
]
},
{
"cell_type": "code",
@@ -1126,7 +1126,7 @@
"\n",
"**Compare the results of your LunarLander-v2 with your classmates** using the [leaderboard](https://huggingface.co/spaces/huggingface-projects/Deep-Reinforcement-Learning-Leaderboard) 🏆\n",
"\n",
"Is moon landing too boring to you? Try to **change the environment**, why not using CartPole-v1, MountainCar-v0 or CarRacing-v0? Check how they works [using the gym documentation](https://www.gymlibrary.ml/) and have fun 🎉."
"Is moon landing too boring to you? Try to **change the environment**, why not using CartPole-v1, MountainCar-v0 or CarRacing-v0? Check how they works [using the gym documentation](https://www.gymlibrary.dev/) and have fun 🎉."
]
},
{
@@ -1177,10 +1177,10 @@
"accelerator": "GPU",
"colab": {
"collapsed_sections": [],
"include_colab_link": true,
"name": "Copie de Unit 1: Train your first Deep Reinforcement Learning Agent 🚀.ipynb",
"private_outputs": true,
"provenance": [],
"include_colab_link": true
"provenance": []
},
"gpuClass": "standard",
"kernelspec": {
@@ -1193,4 +1193,4 @@
},
"nbformat": 4,
"nbformat_minor": 0
}
}

View File

@@ -4,8 +4,8 @@ In this Unit, we're going to dive deeper into one of the Reinforcement Learning
We'll also implement our **first RL agent from scratch**: a Q-Learning agent and will train it in two environments:
- [Frozen-Lake-v1 ⛄ (non-slippery version)](https://www.gymlibrary.ml/environments/toy_text/frozen_lake/): where our agent will need to go from the starting state (S) to the goal state (G) by walking only on frozen tiles (F) and avoiding holes (H).
- [An autonomous taxi 🚕](https://www.gymlibrary.ml/environments/toy_text/taxi/?highlight=taxi) will need to learn to navigate a city to transport its passengers from point A to point B.
- [Frozen-Lake-v1 ⛄ (non-slippery version)](https://www.gymlibrary.dev/environments/toy_text/frozen_lake/): where our agent will need to go from the starting state (S) to the goal state (G) by walking only on frozen tiles (F) and avoiding holes (H).
- [An autonomous taxi 🚕](https://www.gymlibrary.dev/environments/toy_text/taxi/?highlight=taxi) will need to learn to navigate a city to transport its passengers from point A to point B.
<img src="assets/img/envs.gif" alt="unit 2 environments"/>

View File

@@ -23,8 +23,8 @@
"❓ If you have questions, please post them on #study-group-unit2 discord channel 👉 https://discord.gg/aYka4Yhff9\n",
"\n",
"🎮 Environments: \n",
"- [FrozenLake-v1](https://www.gymlibrary.ml/environments/toy_text/frozen_lake/)\n",
"- [Taxi-v3](https://www.gymlibrary.ml/environments/toy_text/taxi/)\n",
"- [FrozenLake-v1](https://www.gymlibrary.dev/environments/toy_text/frozen_lake/)\n",
"- [Taxi-v3](https://www.gymlibrary.dev/environments/toy_text/taxi/)\n",
"\n",
"📚 RL-Library: Python and Numpy\n",
"\n",
@@ -276,12 +276,12 @@
"id": "NAvihuHdy9tw"
},
"source": [
"### Step 1: Create and understand [FrozenLake environment ⛄]((https://www.gymlibrary.ml/environments/toy_text/frozen_lake/)\n",
"### Step 1: Create and understand [FrozenLake environment ⛄]((https://www.gymlibrary.dev/environments/toy_text/frozen_lake/)\n",
"---\n",
"\n",
"💡 A good habit when you start to use an environment is to check its documentation \n",
"\n",
"👉 https://www.gymlibrary.ml/environments/toy_text/frozen_lake/\n",
"👉 https://www.gymlibrary.dev/environments/toy_text/frozen_lake/\n",
"\n",
"---\n",
"\n",
@@ -923,7 +923,7 @@
"source": [
"### Step 10: Evaluate our Q-Learning agent 📈\n",
"- Normally you should have mean reward of 1.0\n",
"- It's relatively easy since the state space is really small (16). What you can try to do is [to replace with the slippery version](https://www.gymlibrary.ml/environments/toy_text/frozen_lake/)."
"- It's relatively easy since the state space is really small (16). What you can try to do is [to replace with the slippery version](https://www.gymlibrary.dev/environments/toy_text/frozen_lake/)."
]
},
{
@@ -1321,12 +1321,12 @@
"source": [
"# Part 2: Taxi-v3 🚖\n",
"\n",
"### Step 1: Create and understand [Taxi-v3 🚕](https://www.gymlibrary.ml/environments/toy_text/taxi/)\n",
"### Step 1: Create and understand [Taxi-v3 🚕](https://www.gymlibrary.dev/environments/toy_text/taxi/)\n",
"---\n",
"\n",
"💡 A good habit when you start to use an environment is to check its documentation \n",
"\n",
"👉 https://www.gymlibrary.ml/environments/toy_text/taxi/\n",
"👉 https://www.gymlibrary.dev/environments/toy_text/taxi/\n",
"\n",
"---\n",
"\n",
@@ -1717,7 +1717,7 @@
"* Try different hyperparameters by looking at what your classmates have done.\n",
"* **Push your new trained model** on the Hub 🔥\n",
"\n",
"Are walking on ice and driving taxis too boring to you? Try to **change the environment**, why not using FrozenLake-v1 slippery version? Check how they work [using the gym documentation](https://www.gymlibrary.ml/) and have fun 🎉."
"Are walking on ice and driving taxis too boring to you? Try to **change the environment**, why not using FrozenLake-v1 slippery version? Check how they work [using the gym documentation](https://www.gymlibrary.dev/) and have fun 🎉."
]
},
{

File diff suppressed because one or more lines are too long

File diff suppressed because one or more lines are too long