diff --git a/units/en/unit7/hands-on.mdx b/units/en/unit7/hands-on.mdx index 84856f7..c65fb12 100644 --- a/units/en/unit7/hands-on.mdx +++ b/units/en/unit7/hands-on.mdx @@ -26,7 +26,7 @@ More precisely, AI vs. AI is three tools: In addition to these three tools, your classmate cyllum created a 🤗 SoccerTwos Challenge Analytics where you can check the detailed match results of a model: [https://huggingface.co/spaces/cyllum/soccertwos-analytics](https://huggingface.co/spaces/cyllum/soccertwos-analytics) -We're going to write a blog post to explain this AI vs. AI tool in detail, but to give you the big picture it works this way: +We're [wrote a blog post to explain this AI vs. AI tool in detail](https://huggingface.co/blog/aivsai), but to give you the big picture it works this way: - Every four hours, our algorithm **fetches all the available models for a given environment (in our case ML-Agents-SoccerTwos).** - It creates a **queue of matches with the matchmaking algorithm.** @@ -46,8 +46,6 @@ In order for your model to get correctly evaluated against others you need to fo What will make the difference during this challenge are **the hyperparameters you choose**. -The AI vs AI algorithm will run until April the 30th, 2023. - We're constantly trying to improve our tutorials, so **if you find some issues in this notebook**, please [open an issue on the GitHub Repo](https://github.com/huggingface/deep-rl-class/issues). ### Chat with your classmates, share advice and ask questions on Discord @@ -57,10 +55,6 @@ We're constantly trying to improve our tutorials, so **if you find some issues ## Step 0: Install MLAgents and download the correct executable -⚠ We're going to use an experimental version of ML-Agents which allows you to push and load your models to/from the Hub. **You need to install the same version.** - -⚠ ⚠ ⚠ We’re not going to use the same version from Unit 5: Introduction to ML-Agents ⚠ ⚠ ⚠ - We advise you to use [conda](https://docs.conda.io/en/latest/) as a package manager and create a new environment. With conda, we create a new environment called rl with **Python 3.9**: @@ -70,10 +64,10 @@ conda create --name rl python=3.9 conda activate rl ``` -To be able to train our agents correctly and push to the Hub, we need to install an experimental version of ML-Agents (the branch aivsai from Hugging Face ML-Agents fork) +To be able to train our agents correctly and push to the Hub, we need to install ML-Agents ```bash -git clone --branch aivsai https://github.com/huggingface/ml-agents +git clone https://github.com/Unity-Technologies/ml-agents ``` When the cloning is done (it takes 2.63 GB), we go inside the repository and install the package @@ -165,7 +159,6 @@ This allows each agent to **make decisions based only on what it perceives local - The solution then is to use Self-Play with an MA-POCA trainer (called poca). The poca trainer will help us to train cooperative behavior and self-play to win against an opponent team. If you want to dive deeper into this MA-POCA algorithm, you need to read the paper they published [here](https://arxiv.org/pdf/2111.05992.pdf) and the sources we put on the additional readings section.