mirror of
https://github.com/huggingface/deep-rl-class.git
synced 2026-02-03 02:14:53 +08:00
Merge pull request #522 from AbhishekRana21/main
a small grammatical error fix
This commit is contained in:
@@ -787,7 +787,7 @@
|
||||
"id": "reBhoODwcXfr"
|
||||
},
|
||||
"source": [
|
||||
"- In my case, I got a mean reward is `200.20 +/- 20.80` after training for 1 million steps, which means that our lunar lander agent is ready to land on the moon 🌛🥳."
|
||||
"- In my case, I got a mean reward of `200.20 +/- 20.80` after training for 1 million steps, which means that our lunar lander agent is ready to land on the moon 🌛🥳."
|
||||
]
|
||||
},
|
||||
{
|
||||
|
||||
@@ -478,7 +478,7 @@ mean_reward, std_reward = evaluate_policy(model, eval_env, n_eval_episodes=10, d
|
||||
print(f"mean_reward={mean_reward:.2f} +/- {std_reward}")
|
||||
```
|
||||
|
||||
- In my case, I got a mean reward is `200.20 +/- 20.80` after training for 1 million steps, which means that our lunar lander agent is ready to land on the moon 🌛🥳.
|
||||
- In my case, I got a mean reward of `200.20 +/- 20.80` after training for 1 million steps, which means that our lunar lander agent is ready to land on the moon 🌛🥳.
|
||||
|
||||
## Publish our trained model on the Hub 🔥
|
||||
Now that we saw we got good results after the training, we can publish our trained model on the hub 🤗 with one line of code.
|
||||
|
||||
Reference in New Issue
Block a user