mirror of
https://github.com/huggingface/deep-rl-class.git
synced 2026-04-13 16:49:44 +08:00
Update train-our-robot.mdx
Adds the inference video to the train-our-robot section.
This commit is contained in:
@@ -45,10 +45,9 @@ en/unit13/onnx_inference_scene.jpg" alt="onnx inference scene"/>
|
||||
|
||||
**Press F6 to start the scene and let’s see what the agent has learned!**
|
||||
|
||||
<Tip>
|
||||
You can see a video of the trained agent in <a href="getting-started.mdx">getting started</a>.
|
||||
</Tip>
|
||||
Video of the trained agent:
|
||||
<video src="https://huggingface.co/datasets/huggingface-deep-rl-course/course-images/resolve/main/en/unit13/onnx_inference_test.mp4" type="video/mp4" controls autoplay loop mute />
|
||||
|
||||
It seems the agent is capable of collecting the key from both positions (left platform or right platform) and replicates the recorded behavior well. **If you’re getting similar results, well done, you’ve successfully completed this tutorial!** 🏆👏
|
||||
|
||||
If your results are different, note that the amount and quality of recorded demos can affect the results, and adjusting the number of steps for BC/GAIL stages as well as modifying the hyper-parameters in the Python script can potentially help. There’s also some run-to-run variation, so sometimes the results can be slightly different even with the same settings.
|
||||
If your results are significantly different, note that the amount and quality of recorded demos can affect the results, and adjusting the number of steps for BC/GAIL stages as well as modifying the hyper-parameters in the Python script can potentially help. There’s also some run-to-run variation, so sometimes the results can be slightly different even with the same settings.
|
||||
|
||||
Reference in New Issue
Block a user