Merge pull request #575 from carschandler/patch-1

Confusing wording in self-play.mdx
This commit is contained in:
Thomas Simonini
2024-12-20 11:54:32 +01:00
committed by GitHub

View File

@@ -37,7 +37,7 @@ The theory behind self-play is not something new. It was already used by Arthur
Self-Play is integrated into the MLAgents library and is managed by multiple hyperparameters that were going to study. But the main focus, as explained in the documentation, is the **tradeoff between the skill level and generality of the final policy and the stability of learning**.
Training against a set of slowly changing or unchanging adversaries with low diversity **results in more stable training. But a risk to overfit if the change is too slow.**
Training against a set of slowly changing or unchanging adversaries with low diversity **results in more stable training. But there is a risk of overfitting if the change is too slow.**
So we need to control: