You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Because the rank of the first selfplay worker is 0, parallel environments are being initialized with the same seed, which might reduce training data diversity.
We could go for a simple fix like replacing self.rank by (self.rank + 1), however this is still problematic if considering multiple workers, as there will be seed overlap between them anyway.
A good option might be to sample a seed for each parallel environment using numpy (which is seeded before launching data workers). For instance:
envs = [self.config.new_game(np.random.randint(10**9)) for i in range(env_nums)]
The text was updated successfully, but these errors were encountered:
I might have found an unexpected behavior in how parallel training environments are being seeded.
I am referring to this line:
EfficientZero/core/selfplay_worker.py
Line 112 in c533ebf
Because the rank of the first selfplay worker is 0, parallel environments are being initialized with the same seed, which might reduce training data diversity.
We could go for a simple fix like replacing
self.rank
by(self.rank + 1)
, however this is still problematic if considering multiple workers, as there will be seed overlap between them anyway.A good option might be to sample a seed for each parallel environment using numpy (which is seeded before launching data workers). For instance:
envs = [self.config.new_game(np.random.randint(10**9)) for i in range(env_nums)]
The text was updated successfully, but these errors were encountered: