Skip to content

Releases: Eclectic-Sheep/sheeprl

v0.3.0

06 Sep 14:05
c51a985
Compare
Choose a tag to compare

v0.3.0 Release Notes

This new release introduces hydra as the default configuration manager. In particular it fixes #74 and automatically #75, since now the cnn_keys and mlp_keys can be specified separately for both the encoder and decoder.
The changes are mainly the following:

  • Dreamer-V3 initialization follows directly Hafner's implementation (adapted from https://github.com/NM512/dreamerv3-torch/blob/main/tools.py)
  • all args.py and the HFArgumentParser have been removed. Configs are now specified under the sheeprl/configs folder and hydra is the default configuration manager
  • Every environment wrapper is directly instantiated through the hydra.utils.instantiate inside the make_env or make_dict_env method: in this way one can easily modify the wrapper passing whatever parameters to customize the env. Every wrapper must take as input the id parameter, which must be specified in the relative config
  • Every optimizer is directly instantiated through the hydra.utils.instantiate and can be modified through the CLI on the experiment run
  • The howto/configs.md has been added in which explain how the configs are organized inside the repo

v0.2.2

11 Aug 15:38
f336e63
Compare
Choose a tag to compare

v0.2.2 Release Notes

  • Fixed Dreamer-V3 test function: it uses its own instead of the Dreamer-V2 ones
  • Added ruff in pre-commit and add pre-commit.ci

v0.2.1

10 Aug 16:41
025fc71
Compare
Choose a tag to compare

v0.2.1 Release Notes

  • Added Dreamer-V3 algorithm from https://arxiv.org/abs/2301.04104
  • Added RestartOnException wrapper, which recreates and restarts the environments whenever somethig bad has happened during the step or reset. This has been added only on Dreamer-V3 algorithm
  • Renamed classes and functions (in particular the Player classes fro both Dreamer-V1/V2)

v0.2

08 Aug 15:21
8f5d54f
Compare
Choose a tag to compare

v0.2 Release notes

  • Added DiambraWrapper
  • Added Multi-encoder/decoder to all the algorithms, but Droq, Sac and PPO Recurrent
  • Added Multi-discrete support to PPO, DreamerV1 and P2E-DV1
  • Modified the make_env function to be able to train the agents on environments that return both pixel-like and vector-like observations
  • Modified the ReplayBuffer class to handle multiple observations
  • Updated howtos
  • Fixed #66
  • Logger creation is moved to sheeprl.utils.logger
  • Env creation is moved to sheeprl.utils.env
  • PPO algo is now a single-folder algorithm (removed ppo_pixel and ppo_continuous folder)
  • sac_pixel has been renamed to sac_ae
  • Added support to gymnasium==0.29.0, mujoco>=2.3.3 and dm_control>=1.0.12

v0.1

08 Aug 13:39
4f7da89
Compare
Choose a tag to compare