From cf8ad134909511a5a8ac7868fbb619532b1bff0b Mon Sep 17 00:00:00 2001 From: niuyazhe Date: Mon, 22 Nov 2021 15:49:02 +0800 Subject: [PATCH] v0.2.1 --- CHANGELOG | 36 ++++++++++++++++++++++++++++++++++++ README.md | 4 ++-- ding/__init__.py | 2 +- 3 files changed, 39 insertions(+), 3 deletions(-) diff --git a/CHANGELOG b/CHANGELOG index 89dc49f..c2e6240 100644 --- a/CHANGELOG +++ b/CHANGELOG @@ -1,3 +1,39 @@ +2021.11.22(v0.2.1) +- env: gym-hybrid env (#86) +- env: gym-soccer (HFO) env (#94) +- env: Go-Bigger env baseline (#95) +- env: add the bipedalwalker config of sac and ppo (#121) +- algo: DQfD Imitation Learning algorithm (#48) (#98) +- algo: TD3BC offline RL algorithm (#88) +- algo: MBPO model-based RL algorithm (#113) +- algo: PADDPG hybrid action space algorithm (#109) +- algo: PDQN hybrid action space algorithm (#118) +- algo: fix R2D2 bugs and produce benchmark, add naive NGU (#40) +- algo: self-play training demo in slime_volley env (#23) +- algo: add example of GAIL entry + config for mujoco (#114) +- feature: enable arbitrary policy num in serial sample collector +- feautre: add torch DataParallel for single machine multi-GPU +- feature: add registry force_overwrite argument +- feature: add naive buffer periodic thruput seconds argument +- test: add pure docker setting test (#103) +- test: add unittest for dataset and evaluator (#107) +- test: add unittest for on-policy algorithm (#92) +- test: add unittest for ppo and td (MARL case) (#89) +- test: polish collector benchmark test +- fix: target model wrapper hard reset bug +- fix: fix learn state_dict target model bug +- fix: ppo bugs and update atari ppo offpolicy config (#108) +- fix: pyyaml version bug (#99) +- fix: small fix on bsuite environment (#117) +- fix: discrete cql unittest bug +- fix: release workflow bug +- fix: base policy model state_dict overlap bug +- fix: remove on_policy option in dizoo config and entry +- fix: remove torch in env +- style: gym version > 0.20.0 +- style: torch version >= 1.1.0, <= 1.10.0 +- style: ale-py == 0.7.0 + 2021.9.30(v0.2.0) - env: overcooked env (#20) - env: procgen env (#26) diff --git a/README.md b/README.md index af545de..1232aa8 100644 --- a/README.md +++ b/README.md @@ -31,7 +31,7 @@ [![Contributors](https://img.shields.io/github/contributors/opendilab/DI-engine)](https://github.com/opendilab/DI-engine/graphs/contributors) [![GitHub license](https://img.shields.io/github/license/opendilab/DI-engine)](https://github.com/opendilab/DI-engine/blob/master/LICENSE) -Updated on 2021.09.30 DI-engine-v0.2.0 (beta) +Updated on 2021.11.22 DI-engine-v0.2.1 (beta) ## Introduction to DI-engine (beta) @@ -52,7 +52,7 @@ DI-engine is a generalized Decision Intelligence engine. It supports most basic ### Other - [DI-engine-docs](https://github.com/opendilab/DI-engine-docs) - [treevalue](https://github.com/opendilab/treevalue) -- [DI-treetensor](https://github.com/opendilab/DI-treetensor) (preview) +- [DI-treetensor](https://github.com/opendilab/DI-treetensor) ## Installation diff --git a/ding/__init__.py b/ding/__init__.py index 065366d..5d8d43f 100644 --- a/ding/__init__.py +++ b/ding/__init__.py @@ -1,7 +1,7 @@ import os __TITLE__ = 'DI-engine' -__VERSION__ = 'v0.2.0' +__VERSION__ = 'v0.2.1' __DESCRIPTION__ = 'Decision AI Engine' __AUTHOR__ = "OpenDILab Contributors" __AUTHOR_EMAIL__ = "opendilab.contact@gmail.com" -- GitLab