CleanRL (@cleanrl_lib) 's Twitter Profile
CleanRL

@cleanrl_lib

High-quality single file implementation of Deep Reinforcement Learning algorithms with research-friendly features (PPO, DQN, C51, DDPG, TD3, SAC, PPG)

ID: 1531787789817331712

linkhttp://cleanrl.dev calendar_today01-06-2022 00:01:07

3 Tweet

637 Followers

0 Following

Costa Huang (@vwxyzjn) 's Twitter Profile Photo

Thanks to @_joaogui1's awesome contribution ๐Ÿ™, CleanRL now has a TD3 + JAX implementation that is 2-4x faster than the TD3 + PyTorch equivalent ๐Ÿ”ฅ. Running on TPU is now possible, too ๐Ÿš€! ๐Ÿ“œ docs: docs.cleanrl.dev/rl-algorithms/โ€ฆ ๐Ÿ’พ code: github.com/vwxyzjn/cleanrโ€ฆ A short ๐Ÿงต1/x

Thanks to @_joaogui1's awesome contribution ๐Ÿ™, <a href="/cleanrl_lib/">CleanRL</a> now has a TD3 + JAX implementation that is 2-4x faster than the TD3 + <a href="/PyTorch/">PyTorch</a> equivalent ๐Ÿ”ฅ. Running on TPU is now possible, too ๐Ÿš€!

๐Ÿ“œ docs: docs.cleanrl.dev/rl-algorithms/โ€ฆ
๐Ÿ’พ code: github.com/vwxyzjn/cleanrโ€ฆ

A short ๐Ÿงต1/x
Chang Ye (@yooceii) 's Twitter Profile Photo

Happy to share that CleanRL now supports Random Network Distillation + envpool, it's 3ร— faster than our first version without envpool and still have comparable performance to the original implementation, say ๐Ÿ‘‹ to the long training time on hard-exploration games! Details๐Ÿ‘‡

Happy to share that <a href="/cleanrl_lib/">CleanRL</a> now supports Random Network Distillation + envpool, it's 3ร— faster than our first version without envpool and still have comparable performance to the original implementation, say ๐Ÿ‘‹ to the long training time on hard-exploration games! 
Details๐Ÿ‘‡