Conversation

✌️ We have added a series of toolchain improvements. Notably, to ensure our code can run without obscure errors, we now use GitHub Action to run our algorithm variant on Windows, Linux, and macOS — things that "just work."
Image
1
3
🤗 What's coming next? We are adding evaluations and integration to save and load models. This will help support recent research efforts that reuse prior computation in RL to bootstrap new agents (e.g., reincarnate RL by ) twitter.com/agarwl_/status
Image
Quote Tweet
Image
tl;dr: What if we didn’t almost always train reinforcement learning (RL) agents from scratch for research? Our work argues for an alternative approach to research, where we build on prior computation, such as learned policies, network weights. arxiv.org/abs/2206.01626 (1/N)
Show this thread
1
11
🧠 We are also adding "RLops" features that help us seamlessly integrate new features and bugs.
Quote Tweet
Experimenting *RLops* at @cleanrl_lib — we will soon be able to compare the library's performance at different versions 🤩! Rough idea: tag experiments in @weights_biases with a PR number and build tools to generate analysis! PR: github.com/vwxyzjn/cleanr
Show this thread
Image
Image
1
3
🤩 Moreover, we plan to support 's Gymnasium in the future to replace the now no-longer-maintained openai/gym. Notably, Shimmy will help us support `dm_env` environments such as `dm_control`!
Quote Tweet
Today we're launching the Farama Foundation, a new nonprofit dedicated to open source reinforcement learning, and we're beginning by maintaining and standardizing all the major open source reinforcement learning environments. Read more here: farama.org/Announcing-The
1
3
Show replies