-
Notifications
You must be signed in to change notification settings - Fork 2k
Closed
Labels
questionFurther information is requestedFurther information is requested
Description
Greetings! I'm a PyTorch RL fan but previously used baselines and stable baselines for research. I notice stable-baselines3 through the origin stable-baselines issue.
Recently there are many PyTorch RL platforms that emerged, including rlpyt, tianshou, etc. I went through their code and compared with stable-baselines3.
| Features | Stable-Baselines3 | rlpyt | tianshou |
|---|---|---|---|
| State of the art RL methods | ✔️ | ✔️ | ✔️ |
| Documentation | ✔️ | ✔️ | ✔️ |
| Custom environments | ✔️ | Just so-so | ✔️ |
| Custom policies | ✔️ | ✔️ | ✔️ |
| Common interface | ✔️ | ✔️ | ✔️ |
| Ipython / Notebook friendly | ✔️ | ✔️ | ✔️ |
| PEP8 code style | ✔️ | ✔️ | ✔️ |
| Custom callback | ✔️ | ❌ | ❌ |
| High code coverage | ✔️ | ❌ | ✔️ |
| Type hints | ✔️ | ❌ | ✔️ |
And for the planned features of stable-baselines3:
| Features | Stable-Baselines3 | rlpyt | tianshou |
|---|---|---|---|
| Tensorboard support | ✔️ | ✔️ | ✔️ |
| DQN extensions | ➖ QR-DQN in SB3 contrib | ✔️ | ✔️ |
| Support for Dict observation spaces | ✔️ | ✔️ | ✔️ |
| Recurrent Policies | ✔️ in contrib | ✔️ | ✔️ |
| TRPO | ✔️ in contrib | ❌ | ✔️ |
Also, the most important feature "modularization", from my perspective, tianshou is the best of all, rlpyt is the second. I hate OpenAI Baselines at this point, but stable-baselines is much better than openai.
Just some of my concerns.
guiambros, rgalljamov, ChenDRAG and chufanchen
Metadata
Metadata
Assignees
Labels
questionFurther information is requestedFurther information is requested