Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Algorithm detail implementation #1

Open
lqhdehub opened this issue Aug 15, 2024 · 2 comments
Open

Algorithm detail implementation #1

lqhdehub opened this issue Aug 15, 2024 · 2 comments

Comments

@lqhdehub
Copy link

Hello, I want to ask if this code is written based on ppo code in stable_baselines3? Are all the algorithmic details included? Thank you for your contributions.

@SlimShadys
Copy link
Owner

Hey! Yes, this code is based on the PPO implementation from Stable Baselines 3. It's a simplified version due to the removal of all the library-specific lines of code. It just shows how PPO truly works (without all the Abstract Classes which SB3 contains, all the various links etc).

It's just missing the Frame-Environment support such as CarRacing (in fact it's in the ToDo list), which processes the game one (or multiple) frames at a time.

Thank you.

@lqhdehub
Copy link
Author

Thank you for your reply.
What you mean is that this code basically contains all the technical details of ppo in SB3 that allow for better control (not abstract classes, all sorts of links).

However, I see that your code seems to have only discrete actions, but no continuous actions, I would like to ask if in the continuous action space, the scope of the action space is inconsistent (divided into two kinds), can the single agent ppo control the object well, or is there any way to make the control effect better?
Thank you very much. Have a nice day

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants