Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Stuck at Local Minimum in PPO with CarRacing-v2 Environment #87

Open
bantu-4879 opened this issue May 15, 2024 · 0 comments
Open

Stuck at Local Minimum in PPO with CarRacing-v2 Environment #87

bantu-4879 opened this issue May 15, 2024 · 0 comments

Comments

@bantu-4879
Copy link

I've been experimenting with various parameters in the Proximal Policy Optimization (PPO) algorithm within the CarRacing-v2 environment. After extensive testing, I've found a combination of parameters that initially shows promising results and learns relatively fast. However, I've encountered a significant challenge where the learning process appears to stagnate after a certain training stage.

Despite extensive training, the agent seems unable to surpass a particular performance threshold. I suspect that the algorithm may be trapped in a local minimum, but it doesn't seem to be a desirable or acceptable minimum given the potential of the environment.

Request for Assistance:
I'm seeking guidance on how to overcome this challenge and help the algorithm escape from the local minimum it's currently stuck in. Any insights, suggestions, or alternative approaches would be greatly appreciated. @simoninithomas

Environment and Configuration:

  • Environment: CarRacing-v2
  • Algorithm: Proximal Policy Optimization (PPO)

My Work
https://github.com/bantu-4879/Atari_Games-Deep_Reinforcement_Learning/tree/main/Notebooks/CarRacing-v2

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant