Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix experimental normalize reward wrapper #277

Merged

Conversation

raphajaner
Copy link
Contributor

Description

Should resolve #271 in a minimal way with simple tests.

Type of change

  • Bug fix (non-breaking change which fixes an issue)

Checklist:

  • I have run the pre-commit checks with pre-commit run --all-files (see CONTRIBUTING.md instructions to set it up)
  • I have commented my code, particularly in hard-to-understand areas
  • I have made corresponding changes to the documentation
  • My changes generate no new warnings
  • I have added tests that prove my fix is effective or that my feature works
  • New and existing unit tests pass locally with my changes

@pseudo-rnd-thoughts
Copy link
Member

We are changing the VectorEnv to not inherit from Env due to their differences in step data. Therefore, we will have different wrappers for Env and VectorEnv so could you remove the vector code

@raphajaner
Copy link
Contributor Author

Okay I see, then it's actually enough to change just a single line:

self.discounted_reward: np.array = np.array([0.0])

which has been a float before.

@pseudo-rnd-thoughts pseudo-rnd-thoughts changed the title Normalize reward vec dim Fix experimental normalize reward wrapper Jan 20, 2023
@pseudo-rnd-thoughts pseudo-rnd-thoughts merged commit 4b5abb6 into Farama-Foundation:main Jan 20, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

[Bug Report] Experimental NormalizeRewardV0 may be broken
2 participants