Skip to content
This repository has been archived by the owner on Mar 31, 2019. It is now read-only.

Original DQN Example #103

Open
ehknight opened this issue Aug 8, 2017 · 1 comment
Open

Original DQN Example #103

ehknight opened this issue Aug 8, 2017 · 1 comment
Assignees

Comments

@ehknight
Copy link

ehknight commented Aug 8, 2017

It would be really nice if we could have an example that re-implements the original DQN paper exactly. The Ms. PacMan one is pretty close but AFAIK it has some subtle differences, such as window augmentation (let me know if I'm wrong on this and I'll close the issue). This would be really helpful so that people can see how the AgentNet syntax relates to other methods that implement the exact same thing.

@justheuristic
Copy link
Collaborator

You're perfectly correct, current implementation is " just some q-learning on a convolutional neural network with experience replay and target networks".
I'll add an implementation of original DQN or prioritized experience replay one within next two weeks.

@justheuristic justheuristic self-assigned this Aug 9, 2017
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants