You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When training small models with low dimensional state spaces, the replay buffer seems to be a bottleneck in DQN. Sampling and updating the buffer for multiple indexes can somewhat easily be vectorized using numpy and can lead to significant speedups of these functions for large batch sizes.
I have implemented the improvement here and wanted to create this issue before starting my pull request, in order to get feedback.
The text was updated successfully, but these errors were encountered:
When training small models with low dimensional state spaces, the replay buffer seems to be a bottleneck in DQN. Sampling and updating the buffer for multiple indexes can somewhat easily be vectorized using numpy and can lead to significant speedups of these functions for large batch sizes.
I have implemented the improvement here and wanted to create this issue before starting my pull request, in order to get feedback.
The text was updated successfully, but these errors were encountered: