-
Notifications
You must be signed in to change notification settings - Fork 43
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
README asset #5
Comments
https://github.com/jcwleo/random-network-distillation-pytorch/blob/master/config.conf I see last pull request is about normalization, maybe UseNorm = True improve reward_per_epi or speed of convergence? And what about UseNoisyNet, when it could better to use? |
@kslazarev |
@jcwleo I see the difference in x-axis scale in reward_per_epi and reward_per_rollout plots. |
Or the x-axis scale (global_update and sample_episode) depends on player survival/experience so on later updates x-axis scale will be the same? |
@kslazarev per_rollout and per_epi is not same scale. per_rollout means just one global update(enter agent.train_model()). but per_epi means Env’s one episode info that is one of parallel env. |
@jcwleo Yes, correct. I have another small questions about code. How could be appropriate to ask? Every question as new issue, or move forward to ask in this issue? |
@kslazarev I want you to create an issue for each question. :) |
Hello, can you tell me how many Gpus you used and how long it took you to see this effect? |
Hello. Not fast. Don't remember exactly, 1 or 2 NV 1080 Ti |
@kslazarev Excuse me, I use 1 3090,2 envs, run for more than 2 hours, the reward is still 0, is this normal? I didn't load the pre-training model |
It was 3 years ago. Could not help, I don't remember exactly what problem could cause. |
@kslazarev Ok, thanks |
@kslazarev |
The text was updated successfully, but these errors were encountered: