Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bug in network_buider #241

Open
xuanyaoming opened this issue Jun 8, 2023 · 5 comments
Open

Bug in network_buider #241

xuanyaoming opened this issue Jun 8, 2023 · 5 comments

Comments

@xuanyaoming
Copy link

Hi! I think the continue should be replaced with pass here, because continue will skip all the remaining lines in the loop.
If my understanding is correct, dense_func refers to nn.Linear or the dense layer in tenseflow. Then in_size should be updated every time when the loop ends. That means line 109 shouldn't be skipped, otherwise in_size will always be the input size.

By the way, is the last element of units the output size?

Looking forward to your reply.

https://github.com/Denys88/rl_games/blob/811af5c19e9f7ebf7ff0725512bfc64e52439fe8/rl_games/algos_torch/network_builder.py#LL106C26-L106C26

@Denys88
Copy link
Owner

Denys88 commented Jun 23, 2023

I think its fine. all other lines a related to the case where we need norm.

@xuanyaoming
Copy link
Author

@Denys88
Copy link
Owner

Denys88 commented Jun 25, 2023

Thanks!
But I am surprised now why it works :)

@xuanyaoming
Copy link
Author

I guess there are two reasons. First, most people prefer to have the same unit size for all hidden layers. Actually, there is a paper claining doing so can reduce the effect of gradient descent in some situations. Secondly, I think most people (including me) have a habit to normalize internal variables. So the code I'm talking about probably has never been run by anyone yet.

@Denys88
Copy link
Owner

Denys88 commented Jun 26, 2023

@xuanyaoming I've got what happened. Btw in the most IsaacGym configs there are like 256-128-64 MLP without normalization.
need_norm is set to true by default and we set it to false only in the case when we want to norm only first layer (I saw this configuration in one of the papers):

                if not need_norm:
                    continue
                if norm_only_first_layer and norm_func_name is not None:
                   need_norm = False 

Thanks for the report I'll fix my code.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants