Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Batch normalization module #6

Open
JingleiSHI opened this issue Feb 2, 2018 · 1 comment
Open

Batch normalization module #6

JingleiSHI opened this issue Feb 2, 2018 · 1 comment

Comments

@JingleiSHI
Copy link

Hello,
Thank you very much for your work, that really helps me a lot. I have noticed that the batch normalization function didn't work very well in my models: with batch normalization it will be trained slower and converge at the wrong point. But we I substitute the 'gamma' and 'beta's <truncated_normal_initializer> with <zeros/ones_initializer>, your batch normalization module works very well and converges at the right point. As I am a tensorflow beginner, I don't understand well the differences it brings. Could you please explain to me why you choose these initializers and why they bring such a difference? Thank you very much!

Best wishes,
J. SHI

@tadax
Copy link
Owner

tadax commented Feb 4, 2018

@JingleiSHI Actually, I have no particular reason for using truncated normal distribution. I not sure why your model doesn't work (zeros/ones initializer are used in tf.layers.batch_normalization).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants