Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About the parameters sensitivity #2

Open
flexibility2 opened this issue May 22, 2019 · 1 comment
Open

About the parameters sensitivity #2

flexibility2 opened this issue May 22, 2019 · 1 comment

Comments

@flexibility2
Copy link

@HongtaoYang , I am very grateful for your source code! However, I have found that your implementation is very sensitive to the parameters of the network, such as :

  • In the batch_normalization layer, it must set the trainable=False, because when setting "trainable=True", the results will drop a lot.

  • In the conv2d layer, it must set the padding to vaild, and the results will also drop when setting the way of padding to "same".

So, I feel very strange and can't understand about this phenomenon, because I think the key of this algorithm is not the design of the network structure, it shouldn't be such sensitive to the parameters of the network, could you explain this phenomenon in detail? Thanks for your kindness!

@HongtaoYang
Copy link
Owner

@flexibility2 , it is correct. First the code is sensitive to hyper-parameters and that's why I failed to implement this for so long time until the author released their version, which provides a guideline. It is also correct that the core of the algorithm is not the design of network. However, as with many unsupervised learning algorithms, the network is sensitive to the initial conditions. The core of DAC depends on the assumption that objects of the same class should have more similar representations. If the network is initialised in a way that this assumption does not hold, then the self-reinforcing training process will diverge rapidly.

That is my personal understanding, I guess you can test if the authors' code is also sensitive to such conditions.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants