Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Memory #7

Open
WangZhouTao opened this issue Dec 17, 2017 · 10 comments
Open

Memory #7

WangZhouTao opened this issue Dec 17, 2017 · 10 comments

Comments

@WangZhouTao
Copy link

hello, I am a student.recently,I am reading the HED paper. I found that your code is very interesting,but some error is happened when I try to train the model with my own data.Error message show that Memory error.My computer in-build 32GB memory, would you give me some suggestion.Think you very much.

@CangHaiQingYue
Copy link

my GPU only has 6G memory which is fine to run the code. dont know your setting

@WangZhouTao
Copy link
Author

Sorry,I forgot to tell you.I use 1080ti GPU and 32GB memory.I can run your code with my own data.But Memory error is happened when the iteration is about 13000.I obverse top command in ubuntu, i find Memory is insufficient(not gpu memory, is computer memory).If you could tell me some suggestion,I really appreciate what you’ve done。

@CangHaiQingYue
Copy link

Sorry, I have not had such error. This code will save a checkpoint each 100 steps, maybe 13000 is too big in your dataset( I trained at 30000 steps without memory error).

@WangZhouTao
Copy link
Author

I already changed to save a checkpoint with each 1000 steps, but the error is still happen.Did you run this code on windows? Thank you.

@CangHaiQingYue
Copy link

I run it on Ubuntu 16.04.

@sandhawalia
Copy link
Member

Could you please give details of your data-set ? It might be the case that your image sizes are quite large and since the VGG base model is full convolutional the intermediate representations overflow they GPU memory.

@wutachiang
Copy link

I also encountered the same problem with you, my server configuration is 8 Tesla p20, memory is 512G. Even with this configuration, memory error occurs after approximately 13,000 iterations during training. Will you solve the problem?@CangHaiQingYue

@CangHaiQingYue
Copy link

@Jasontachiangwu Well, when I cancelled the 'summary_write' operator, this problem was gone.
So I guess there maybe some bug in 'summary'. You may rebuild your own code.

@wutachiang
Copy link

@CangHaiQingYue,After I updated the tensorflow version to r1.8, no problem was found after training. It may be that there is a bug in the summary.

@yuezhilanyi
Copy link

yuezhilanyi commented Jun 26, 2018

with same code on windows 10 and ubuntu 16.04, tensorflow v1.4.0, 20000 iterations:
memory error occurs in ubuntu 16.04
while no error occurs in windows 10

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants