Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Convolutions with cudaconv2 to reduce memory consumption #852

Closed
hosang opened this issue Aug 4, 2014 · 3 comments
Closed

Convolutions with cudaconv2 to reduce memory consumption #852

hosang opened this issue Aug 4, 2014 · 3 comments

Comments

@hosang
Copy link

hosang commented Aug 4, 2014

I want to run a caffe with small stride on big images and ran into memory issues. I tried out PR #520, but even on quite small images (480x640) and an okay sized model (params and blobs take about 2GB on that image) the Consumption on CPU is at ~12GB. I assume the difference is in colbuffers since that's where I get the out of memory error. My understanding of FFT based convolutions is, that it also won't solve my memory problems.

What do you think about adding a convolution implementation that doesn't use additional memory? cuda-convnet [1] seems to be quite fast judging from the benchmark at [2]. The convolution code doesn't exactly look simple, so it doesn't look like a no-brainer to me to add it into caffe. Does it make any sense?

[1] https://code.google.com/p/cuda-convnet/
[2] https://github.com/soumith/convnet-benchmarks

@hosang hosang changed the title Convolutions with cudaconv2 to reduce memory consumtion Convolutions with cudaconv2 to reduce memory consumption Aug 4, 2014
@Yangqing
Copy link
Member

Yangqing commented Aug 5, 2014

It is a little non-trivial, and cuda-convnet actually uses a different order. We are exploring alternate approaches which may achieve the same (or better) goal, so incorporating cuda-convnet may not be on our radar (at least for now).

@shelhamer
Copy link
Member

Closing as duplicate of #830 to focus the conversation now that the memory aspect has been noted there. While we expect our alternative approach to address memory usage and speed, you are welcome to try integrating cuda-convnet2 convolution for comparison.

@hosang
Copy link
Author

hosang commented Aug 18, 2014

We are exploring alternate approaches which may achieve the same (or better) goal

@Yangqing: What approaches do you mean? I didn't find anything like that in the bugtracker/mailing list.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants