-
Notifications
You must be signed in to change notification settings - Fork 18.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
cuDNN acceleration #1046
cuDNN acceleration #1046
Conversation
@shelhamer the slides are not public |
Thank you for pointing that out. The slides are now public. On Sunday, September 7, 2014, Sergio Guadarrama [email protected]
|
Awesome! |
It only published a lib, but no source code. On Mon, Sep 8, 2014 at 11:23 AM, NIU ZHIHENG [email protected]
Best wishes, |
Just for the record - when compiled with cudnn and no changes are made to the pre-cudnn protobuffers, the default behavior is to use the cudnn implementation of things. |
cool component |
@shelhamer it is a bit annoying to get so many warnings when Falling back to standard Caffe. I think it could say it during the Setup and then don't say it again.
See #1170 |
cuDNN acceleration
cuDNN acceleration
Caffe + cuDNN is an NVIDIA-Caffe collaboration for deep learning. cuDNN is an acceleration library for deep network operations with drop-in integration to Caffe. It is a free library downloadable with CUDA developer registration. It requires CUDA >= 6.5. This combination is the fastest public framework for deep learning in vision when benchmarked on the AlexNet / CaffeNet architectures with overall model speedups of 1.2-1.5x and layer-wise speedups of 1.2-3x over standard Caffe. Caffe + cuDNN lets you define your models just as before while taking advantage of these computational speedups.
In this first release cuDNN includes
These operations are drop-in accelerations of the Caffe layers. To switch on acceleration, set
in your
Makefile.config
during installation. Layers will be accelerated by default.NVIDIA and Caffe will coordinate future releases to further accelerate computation and introduce new features. NVIDIA has committed to tuning cuDNN to current and future GPU architectures.
Caffe is free and open-source and cuDNN is a CUDA developer library like cuBLAS and cuRAND.
Check out the cuDNN site, the Caffe's latest roast slides, and NVIDIA parallel forall blog announcement!
Thanks to the cuDNN team for this collaboration and special thanks to Cliff Woolley for his attention to detail.
Note on convolution: the cuDNN convolution aims to match or exceed the speed of Caffe's own matrix-multiplication approach while reducing memory usage. In many input and model regimes it accelerates the computation 1.3-3x and never requires buffers. In certain cases of fully-convolutional models or large inputs the Caffe convolution is slightly faster at the cost of more memory usage -- this is a direction for further optimization.
To pick the computational engine per-layer in your models, set the
engine: CAFFE
orengine: CUDNN
field in the{convolution,pooling,relu,sigmoid,tanh,softmax}_param
in your model definition: