On 05/07/2017 10:05, Andreas Beckmann wrote:
I had seen them, too, but haven't investigated them at all :-)
Should I upload the version we have right now? Is there an upstream
releasse announcement? What changed?

I found this [1]:

cuBLAS 8.0.61.2 New Features

cuBLAS Library

This update contains performance enhancements and bug-fixes to the cuBLAS library in CUDA Toolkit 8. Deep Learning applications based on Recurrent Neural Networks (RNNs) and Fully Connected Networks (FCNs) will benefit from new GEMM kernels and improved heuristics in this release.

This update supports the x86_64 architecture on Linux, Windows, and Mac OS operating systems, and the ppc64le architecture on Linux only.

The highlights of this update are as follows:
‣ Performance enhancements for GEMM matrices used in speech and natural
language processing
‣ Integration of OpenAI GEMM kernels
‣ Improved GEMM heuristics to select optimized algorithms for given input sizes
‣ Heuristic fixes for batched GEMMs
‣ GEMM performance bug fixes for Pascal and Kepler platforms

> solution for adding the patch will be to create more source tarballs
> (.orig-patch1-$arch.tar.gz) and unpack them, overriding what we unpacked
> first.

Is there any point to uploading 8.0.61, or should we just skip to 8.0.61.2? I haven't been able to do any testing yet, but I am at the point now where I can try building the SDK samples.


[1] http://docs.nvidia.com/cuda/pdf/CUDA_Toolkit_Release_Notes.pdf

Reply via email to