![]() A NVIDIA employee on stackoverflow suggested, I may be hitting a lengthy JIT compile step, because the GTX 1080 has compute capability of 6.1, which the wheel I used may not be compiled for. When I tested this with another wheel ( which is linked in this tutorial, I did not compile it myself.) on cuda 9.1/cudnn7.0.5, I had the same issues. ![]() The issue is still the same.ĬUDA works, since it prints the 'Hello, TensorFlow!', when I use the official test example, but before that it takes like 2minutes every time! To make sure, I pip-installed tensorflow-gpu into a fresh anaconda env. I have CUDA 9.0, so I downloaded CuDNN 7.0.5 for CUDA 9.0 and pasted the files to *C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v9.0*, overwriting the ones form cuDNN 7.1.2, which I tested earlier. Tensorflow site says, I should use CUDA® Toolkit 9.0 and cuDNN v7.0. I've been trying to find out, why this happens, and nothing really worked so far. ![]() I noticed that tensorflow always takes about ~2min before it actually starts to compute.
0 Comments
Leave a Reply. |