After installing TensorFlow and its dependencies on a g2.2xlarge EC2 instance I tried to run an MNIST example from the getting started page:
python tensorflow/models/image/mnist/convolutional.py
But I get the following warning:
I tensorflow/core/common_runtime/gpu/gpu_device.cc:611] Ignoring gpu device
(device: 0, name: GRID K520, pci bus id: 0000:00:03.0) with Cuda compute
capability 3.0. The minimum required Cuda capability is 3.5.
Is this a hard requirement? Any chance I could comment that check out in a fork of TensorFlow? It would be super nice to be able to train models in AWS.
There is a section in the official installation page that guides you to enable Cuda 3, but you need to build Tensorflow from source.
$ TF_UNOFFICIAL_SETTING=1 ./configure
# Same as the official settings above
WARNING: You are configuring unofficial settings in TensorFlow. Because some
external libraries are not backward compatible, these settings are largely
untested and unsupported.
Please specify a list of comma-separated Cuda compute capabilities you want to
build with. You can find the compute capability of your device at:
https://developer.nvidia.com/cuda-gpus.
Please note that each additional compute capability significantly increases
your build time and binary size. [Default is: "3.5,5.2"]: 3.0
Setting up Cuda include
Setting up Cuda lib64
Setting up Cuda bin
Setting up Cuda nvvm
Configuration finished
Currently only GPUs with compute capability >= 3.5 are officially supported. However, GitHub user @infojunkie has offered a patch that makes it possible to use TensorFlow with a GPU with compute capability 3.0.
The official fix is in development. Meanwhile, check out the discussion on the GitHub issue for adding this support.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With