Tensorflow 2.1.0 error resolution: failed call to cuinit: CUDA_ ERROR_ NO_ DEVICE: no CUDA-capable device is detected

today in the use of keras-gpu in jupyter notebook error, at first did not pay attention to the console output, only from jupyter see error messages. So, check the solution, roughly divided into two, one version back, two specified running equipment. Because I felt that it was not the version problem, I still used the latest version without going back, so I tried the second method to solve the problem, that is, the program was designed to run the device, and the program could run without error. I accidentally noticed the console output and found that the CPU was running?Then found that the name of the specified device error, resulting in the system can not find the device, so first query the name of the device and then specify the device, to solve the problem.

native environment:

  • cudatoolkit = 10.1.243
  • cudnn = 7.6.5
  • tensorflow – gpu = 2.1.0
  • keras – gpu = 2.3.1

Jupyter notebook print error:

UnknownError:  Failed to get convolution algorithm. This is probably because cuDNN failed to initialize, so try looking to see if a warning log message was printed above.
	 [[node time_distributed_1/convolution (defined at C:\anaconda3\envs\keras\lib\site-packages\keras\backend\tensorflow_backend.py:3009) ]] [Op:__inference_keras_scratch_graph_1967]

Function call stack:

Jupyter notebook console print error: failed call to cuInit: CUDA_ERROR_NO_DEVICE: no cuda-capable device is detected


  1. to view native GPU/CPU information
from tensorflow.python.client import device_lib


[name: "/device:CPU:0"
 device_type: "CPU"
 memory_limit: 268435456
 locality {
 incarnation: 16677593686354176255,
 name: "/device:GPU:0"
 device_type: "GPU"
 memory_limit: 1440405913
 locality {
   bus_id: 1
   links {
  incarnation: 787265797177696422
 physical_device_desc: "device: 0, name: GeForce 940MX, pci bus id: 0000:01:00.0, compute capability: 5.0"]
  1. can see that the native device has GPU, so use the following statement to specify the name of the running device:
    The attention! = The following must be the device name, each person’s device name may be different, do not copy, must first use the command query, then specify!
import os
os.environ['CUDA_VISIBLE_DEVICES'] = '/device:GPU:0'
  1. to view the currently loaded device:
import tensorflow as tf
sess = tf.compat.v1.Session(config=tf.compat.v1.ConfigProto(log_device_placement=True))


Device mapping:
/job:localhost/replica:0/task:0/device:GPU:0 -> device: 0, name: GeForce 940MX, pci bus id: 0000:01:00.0, compute capability: 5.0
  1. console output similar to the following information explains that the code has been run on the GPU:

Read More: