CUDNN_STATUS_INTERNAL_ERROR/ResourceExhaustedErro

运行pycharm时,遇到如下错误:
tensorflow/stream_executor/cuda/cuda_dnn.cc:503] could not create cudnn handle: CUDNN_STATUS_INTERNAL_ERROR
2018-08-06 12:38:55.926801: F tensorflow/core/kernels/conv_ops.cc:713] Check failed: stream->parent()->GetConvolveAlgorithms( conv_parameters.ShouldIncludeWinogradNonfusedAlgo(), &algorithms)
或者
ResourceExhaustedError (see above for traceback): OOM when allocating tensor with shape[36,64,41,41] and type float on /job:localhost/replica:0/task:0/device:GPU:0 by allocator GPU_0_bfc
[[Node: Conv2D = Conv2D[T=DT_FLOAT, data_format=”NCHW”, dilations=[1, 1, 1, 1], padding=”SAME”, strides=[1, 1, 1, 1], use_cudnn_on_gpu=true, _device=”/job:localhost/replica:0/task:0/device:GPU:0”](Conv2D-0-TransposeNHWCToNCHW-LayoutOptimizer, w_start/read)]]
Hint: If you want to see a list of allocated tensors when OOM happens, add report_tensor_allocations_upon_oom to RunOptions for current allocation info.

 [[Node: add_19/_7 = _Recv[client_terminated=false, recv_device="/job:localhost/replica:0/task:0/device:CPU:0", send_device="/job:localhost/replica:0/task:0/device:GPU:0", send_device_incarnation=1, tensor_name="edge_246_add_19", tensor_type=DT_FLOAT, _device="/job:localhost/replica:0/task:0/device:CPU:0"]()]]

Hint: If you want to see a list of allocated tensors when OOM happens, add report_tensor_allocations_upon_oom to RunOptions for current allocation info.
解决方式:查看Nvidia x server settings窗口下的gpu0-(Geforce gtx 1080 ti)下的usd dedicated menmory ,此时若占用96%,则关机重启,即可解决问题。

猜你喜欢

转载自blog.csdn.net/u014727716/article/details/81452373