You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@mxnet.apache.org by GitBox <gi...@apache.org> on 2020/04/04 16:45:44 UTC
[GitHub] [incubator-mxnet] r2d3 commented on issue #17826: [quetions] Check
failed: e == cudaSuccess: CUDA: initialization error
r2d3 commented on issue #17826: [quetions] Check failed: e == cudaSuccess: CUDA: initialization error
URL: https://github.com/apache/incubator-mxnet/issues/17826#issuecomment-609056307
HI @rondogency and @Rainweic,
even with an "import mxnet" in each process, I got similar issue.
For example: https://github.com/aws-samples/parallelize-ml-inference
is creating `global model` in `init_worker`.
It does not do `import mxnet` in each process. I modified the code to do the import in `init_worker` but still gets the CUDA error
` File "/home/ubuntu/anaconda3/envs/mxnet_p36/lib/python3.6/site-packages/mxnet/symbol/symbol.py", line 1776, in simple_bind
ctypes.byref(exe_handle)))
File "/home/ubuntu/anaconda3/envs/mxnet_p36/lib/python3.6/site-packages/mxnet/base.py", line 255, in check_call
raise MXNetError(py_str(_LIB.MXGetLastError()))
mxnet.base.MXNetError: [16:40:00] src/engine/./../common/cuda_utils.h:379: Check failed: e == cudaSuccess || e == cudaErrorCudartUnloading: CUDA: initialization error`
https://github.com/aws-samples/parallelize-ml-inference/issues/3
Regards
David
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
users@infra.apache.org
With regards,
Apache Git Services