You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I met an issue related to the OpenCL targeting with mxnet model on TVM processing:
This is message error:
[14:32:18] src/nnvm/legacy_json_util.cc:190: Loading symbol saved by previous version v0.12.0. Attempting to upgrade...
[14:32:18] src/nnvm/legacy_json_util.cc:198: Symbol successfully upgraded!
model compiled.
[14:32:23] src/runtime/opencl/opencl_device_api.cc:235: Initialize OpenCL platform 'NVIDIA CUDA '
[14:32:24] src/runtime/opencl/opencl_device_api.cc:260: opencl(0)='GeForce GTX 750 Ti ' cl_device_id=0x83dde60
Traceback (most recent call last):
File "/home/abc/work/code.py", line 81, in
m.run()
File "/usr/local/lib/python3.5/dist-packages/tvm-0.4.0-py3.5-linux-x86_64.egg/tvm/contrib/graph_runtime.py", line 113, in run
self._run()
File "/usr/local/lib/python3.5/dist-packages/tvm-0.4.0-py3.5-linux-x86_64.egg/tvm/_ffi/_ctypes/function.py", line 183, in call
ctypes.byref(ret_val), ctypes.byref(ret_tcode)))
File "/usr/local/lib/python3.5/dist-packages/tvm-0.4.0-py3.5-linux-x86_64.egg/tvm/_ffi/base.py", line 66, in check_call
raise TVMError(py_str(_LIB.TVMGetLastError()))
tvm._ffi.base.TVMError: [14:32:26] src/runtime/module_util.cc:52: Check failed: ret == 0 (-1 vs. 0) [14:32:26] src/runtime/opencl/opencl_module.cc:141: OpenCL build error for device=0x83dde60ptxas error : Entry function 'fuse_conv2d_relu_9__kernel2' uses too much shared data (0x1ca64 bytes, 0xc000 max)
ptxas error : Entry function 'fuse_conv2d_relu_7__kernel2' uses too much shared data (0xda44 bytes, 0xc000 max)
I also encounter this problem when I run one complex model. So I want to answer here, because I don't find you ask this question in discuss.tvm.ai. I have pointed out what's wrong of OpenCL kernel code. My wrong OpenCL kernel code is:
I think your model should also have similar OpenCL code. The problem is 33800. NV only allows shared data's size 48K and Intel's GPU allows 64K. You can call clinfo command to check it out.
You can refer this: #525 When I tune as the suggestion, I can reduce the array size below 32K. I think it can also help you.
@FrozenGene thanks for your information. But i think the fuse function here use too much memory (~111kb) so we need to reduce array size as ur suggestion. Can you share the way you tune your model in detail?
Dear contributors,
I met an issue related to the OpenCL targeting with mxnet model on TVM processing:
This is message error:
Have you got any idea to resolve this trouble?
Thanks a lot
The text was updated successfully, but these errors were encountered: