-
Notifications
You must be signed in to change notification settings - Fork 5.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG/Help] Windows 下 INT-4 量化模型无法加载 #162
Comments
不知道这个对开发人员有没有帮助 设置如下: 启动如下:
|
请确保已安装gcc和openmp。
编译时加入-O3会极大的加速,但在某些平台上可能造成错误,请根据情况自行添加优化参数。
推测错误原因有可能是没有openmp无法加载并行的kernel,也有可能是路径过于复杂ctypes没有正确处理。 |
非常感谢你的帮助,这边更换到Linux平台可以使用了 |
No compiled kernel found.
(2)然后在原先模型加载后手动加载一下手动编译的kernel
还是会报编译错误,但是已经可以使用了。 |
报编译错误,但可以使用这是什么情况 |
@sgb25sgb 使用默认的加载cpu kernel 失败,但是model = model.quantize(bits=4, kernel_file="Your Kernel Path")加载的cpu kernel 成功了 |
fxb392 谢谢你! |
试下这个: |
Is there an existing issue for this?
Current Behavior
Expected Behavior
No response
Steps To Reproduce
在 Windows 下加载 INT-4 量化模型,显示 CPU kernal 编译成功,但无法加载已编译的 kernal。经检查 quantization_kernels_parallel.so 成功编译,且使用
os.path.exists()
检测文件也返回True
。在 WSL 下一切正常。
Environment
Anything else?
No response
The text was updated successfully, but these errors were encountered: