-
Notifications
You must be signed in to change notification settings - Fork 48
Issues: QwenLM/qwen.cpp
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
crash if compliing in debug mode, everything is ok if in release mode
#75
opened Jan 31, 2024 by
feixyz10
请问用qwen.cpp量化后的模型如何使用optimum-benchmark进行性能基准测试,现在参照readme中所述只得到一个build文件夹,不清楚如何进行下一步的测试
#72
opened Jan 18, 2024 by
suyu-zhang
[BUG] Qwen-1.8-Chat,用llama.cpp量化为f16,然后推理回答错乱,请问1.8在llama.cpp还不支持吗?
#69
opened Dec 26, 2023 by
Lyzin
2 tasks done
💡 [Question] - 您好,请教个问题,qwen-cpp BaseStreamer 如何通过std::string 构造一个 BaseStreamer?C++代码少一个构造方式
question
Further information is requested
#62
opened Dec 18, 2023 by
micronetboy
您好,请教个问题,qwen-cpp BaseStreamer 如何通过std::string 构造一个 BaseStreamer?C++代码少一个构造方式
#61
opened Dec 18, 2023 by
micronetboy
💡 [Question] - <title>qwen-cpp 只使用 cpu 和 启用 cpu BLAS 加速, 在都不使用GPU的情况下,速度有多大差别?我测试没有差别
question
Further information is requested
#63
opened Dec 15, 2023 by
micronetboy
💡 [Question] - QwenCPP Python Binding 如何 支持 BLAS CPU 加速
question
Further information is requested
#64
opened Dec 15, 2023 by
micronetboy
💡 [REQUEST] - CPU 的 qwen-cpp 如何封装为一个 http 服务?
question
Further information is requested
#65
opened Dec 14, 2023 by
micronetboy
CUDA error 2 at /home/qwen.cpp/third_party/ggml/src/ggml-cuda.cu:7196: out of memory
#55
opened Dec 8, 2023 by
youngallien
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.