-
Notifications
You must be signed in to change notification settings - Fork 1.3k
Issues: intel-analytics/ipex-llm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Fundamental issue in response with IPEX-LLM with Ollama
user issue
#12636
opened Dec 30, 2024 by
anandnandagiri
[LNL][npu_llm] LNL npu doesn't support codegeex4-all-9b and codegeex4-all-9b model
#12633
opened Dec 29, 2024 by
johnysh
Docker Documentation incorrect. Benchmarks are all not working. Testing needed
user issue
#12629
opened Dec 27, 2024 by
TimoGoetze
PI_ERROR_BUILD_PROGRAM_FAILURE error when running Ollama using ipex-llm on 12450H CPU
#12597
opened Dec 23, 2024 by
qadzhang
WSL / Docker ipex-llm-inference-cpp-xpu:latest SIGSEGV on model load
#12592
opened Dec 21, 2024 by
vladislavdonchev
[BMG dgfx][ipex-llm[cpp]] low performance and gpu using when running llama.cpp inference on B580
user issue
#12586
opened Dec 20, 2024 by
jianjungu
RuntimeError: XPU out of memory on WSL2 vLLM running Qwen2.5-7B-Instruct, sym_int4, Arc A770
user issue
#12584
opened Dec 19, 2024 by
nkt-dk
Add Model Accuracy Benchmark support for large Models/Multi-dGPUs
multi-arc
user issue
#12572
opened Dec 18, 2024 by
RobinJing
Question: What's the recommended way to finetune newish models?
#12571
opened Dec 18, 2024 by
0xDEADFED5
Frequent Crashes with Ollama on ARC770 and i7-14700K
user issue
#12555
opened Dec 16, 2024 by
sirlegendary
IndexError while conducting GPT-2 Large benchmark for 1k input tokens.
user issue
#12551
opened Dec 16, 2024 by
JoAnnHang
GPU Runner crash in Ollama when offloading multiple layers
user issue
#12513
opened Dec 8, 2024 by
pauleseifert
[torch 2.3 + bigdl-core-xe-23] AttributeError: module 'xe_linear' has no attribute 'forward_qkv'
user issue
#12506
opened Dec 5, 2024 by
Nuullll
How can I use NPU to run Ollama on Intel Ultra7 155H chip in a laptop?
user issue
#12504
opened Dec 5, 2024 by
Muzixin
LLaVA-Video-7B-Qwen2 int4 quantization enabling on ARC
user issue
#12482
opened Dec 3, 2024 by
zhangcong2019
Using bf16 for inference on a CPU is slower than using float32.
user issue
#12472
opened Dec 2, 2024 by
fousdfrf
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.