-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
llama cpp Quickstart #12894
base: main
Are you sure you want to change the base?
llama cpp Quickstart #12894
Conversation
rem under most circumstances, the following environment variable may improve performance, but sometimes this may also cause performance degradation | ||
set SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=1 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Move these two to Note
``` | ||
- For multi-GPUs user, go to Tips for how to select specific GPU. | ||
|
||
### Step 3: Running community GGUF models with IPEX-LLM |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Run GGUF models
|
||
### Step 3: Running community GGUF models with IPEX-LLM | ||
|
||
Here we provide a simple example to show how to run a community GGUF model with IPEX-LLM. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
how to run a standard GGUF model
Description
llama cpp Quickstart