Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

VERY VERY Slow on the rtx 4050 and i5-12455 and 16 gb ram #1719

Closed
Asory2010 opened this issue Jun 6, 2023 · 6 comments
Closed

VERY VERY Slow on the rtx 4050 and i5-12455 and 16 gb ram #1719

Asory2010 opened this issue Jun 6, 2023 · 6 comments
Labels

Comments

@Asory2010
Copy link

I also have cublas enabled and i have tried both 13b and 7b models and it takes ages to even spell on token. I am using these parameters:

main -i --interactive-first -r "### Human:" --temp 0 -c 2048 -n -1 --reapeate_penalty 1.2 --instruct --color -m wizard-mega-13b.ggml.q4_0.bin

@ghost
Copy link

ghost commented Jun 6, 2023

Add the -t parameter to your prompt, perhaps -t 4.

You might try lowering the batch # for the model to begin responding quicker with -b 10 in your prompt.

@Asory2010
Copy link
Author

Add the -t parameter to your prompt, perhaps -t 4.

You might try lowering the batch # for the model to begin responding quicker with -b 10 in your prompt.

did not work): plus it crashes after a while in the loading procces

@ghost
Copy link

ghost commented Jun 6, 2023

Add the -t parameter to your prompt, perhaps -t 4.
You might try lowering the batch # for the model to begin responding quicker with -b 10 in your prompt.

did not work): plus it crashes after a while in the loading procces

If your token generation is extremely slow, then try -t 1 and work your way up from there. Here's more information, including GPU with cuBlas:

https://github.com/ggerganov/llama.cpp/blob/master/docs/token_generation_performance_tips.md

This is the limit of my knowledge on the subject, so if it continues to crash then I suggest someone else troubleshoot with @Asory2010

@gjmulder
Copy link
Collaborator

gjmulder commented Jun 7, 2023

Run top or atop to see how many threads are active on your CPU. As a rough rule of thumb you want to set -t to the number of physical cores on your CPU (usually half the number of hypercores the system reports).

Run nvidia-smi to see what is happening on your GPU. If your CPU isn't the bottleneck you should see 25-50% GPU utilisation after configuring -ngl.

EDIT: The Intel® Core™ i5-1245U Processor has 2 fast and 8 slow CPU cores. I'd try to set -t to 2, 4, 6, 8, 10 to see if the slow CPU cores actually help performance.

@Asory2010
Copy link
Author

Run top or atop to see how many threads are active on your CPU. As a rough rule of thumb you want to set -t to the number of physical cores on your CPU (usually half the number of hypercores the system reports).

Run nvidia-smi to see what is happening on your GPU. If your CPU isn't the bottleneck you should see 25-50% GPU utilisation after configuring -ngl.

EDIT: The Intel® Core™ i5-1245U Processor has 2 fast and 8 slow CPU cores. I'd try to set -t to 2, 4, 6, 8, 10 to see if the slow CPU cores actually help performance.

Quick Update after some testing the text gen became wayyyyy faster but the loading time still remained slow, why is that?

@github-actions github-actions bot added the stale label Mar 25, 2024
Copy link
Contributor

This issue was closed because it has been inactive for 14 days since being marked as stale.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants