Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

llama : switch KQ multiplication to use F32 precision by default #10015

Merged
merged 1 commit into from
Oct 27, 2024

Conversation

ggerganov
Copy link
Owner

@ggerganov ggerganov commented Oct 23, 2024

ref #10005, #9991 (comment)

The list of models that require higher floating point range in the attention keeps growing, so to be on the safe side, default to F32 for the KQ multiplication.

Nexesenex added a commit to Nexesenex/croco.cpp that referenced this pull request Oct 24, 2024
@ggerganov ggerganov merged commit 8841ce3 into master Oct 27, 2024
60 checks passed
@ggerganov ggerganov deleted the gg/default-kq-f32-prec branch October 27, 2024 19:00
@LostRuins
Copy link
Collaborator

Hi, just forwarding some findings from community members - apparently this change to using F32 precision causes a significant speed regression (-30% T/s) on HIPBLAS. I personally use a Nvidia GPU and there's been no issues there for me.

They are using ROCM rx 6800xt and 5800x3d, 7900 xtx (driver 24.9.1)
models compared: llama 2 7b chat q8_0 and Mistral Small Instruct Q6_k

Regression was found by bisecting commits until arriving at this one. Not sure if anything can be done, but just thought I'd highlight it to see if anyone else using AMD GPUs has observed similar issues.

Also tagging @YellowRoseCx as they use AMD and can assist in testing/verifying if needed.

LostRuins added a commit to LostRuins/koboldcpp that referenced this pull request Nov 4, 2024
@ggerganov ggerganov mentioned this pull request Nov 4, 2024
2 tasks
arthw pushed a commit to arthw/llama.cpp that referenced this pull request Nov 15, 2024
arthw pushed a commit to arthw/llama.cpp that referenced this pull request Nov 18, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Bug: Unexpected output from Granite 3.0 MoE 1b when all layers on NVIDIA GPU
2 participants