Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

cmake : Resolve quantized model issue when CUBLAS enabled #1667

Merged
merged 1 commit into from
Dec 21, 2023

Conversation

bobqianic
Copy link
Collaborator

Thanks for the help provided by @slaren, the problem is finally solved!
See #1661 for details

@bobqianic bobqianic linked an issue Dec 20, 2023 that may be closed by this pull request
@Sing303
Copy link

Sing303 commented Dec 21, 2023

Something similar has been done here #1499

@ggerganov ggerganov merged commit 8986690 into ggerganov:master Dec 21, 2023
37 checks passed
@bobqianic bobqianic deleted the fix-quantized branch December 21, 2023 14:16
bygreencn added a commit to bygreencn/whisper.cpp that referenced this pull request Dec 25, 2023
* ggerganov/master:
  whisper : Replace WHISPER_PRINT_DEBUG with WHISPER_LOG_DEBUG (ggerganov#1681)
  sync : ggml (ggml_scale, ggml_row_size, etc.) (ggerganov#1677)
  docker :  Dockerize whisper.cpp (ggerganov#1674)
  CI : Add coverage for talk-llama when WHISPER_CUBLAS=1 (ggerganov#1672)
  examples : Revert CMakeLists.txt for talk-llama (ggerganov#1669)
  cmake : set default CUDA architectures (ggerganov#1667)
viktor-silakov pushed a commit to viktor-silakov/whisper_node_mic.cpp that referenced this pull request May 11, 2024
iThalay pushed a commit to iThalay/whisper.cpp that referenced this pull request Sep 23, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Quantized model is not working properly when CUBLAS is ON
3 participants