-
Notifications
You must be signed in to change notification settings - Fork 3.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add q8_0
models to download-ggml-model.sh
#2589
Merged
Merged
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
ggerganov
approved these changes
Nov 28, 2024
bygreencn
added a commit
to bygreencn/whisper.cpp
that referenced
this pull request
Dec 3, 2024
* ggerganov/master: (447 commits) ruby : Add low-level methods to transcribe (ggerganov#2585) models : add `q8_0` models to `download-ggml-model.sh` (ggerganov#2589) ruby : Follow source tree change (ggerganov#2580) whisper : use backend registry (#0) ggml/sched : do not skip views in pre-assignments whisper : adapt to new ggml (wip) talk-llama : sync llama.cpp sync : ggml ggml : sync resolve (skip) (#0) Add required ggml-base and backend libs to cmake pkg (llama/10407) cuda : fix CUDA_FLAGS not being applied (llama/10403) sycl : Add option to set the SYCL architecture for all targets (llama/10266) vulkan: Optimize soft_max (llama/10301) sycl: Revert MUL_MAT_OP support changes (llama/10385) cuda : only use native when supported by cmake (llama/10389) vulkan: remove use of null initializer (llama/10372) metal : fox offset integer overflows in im2col (ggml/1015) Vulkan: Fix device info output format specifiers (llama/10366) metal : add `GGML_UNARY_OP_ELU` kernel (ggml/1018) CUDA: fix MMV kernel being used for FP16 src1 (llama/10357) ...
bygreencn
added a commit
to bygreencn/whisper.cpp
that referenced
this pull request
Dec 5, 2024
* ggerganov/master: (44 commits) ruby : Add low-level methods to transcribe (ggerganov#2585) models : add `q8_0` models to `download-ggml-model.sh` (ggerganov#2589) ruby : Follow source tree change (ggerganov#2580) whisper : use backend registry (#0) ggml/sched : do not skip views in pre-assignments whisper : adapt to new ggml (wip) talk-llama : sync llama.cpp sync : ggml ggml : sync resolve (skip) (#0) Add required ggml-base and backend libs to cmake pkg (llama/10407) cuda : fix CUDA_FLAGS not being applied (llama/10403) sycl : Add option to set the SYCL architecture for all targets (llama/10266) vulkan: Optimize soft_max (llama/10301) sycl: Revert MUL_MAT_OP support changes (llama/10385) cuda : only use native when supported by cmake (llama/10389) vulkan: remove use of null initializer (llama/10372) metal : fox offset integer overflows in im2col (ggml/1015) Vulkan: Fix device info output format specifiers (llama/10366) metal : add `GGML_UNARY_OP_ELU` kernel (ggml/1018) CUDA: fix MMV kernel being used for FP16 src1 (llama/10357) ...
mrienstra
added a commit
to mrienstra/whisper.cpp
that referenced
this pull request
Dec 11, 2024
Introduced in ggerganov#2589
lyapple2008
pushed a commit
to lyapple2008/whisper.cpp.mars
that referenced
this pull request
Feb 4, 2025
lyapple2008
pushed a commit
to lyapple2008/whisper.cpp.mars
that referenced
this pull request
Feb 4, 2025
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Models names as per v1.7.2 announcement.
PR adds
-q8_0
models fortiny
,base
,small
,medium
,large-v2
, &large-v3-turbo
.With the changes in this PR, the remaining differences between the model list in the v1.7.2 announcement and
models/download-ggml-model.sh
are:Only in
models/download-ggml-model.sh
tiny.en
tiny.en-q5_1
base.en-q5_1
small.en
small.en-tdrz
small.en-q5_1
medium.en
medium.en-q5_0
large-v1
large-v3
large-v3-q5_0
Only in v1.7.2 announcement
tiny
tiny-q5_0
base-q5_0
small-q5_0
medium-q5_1
medium-dis
large-v2-q5_1
large-v2-dis