Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

talk-llama: only copy used KV cache in get / set state #890

Merged

Conversation

herrera-luis
Copy link
Contributor

@herrera-luis herrera-luis commented May 8, 2023

adding the KV cache improvement made in the llama repo:

ggerganov/llama.cpp#1272

---------

Co-authored-by: ejones <evan.q.jones@gmail.com>
@ggerganov ggerganov merged commit 4e4d00c into ggerganov:master May 8, 2023
jacobwu-b pushed a commit to jacobwu-b/Transcriptify-by-whisper.cpp that referenced this pull request Oct 24, 2023
---------

Co-authored-by: ejones <evan.q.jones@gmail.com>
jacobwu-b pushed a commit to jacobwu-b/Transcriptify-by-whisper.cpp that referenced this pull request Oct 24, 2023
---------

Co-authored-by: ejones <evan.q.jones@gmail.com>
landtanin pushed a commit to landtanin/whisper.cpp that referenced this pull request Dec 16, 2023
---------

Co-authored-by: ejones <evan.q.jones@gmail.com>
iThalay pushed a commit to iThalay/whisper.cpp that referenced this pull request Sep 23, 2024
---------

Co-authored-by: ejones <evan.q.jones@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants