koboldcpp-1.0.8beta #13
Closed
LostRuins
announced in
Announcements
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
koboldcpp-1.0.8beta
-O3
to-Ofast
. This should increase generation speed even more, but I dunno if anything will break, please let me know if so.To use, download and run the koboldcpp.exe
Alternatively, drag and drop a compatible quantized model for llamacpp on top of the .exe, or run it and manually select the model in the popup dialog.
and then once loaded, you can connect like this (or use the full koboldai client):
http://localhost:5001
This discussion was created from the release koboldcpp-1.0.8beta.
Beta Was this translation helpful? Give feedback.
All reactions