You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Jun 24, 2024. It is now read-only.
Ok I sort of figured it out. For others looking into this:
Use the GGUF branch
Run it like this: cargo run --release -- infer -m ~/llms/mistral-7b-instruct-v0.1.Q4_K_S.gguf -p "Write a long story" -n 5000 -r mistralai/Mistral-7B-v0.1
The important bit is to tell it to use the tokenizer at the end.
Yup - the gguf branch's embedded tokenizer support doesn't quite work, so I've disabled it for now. I'm rebuilding the library in develop to target the latest llama.cpp, but that's going to take some time.
Do you all support this? https://mistral.ai/news/announcing-mistral-7b/
The text was updated successfully, but these errors were encountered: