Skip to content
This repository has been archived by the owner on Jun 24, 2024. It is now read-only.

Fails to run with gpt4-x-alpaca-13b-native-ggml-model-q4_0 #102

Closed
rumatoest opened this issue Apr 2, 2023 · 2 comments
Closed

Fails to run with gpt4-x-alpaca-13b-native-ggml-model-q4_0 #102

rumatoest opened this issue Apr 2, 2023 · 2 comments

Comments

@rumatoest
Copy link

rumatoest commented Apr 2, 2023

See https://huggingface.co/Bradarr/gpt4-x-alpaca-13b-native-ggml-model-q4_0

cargo run --release -- -m ./gpt4-x-alpaca-13b-native-ggml-model-q4_0.bin -p "How do you do ?"
thread 'main' panicked at 'Could not load model: InvalidMagic { path: ".../gpt4-x-alpaca-13b-native-ggml-model-q4_0.bin" }', llama-cli/src/main.rs:206:6
note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace

@philpax
Copy link
Collaborator

philpax commented Apr 2, 2023

My guess is that it's using the GGJT format, which we don't support yet (#93)

@rumatoest
Copy link
Author

I was focused mostly on ggml and q4 keywords. Should also read whether it is compatible with llama.cpp

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants