Skip to content

Running llama.cpp on android just prints out the question #712

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
Shreyas-ITB opened this issue Apr 2, 2023 · 2 comments
Closed

Running llama.cpp on android just prints out the question #712

Shreyas-ITB opened this issue Apr 2, 2023 · 2 comments
Labels
android Issues specific to Android stale

Comments

@Shreyas-ITB
Copy link

I ran llama.cpp on my android phone which has 8 threads and 8GB of ram in which around 7.16 GB is available, that is more than enough to run the 7B Alpaca model on it. But when i run it, it just repeats the question that i provided to it. I am using the ./examples/chat.sh file. Why does it do that? How do i solve it?

@cmp-nct
Copy link
Contributor

cmp-nct commented Apr 2, 2023

Just guessing: after prompt was processed there can be a noticeable delay until the completions start.
Also there are interactive modes that wait for return/enter.

@sw sw added the android Issues specific to Android label Apr 4, 2023
@github-actions github-actions bot added the stale label Mar 25, 2024
Copy link
Contributor

This issue was closed because it has been inactive for 14 days since being marked as stale.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
android Issues specific to Android stale
Projects
None yet
Development

No branches or pull requests

3 participants