Skip to content

Running llama.cpp on android just prints out the question #712

New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

Closed
Shreyas-ITB opened this issue Apr 2, 2023 · 2 comments
Closed

Running llama.cpp on android just prints out the question #712

Shreyas-ITB opened this issue Apr 2, 2023 · 2 comments
Labels
android Issues specific to Android stale

Comments

@Shreyas-ITB
Copy link

I ran llama.cpp on my android phone which has 8 threads and 8GB of ram in which around 7.16 GB is available, that is more than enough to run the 7B Alpaca model on it. But when i run it, it just repeats the question that i provided to it. I am using the ./examples/chat.sh file. Why does it do that? How do i solve it?

@cmp-nct
Copy link
Contributor

cmp-nct commented Apr 2, 2023

Just guessing: after prompt was processed there can be a noticeable delay until the completions start.
Also there are interactive modes that wait for return/enter.

@sw sw added the android Issues specific to Android label Apr 4, 2023
@github-actions github-actions bot added the stale label Mar 25, 2024
Copy link
Contributor

This issue was closed because it has been inactive for 14 days since being marked as stale.

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
android Issues specific to Android stale
Projects
None yet
Development

No branches or pull requests

3 participants