-
Notifications
You must be signed in to change notification settings - Fork 11.5k
llama.cpp acts too dumb while running on phone!! #802
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
try playing with settings, like increasing temperature |
Okay ill try it and see then ill let you know |
@BarfingLemurs Nope it doesnt work, it stays the same (acts way way dumber and still repeats the question) |
You may want to log an issue with the Stanford Alpaca. It is the training set fo the Alpaca model you are using. |
@gjmulder im using alpaca lora maybe thats the issue? I mean its not the problem of the alpaca model tho it worked fine on my laptop and for many users on their PCs on mobile it malfunctions |
Maybe because of termux. I don't even know how to use the Firefox in termux to watching YouTube 720p without crash😅😂 |
@FNsi lol its not due to termux. The same issue happens with userland (an app thats intended to run ubuntu on android without root) |
The same model should work the similarly with |
There are various optimized code paths that are only enabled for certain platform and feature sets, there could be differences in the implementation of those. Could you post the initial output with the system_info and model parameters? |
@unbounded the output i get is in the start of the issue (there is a screenshot of what the model is saying) System Info |
Closing as assumed fixed by 684da25 , feel free to reopen if this still happens with the latest version. |
I was trying llama.cpp on phone with termux installed. but look at this image

Specifications
The phone has 8 gigs of RAM and 7 gigs is free and the CPU has 8 cores so its not the issue of the RAM and CPU.
Model used: alpaca-7B-lora
llama.cpp version: latest
prompt: chat-with-bob.txt
I really don't know what is causing the issue here. The problem happening is, when i ask a question to it, it just either answers the question in a very dumb way or it just repeats the same question not answering anything. With the same model, prompt and llama.cpp version on my PC with 4GB ram works as expected it answers every question with almost 98% accuracy. Can any of you guys help me out with this? or update the llama.cpp and fix the mobile issues please?
Thankyou
The text was updated successfully, but these errors were encountered: