-
Notifications
You must be signed in to change notification settings - Fork 11.5k
FP16 and 4-bit quantized model both produce garbage output on M1 8GB #137
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Thtats nothing to do with this project. |
@v3ss0n could you please elaborate? |
I started messing with this project two hours ago and had exactly same issue. Completely mangled output. |
I found the solution to my issue! make sure that when you're using the |
Improve readme: clarify dependencies and other things to install
Both the
ggml-model-q4_0
andggml-model-f16
produce a garbage output on my M1 Air 8GB, using the 7B LLaMA model. I've seen the quantized model having problems but I doubt the quantization is the issue as the non-quantized model produces the same output.The text was updated successfully, but these errors were encountered: