Replies: 2 comments 1 reply
-
Beta Was this translation helpful? Give feedback.
1 reply
-
This repo supports 4 bit quantization. Models released by researchers are usually in full or half precision. Those who have supercomputers are not pressured to squeeze models into Raspberry Pi and 4 bit quantization doesn't work well with training anyways. Different priorities, simply speaking. Researchers are less interested in delivering working product. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello,
first of all, thank you very much for this authentic piece of GOLD!
I just wanted to understand why the official version from facebook, as well as alpaca and vicuna, require a lot RAM to run on the CPU or a lot of memory to run on the GPU, while this version can successfully be used on my CPU with only 16GB RAM decently "fast".
Thanks
Beta Was this translation helpful? Give feedback.
All reactions