-
Notifications
You must be signed in to change notification settings - Fork 11.5k
Error to loading the latest Chinese LLaMa Alpaca model #1464
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Download the previous version : |
I'v encountered the same problem. The some steps above , no errors while merging and quantizing. just hangs when loading the model, before ">" appears. but the model seems loaded without error. |
the previous version don't work , even can't load 8-bit quantized model |
I found the problem of it. The original document suggest to convert the model using the command like this: I read the convert.py carefully and found it has a parameter of vocab-dir: The document ask to put the tokenizer.model in the upper level directory, I guess maybe it can't use this tokenizer.model file and in fact the tokenizer.model in the Chinese Alpaca model is different with the original LLaMa model. So at last I add the --vocab-dir parameter to specify the directory of the Chinese Alpaca's tokenizer.model. |
I just reinstalled ubuntu vm and clone the latest version, re-compiled it. then it works. that's weird |
I pull the latest release version, rerun make and ./main, and it worked. The reason for me is that I convert and quantize with the newer version but trying to run with an older version. |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
Errors when loading the latest Chinese LLaMA/Alpaca Plus-13B model:
./main -m ../ggml-alpaca13b-q5_1.bin -n 256 --repeat_penalty 1.0 --color -i -r "[Steve]:" -f chat-with-vicuna-v1.txt
main: build = 526 (e6a46b0)
main: seed = 1684135558
llama.cpp: loading model from ../ggml-alpaca13b-q5_1.bin
error loading model: unknown (magic, version) combination: 67676a74, 0000000; is this really a GGML file?
llama_init_from_file: failed to load model
llama_init_from_gpt_params: error: failed to load model '../ggml-alpaca13b-q5_1.bin'
main: error: unable to load model
Here are the steps to combine the Chinese alpaca model with original llama model:
No errors in the above steps.
But when loading the model using main program, there are errors like above.
It seemed that (magic, version) combination: 67676a74, 0000000 are not supported when loading the model.
Any solution or suggestion about this? Thanks!
The text was updated successfully, but these errors were encountered: