-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
main: failed to quantize model from './models/7B/ggml-model-f16.bin' #18
Comments
try using #16 with virtualenv |
you can move the models you already download to llama.cpp folder in the #16 repo |
Hi, I am getting this error maybe there is a relation with the other error.
|
I am getting the "Illegal instruction: 4" error as well. It looks like it is compiling it for x86_64 rather than ARM.
|
I am getting the same error as spullara. I am on an M1, and curiously, when I run the make command directly from the Edit: I got it working, but this might be very unique to my computer. These are the steps I took:
I'm not 100% sure why this worked, but I did it twice just to make sure. |
ggerganov/llama.cpp#41 (comment) this comment works! |
It looks like this assumesnumpy
is installed but does not install it if it is not. I don't notice a mention of needingnumpy
to be installed before running in the README. Maybe a doc update and check / exit if not installed would be sufficient.This does try to install
numpy
andtorch
but if that fails for some reason the build continues and then fails at the end.The server also starts without error in this case... but has no response to any prompts (error or output).
1st Error - This should probably stop the build when it fails
2nd Error
The text was updated successfully, but these errors were encountered: