-
Notifications
You must be signed in to change notification settings - Fork 7.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support for QWEN and Baichuan2 models #1731
Comments
Baichuan2 should be supported in the next release - the current release isn't using the latest version of our llama.cpp fork. Qwen got merged upstream a little too late, but it should be supported here soon. |
There is a problem. When I use Qwen, no matter the graphics card or CPU specified by my device, it has been calling the CPU and memory for calculation instead of calling the GPU. |
As of #2005, Qwen, Qwen2, and Baichuan all have GPU support enabled on both the Metal and Vulkan backends. |
This is implemented in the v2.7.2 release. |
Feature request
recently, https://github.com/ggerganov/llama.cpp has add support for both QWEN and Baichuan2.
It has added QWEN at 1610.
ggerganov/llama.cpp#4281
I have look up the Nomic Vulkan Fork of LLaMa.cpp,
it does have support for Baichuan2 but not QWEN, but GPT4ALL itself does not support Baichuan2.
Motivation
I failed to load baichuan2 and QWEN models, GPT4ALL supposed to be easy to use
Your contribution
Not quite as i am not a programmer but i would look up if that helps
The text was updated successfully, but these errors were encountered: