Is There Any Other Settings To Use It In GPU? #75
Replies: 2 comments 2 replies
-
Curious about this for M1/M2 gpus/neural engines |
Beta Was this translation helpful? Give feedback.
-
Llama.cpp is looking at ways to utilize GPUs, but it's not a major goal of the repo (which is CPU oriented). Take a look here for some info that might help: Just including this in case it helps: Personally, I think it's cool, but I hate interacting with GPTQ code/quantized models because it has so many different versions and changes often. It's not very stable imo. |
Beta Was this translation helpful? Give feedback.
-
Hello,
Is there any other settings to use it in GPU? I use 'Vicuna GGML' version. I know 'GPTQ' version is for GPU. But, I found a huggingface GPTQ version is more stable.
Please, Let me know.
Thank You!
Beta Was this translation helpful? Give feedback.
All reactions