Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Only Distribute CUDA Binaries? #298

Closed
martindevans opened this issue Nov 15, 2023 · 2 comments
Closed

Only Distribute CUDA Binaries? #298

martindevans opened this issue Nov 15, 2023 · 2 comments

Comments

@martindevans
Copy link
Member

As of this PR to llama.cpp the CUDA binaries are capable of running with CPU only, as long as n_gpu_layers = 0.

This might mean that we can significantly simplify our distribution of binaries by removing the CPU only variants and only shiping CUDA ones.

@martindevans martindevans moved this to 🔖 In Discussion in LLamaSharp Dev Nov 15, 2023
@martindevans
Copy link
Member Author

There's some further discussion over in the linked PR with some potential issues that may make this less attractive:

  • Even with ngl=0 some work is done on GPU (maybe)

@AsakusaRinne
Copy link
Collaborator

It's a good news to hear this feature supported by llama.cpp. However, I have some disagreements.

If the user does not have the cuda installed, using cuda backend makes little sense. Besides, the cpu backend is much smaller than the cuda backend, which may be desired in some conditions. I prefer to keep separating these two backends. :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
Status: ✅ Done
Development

No branches or pull requests

2 participants