Debian and Ubuntu packages - Call for testing and feedback #20042
Replies: 3 comments 29 replies
-
|
@ckastner ? |
Beta Was this translation helpful? Give feedback.
-
|
Here's an interesting problem that we've been hearing from our users, for which we could really use more feedback: Which of ggml's compute backends should the Currently, it only installs the CPU backend. So people with GPUs won't benefit from them unless they know to additionally install the HIP, CUDA or Vulkan backends. Possible strategies1. ggml backend meta-package This strategy involves creating a ggml meta-package With such a setup, by default,
2. llama.cpp meta-packages This strategy adds new packages
3. Vulkan backend This strategy installs the Vulkan backend by default.
|
Beta Was this translation helpful? Give feedback.
-
|
Based on your feedback, we have been working over the last two weeks on the packaging of a ggml CUDA backend built against the Nvidia-provided repositories (cc @aendk). It is available for Debian https://salsa.debian.org/deeplearning-team/ggml/-/blob/debian/unstable/debian/README.vendor-cuda.md There are various use cases where using the Nvidia-provided packages can be useful or necessary, which in turn require a dedicated build of the ggml CUDA backend:
Thanks to the ggml and llama.cpp architectures, all the other packages (currently in the Debian AI team repository, and soon in the We have tried to provide it for Ubuntu These ggml CUDA backend packages for deployments using the Nvidia-provided repositories have been systematically tested with Debian Trixie on AWS instances with Nvidia GPUs (amd64 and arm64). As usual, your feedback is welcome and very useful! |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
we have been working on shipping llama.cpp, whisper.cpp, ggml and other ggml-org projects directly from Debian's and Ubuntu's official repositories.
We want users to be able to just
sudo apt-get install llama.cpp.Our packages should be in good shape now (as tested by the CI we built for this task), so now we would like to ask you, the community, for feedback:
Please check ggml-Debian for a summary of what we have, how to install/use it, and what we are working on.
Note that packages for
trixie-backportsandnoble-backportsare currently still being shipped from our own development repository (instructions included in the link above), but these will be made available in the official channels soon. We want to ship ~monthly updates this way.Also note that newer GPUs might not be supported by HIP/CUDA in the above releases. Try the Vulkan backend in that case.
Beta Was this translation helpful? Give feedback.
All reactions