On Fri, 2025-01-31 at 12:31 +0100, Simon Josefsson wrote: > Package: wnpp > Severity: wishlist > Owner: Simon Josefsson <si...@josefsson.org> > X-Debbugs-CC: debian-de...@lists.debian.org, debian...@lists.debian.org > > * Package name : ollama > Version : 0.5.7-1 > Upstream Author : Ollama > * URL : https://github.com/ollama/ollama > * License : Expat > Programming Lang: Go > Description : large language model tools > > Ollama: Get up and running with large language models. > > https://salsa.debian.org/go-team/packages/ollama > https://salsa.debian.org/jas/ollama/-/pipelines
@ckk is planning to package llama.cpp within debian deep learning team (debian-ai@l.d.o). Maybe you want to discuss with the team whether you want to deal with the embedded copy of llama.cpp inside ollama source tree? I did not look into how ollama enables ROCm and CUDA support, but that's also something the team care about. Do you want to enable any of them?