On 2025-02-06 01:33, Petter Reinholdtsen wrote:
> [Christian Kastner]
> I checked in a few minor fixes.

Look fine, though I deliberately skipped the poetry dependency for now
as it looked more like a false positive.

> I noticed llama.cpp depend on llama.cpp-backend with no concrete
> dependency first.  This lead to unpredictable behaviour, and I suggest
> depending on for example 'llama.cpp-cpu | llama.cpp-backend' to make
> sure 'apt install llama.cpp' behave predictably.

This was my intention, but I initially wasn't sure what the default
would be (-cpu or -blas). Looks like I forgot to add one before upload.

> I was sad to discover the server example is missing, as it is the
> llama.cpp progam I use the most.  Without it, I will have to continue
> using my own build.

It'll be re-enabled soon. The were a few generated and minified files in
that example, so I just opted to skip those for now, and focus on the
build process.

> I hope to get whisper.cpp to the same state, so it can have a fighting
> chance to get into testing before the freeze.

Seeing as how closely llama.cpp and whisper.cpp are related, in the
ideal case, you should be able to just carry over some patches, and
mostly just copy d/rules, as llama.cpp and whisper.cpp share the ggml
library on a source basis.

Note that for amd64, there actually is a type of dynamic dispatching for
ggml, though it needs to be patched as the lowest supported level is
AVX. Also something I had planned for the next iteration.

Best,
Christian

Reply via email to