On 2025-02-06 02:42, M. Zhou wrote:
> I second this. llama-server is also the service endpoint for DebGPT.

I'll prioritize fixing this.

> I pushed a fix for ppc64el. The hwcaps works correctly for power9, given the 
> baseline is power 8.

Ah good catch. The broken install pattern was due to a last-minute fix
that I only tested on amd64...

I meant to ask anyway: performance-wise, is it comparable to your local
build? I mean, I wouldn't know what in the code would alter this, but I
built and tested this on platti.d.o and performance was poor, so another
data point would be useful.

Best,
Christian

Reply via email to