Hi Mo,

before going into criticizing things, I would like to thank you for your
continued work in the AI space. In particular, your way of classifying
models into different degrees of freedom demonstrates how much you care
about Debian's values. I see that your focus is on enabling users and
that's amazing!

On Sun, Jan 12, 2025 at 05:01:33PM -0500, M. Zhou wrote:
> One long term goal of debian deep learning team is to host an LLM with
> the team's AMD GPUs and expose it to the members. That said, the necessary
> packages to run that kind of service are still missing from our archive.
> It is a good way to use existing GPUs any way.

Disregarding the ecological and political aspects that have been
discussed elsewhere at length, let me also note that successfully using
a LLM is not trivial. Whether you get something useful very much depends
on what you ask. Typically, you should expect that something between 30%
and 70% of answers are factually wrong in some regard. As a result,
asking questions where the answer is not verifiable is not very useful.
Using the answers without verifying them poses a risk to us as a
community that myths are propagated and become harder to falsify. We are
talking about ai dementia already and can observe it happening. How do
you see us mitigating this problem?

That said, it can be useful to use a LLM if your premise is that you
verify the answer. For instance if you are searching for a library
function that does something particular, searching the documentation can
be time consuming. Describing your function to a LLM and then looking up
the documentation of the presented suggestions has a significant chance
of turning up something useful (in addition to the rubbish included).
Likewise, searching for a particular statement in a mailing list thread
can be a daunting task where a LLM may be able to interpret your vague
memory sufficiently well that it can provide a few candidates.

This way of using LLMs is effectively limiting it to behave as a search
engine with a different query language. I suspect that most of us use
search engines all day without thinking much about them being driven by
corporate actors running big clusters. Is using LLMs in this way that
much different from using a search engine?

Ecological and economical reasons aside, would anyone see an issue with
providing an AI-driven search engine to Debian's documentation, mailing
lists, bugs and the wiki?

> Even if we get no commercial sponsorship of API calls, we will eventually
> experiment and evaluate one with the team's infrastructure. We are still
> working towards that.

I am actually looking forward to this as I trust you to do it in a
responsible way given your earlier work.

> First, DebGPT is designed to conveniently put any particular information
> whether or not Debian-specific, to the context of LLM. I have also implemented
> some mapreduce algorithm to let the LLM deal with extremely overlength
> context such as a whole ratt buildlog directory.

Indeed. I can imagine that a LLM may be able to also suggest particular
lines in a build log that may hint at the cause of a failure whereas we
now have a set of patterns (e.g. "Waiting for unfinished jobs") and hope
that one of them locates the problem.

> My intention is always to explore possible and potential ways to make LLM
> useful in any possible extent. To support my idea, I wrote DebGPT, and
> I tend to only claim things that is *already implemented* and *reproducible*
> in DebGPT.

I have not experimented with DebGPT yet. Maybe I should. Unless already
done so, let me suggest that you engineer the default prompt in such a
way that your LLM references its information sources whenever possible.
Also including some text that asks the user to verify the answer using
external sources may help (in an opt-out way).

Practically speaking, the use of neural networks is not something we can
stop even if we wanted to. In ten years, all of us will be using neural
networks on a daily basis. Even today, they're difficult to avoid (e.g.
when dealing with customer service of any corporation). The best we can
do here is making the use of them as free as possible and that's what I
see Mo is doing.

Helmut

Reply via email to