Ihor Radchenko <[email protected]> writes:

> Christian Moe <[email protected]> writes:
>
>> Looks like it's time to add a note to this effect to
>> https://orgmode.org/worg/org-contribute.html, and possibly elsewhere as
>> well. The points you make that go beyond copyright are well put, and I
>> think it is worth setting them out. Should the copyright issue somehow
>> be resolved to GNU's satisfaction, it doesn't mean that we want to open
>> the floodgates.
>
> Yes, I plan to write a draft, once this and another thread settle.
> I also coined some idea about copyright handling that we can use before
> the official GNU guidance on LLMs is issued. But I still need to check -
> several people on private GNU lists raised concerns.
>
> And GNU guidance seems to be coming (after they consult lawyers). Maybe
> in several months or so.

Hmm.  You've probably also seen this, but...

OpenAI has proven that LLMs have a fundamental problem -- they lie and
their lying is getting more pronounced in the newer models.  The basic
problem is they are trained to *not* say "I don't know" because saying
that would break the foundation of their business plan.  Something to
incorporate in your draft...

https://www.science.org/content/article/ai-hallucinates-because-it-s-trained-fake-answers-it-doesn-t-know


Reply via email to