Caro Fabio,

convengo con te che il (c) abbia dei limiti in tale contesto, e soprattutto non credo che il /fair use/ sia l'unico criterio utile.

Le corti in varie parti del mondo saranno sempre più investite in merito, vedremo che orientamento prenderanno e se la legge sul (c) sarà l'unico strumento azionato. Gli LLM sfidano un apparato giuridico che non è pensato per la produzione industriale di testi, fenomeno che finora non esisteva.

In merito a quanto dici sull'equivalenza per macchine ed umani di '/istruirsi/' sullo stesso testo, per parte mia credo che l'/apprendimento/ umano e il /training/ del modello LLM abbiano una enorme differenza.

L'umano è in grado di esprimersi col linguaggio anche senza i testi in questione, e di estrarre le 'idee' in esso contenute prescindendo dalla formulazione esatta, mentre la macchina produce linguaggio statisticamente correlato con la semantica associata a quelle idee solo seguendo la formulazione linguistica dei  testi pertinenti, e solo con tali testi. Non potrei dire la stessa cosa pensando a uno studente che /si istruisce/ dai libri.

Per la macchina l'idea (che per noi è significante, contenuto) non esiste, ma solo il linguaggio (il significante): anche se questa non produrrà frasi che copiano letteralmente l'input di training, lo specifico input è essenziale alla produzione di testi con la semantica dell'input in questione.

Vedrei poi altri aspetti che le corti potrebbero tenere in considerazione, che emergeranno forse maggiormente in futuro, ma che meriterebbero approfondimento ora.

- l'appropriazione del lavoro linguistico non riconosciuto dell'autore, anche come collettività autorale. Questo mi pare si veda già con la produzione automatica di codice informatico attingendo ai repository;

- la responsabilità sulle conseguenze del contenuto dei testi e eventuali danni derivanti dalla scarsa qualità dello stesso (già vediamo ricette velenose, fake news, suicidi, induzione a comportamenti pericolosi, bug nel codice ...);

- l'inquinamento ambientale. Non solo quello energetico (per la produzione e l'aggiornamento degli LLM), ma inquinamento informativo dell'ecosistema linguistico (o in generale simbolico). Ammettendo il linguaggio come bene sistemico e patrimonio comune, l'immissione massiva di testi (o immagini) generati artificialmente interferisce con l'ecosistema e la sua naturale evoluzione.

Questo ultimo aspetto è quello meno immediatamente visibile ma sarà il più intenso, e investirà per primi i motori di ricerca e gli altri attori dell'ecosistema digitale, che vedranno diluirsi il rapporto segnale/rumore all'aumentare dei testi generati artificialmente, e di conseguenza il valore del loro servizio. Dovranno decidere da che parte stare... Anche la produzione di software a codice aperto risentirà dello stesso problema.

In generale le collettività che producono testi, codice e immagini e che li riversano nei commons ne subiscono un danno dal momento che la produzione industriale artificiale sommergerà il loro ecosistema con prodotti di qualità dubbia e che sottraggono loro lavoro e riconoscimento.

Stiamo (anche) a vedere...

Alberto


On 29/09/23 10:24, Fabio Alemagna wrote:
L'idea che istruire un modello su dei testi coperti da copyright sia
una violazione del suddetto copyright è altamente opinabile, e il
ragionamento è in realtà abbastanza semplice: se istruirsi su un testo
ne violasse il copyright, saremmo tutti dei criminali. Di quei testi
non c'è fisicamente traccia all'interno dei modelli, non viene copiato
niente. I modelli sono un'opera trasformativa di quei testi, non
derivativa.

Lo argomenta molto bene Creative Commons:
https://creativecommons.org/2023/02/17/fair-use-training-generative-ai/

Detto questo, cito le parole di un altro autore, Jeff Jarvis:
https://www.facebook.com/jeff.jarvis/posts/pfbid0LMFeqdTYoxnGHQAZwp5HMmeeVqgMSjL2dkcwMcBojkb2cinBpgYTHyc7Fhq1B9NPl

«I, for one, am not complaining about my books being in in large
language model training sets. I write to enter ideas into public
discourse. I prefer informed over ignorant AI. I believe it is fair
use for anyone to read & use books for transformative work. In fact,
I'd probably feel snubbed if my books were not there. I'm happy when
they are in libraries. I'm fine that they're here.»

Fabio

Il giorno ven 29 set 2023 alle ore 07:52 Alberto Cammozzo via nexa
<nexa@server-nexa.polito.it>  ha scritto:
<https://www.theguardian.com/australia-news/2023/sep/28/australian-books-training-ai-books3-stolen-pirated>


Thousands of books from some of Australia’s most celebrated authors have 
potentially been caught up in what Booker prize-winning novelist Richard 
Flanagan has called “the biggest act of copyright theft in history”.

The works have allegedly been pirated by the US-based Books3 dataset and used 
to train generative AI for corporations such as Meta and Bloomberg.

Flanagan, who found 10 of his works, including the multi-international 
award-winning 2013 novel The Narrow Road to the Deep North, on the Books3 
dataset, told Guardian Australia he was deeply shocked by the discovery made 
several days ago.

“I felt as if my soul had been strip mined and I was powerless to stop it,” he 
said in a statement.

“This is the biggest act of copyright theft in history.”

AI could ‘turbo-charge fraud’ and be monopolised by tech companies, Andrew 
Leigh warns


The Australian Publishers Association confirmed to Guardian Australia on 
Wednesday that as many as 18,000 fiction and nonfiction titles with Australian 
ISBNs (unique international standard book numbers) appeared to be affected by 
the copyright infringement, although it is not yet clear what proportion of 
these are Australian editions of internationally authored books.

“We’re still working through [the data] to work out the impact in terms of 
Australian authors,” APA spokesperson Stuart Glover said.

“This is a massive legal and ethical challenge for the publishing industry and 
for authors globally.”

A search tool published on Monday by US media platform The Atlantic and 
uploaded by the US Authors Guild on Wednesday revealed the works of Peter 
Carey, Helen Garner, Kate Grenville, Anna Funder, Christos Tsiolkas and Thomas 
Keneally, as well as Flanagan and dozens of other high-profile Australian 
authors, were included in the pirated dataset containing more than 180,000 
titles.

On Thursday, the Australian Society of Authors issued a statement saying it was 
“horrified” to learn that the works of Australian writers were being used to 
train artificial intelligence without permission from the authors.

ASA chief executive, Olivia Lanchester, described the Books3 dataset as piracy 
on an industrial scale.

“Authors appropriately feel outraged,” Lanchester said. “The fact is this 
technology relies upon books, journals, essays written by authors, yet 
permission was not sought nor compensation granted.”

Lanchester said the Australian literary industry, while not objecting per se to 
emerging technologies such as AI, was deeply concerned about the lack of 
transparency evident in the development and monetisation of AI by global tech 
companies.

“Turning a blind eye to the legitimate rights of copyright owners threatens to 
diminish already precarious creative careers,” she said.

“The enrichment of a few powerful companies is at the cost of thousands of 
individual creators. This is not how a fair market functions.”

Josephine Johnston, chief executive of Australia’s Copyright Agency, described 
the Books3 development as “a free kick to big tech” at the expense of 
Australia’s creative and cultural life.

“We’re going to need greater transparency – how these tools have been 
developed, trained, how they operate – before people can truly understand what 
their legal rights might be,” she said.

“We seem to be in this terrible position now where content owners – remembering 
that the vast majority of them will be individual authors – may actually have 
to take out court cases to enforce their rights.”

Australian copyright law protects creators of original content from data 
scraping.

Litigation in the US against ChatGPT creator OpenAI over use of allegedly 
pirated book datasets, Books1 and Books2 (which do not appear to be affiliated 
with Books3) has already commenced.


In July, North American horror/fantasy writers Mona Awad (author of Bunny) and 
Paul Tremblay (author of The Cabin at the End of the World) filed a lawsuit in 
a San Francisco federal court, alleging ChatGPT unlawfully digested their books 
as part of its AI training data.

On 28 August, OpenAI filed a motion to dismiss the lawsuit, arguing that the 
authors “misconceive the scope of copyright, failing to take into account the 
limitations and exceptions (including fair use) that properly leave room for 
innovations like the large language models now at the forefront of artificial 
intelligence”.

On 19 September the Writers Guild and 17 of its members, including bestselling 
novelists John Grisham, George RR Martin and Jodi Picoult, filed a complaint in 
a New York district court against OpenAI, seeking redress for “flagrant and 
harmful infringements” of guild members’ registered copyrights.

In a statement on its website, the guild says while it is aware that companies 
such as Meta and Bloomberg have used the Books3 dataset to train their LLMs, it 
is not yet clear whether OpenAI is using Books3 to train its ChatGPT models GPT 
3.5 or GPT 4.

Democracies face ‘truth decay’ as AI blurs fact and fiction, warns head of 
Australia’s military


Guardian Australia has sought comment from OpenAI, which has yet to officially 
respond to the guild’s complaint, and Meta.

On 4 September, US technology magazine Wired reported that a Danish anti-piracy 
group called Rights Alliance had been told by Bloomberg that the company did 
not plan to train future versions of its BloombergGPT using Books3.

Bloomberg declined to respond to the Guardian’s queries.

The APA said the global nature of the issue would present significant 
challenges in enforcement and prosecution, and has joined the authors’ society 
in calling for AI technologies to be regulated.

Consultation closed last month for a Department of Industry, Science and 
Resources discussion paper on supporting responsible AI.

A parliamentary inquiry is under way examining the use of generative artificial 
intelligence in the Australian education system.

Flanagan said it was up to the Australian government to act to protect 
Australia’s writers.

“It has power and we do not,” he said.

“If it cares for our culture it must now stand up and fight for it.”

_______________________________________________
nexa mailing list
nexa@server-nexa.polito.it
https://server-nexa.polito.it/cgi-bin/mailman/listinfo/nexa
_______________________________________________
nexa mailing list
nexa@server-nexa.polito.it
https://server-nexa.polito.it/cgi-bin/mailman/listinfo/nexa

Reply via email to