whatever, ma fa un po' pensare che mentre i ricercatori di Oxford parlano di introdurre una "Legal Duty to Tell the Truth" per i LLM, a pochi chilometri migliaia di hooligan imbufaliti bruciano le macchine perché qualcuno ha diffuso fake news sull'assassino di quelle tre bambine a Southport, dicendo che era un immigrato clandestino musulmano.

A parte l'imbecillità neopositivista degli oxfordiani sulla quale potremo tornare, il problema come dice Stefano è la diffusione, non la genesi.

Isn't it?

G.

On Fri, 9 Aug 2024 at 15:20, Alberto Cammozzo via nexa <nexa@server-nexa.polito.it> wrote:

Il politico e un LLM non sono confrontabili. 

Per favore smettiamola di paragonare sistemi industriali e persone solo perché producono la stessa tipologia di output (o di inquinamento).

Sarebbe come confrontare uno che brucia un pezzo di carta plasticata in giardino con il disastro di Seveso solo perché in entrambi i casi si è prodotta diossina.

A.

PS: No god in the machine: the pitfalls of AI worship
<https://www.theguardian.com/news/article/2024/aug/08/no-god-in-the-machine-the-pitfalls-of-ai-worship>

The rise of artificial intelligence has sparked a panic about computers gaining power over humankind. But the real threat comes from falling for the hype

[...]

On 07/08/24 15:54, Fabio Alemagna wrote:

Mi sembra la descrizione perfetta del comportamento di un politico medio. Se dovessimo  obbligare per legge gli LLM a dire la verità, per quale ragione dovremmo esentare i politici dal fare altrettanto?


Il mer 7 ago 2024, 12:55 J.C. DE MARTIN <juancarlos.demartin@polito.it> ha scritto:
OII | Large Language Models pose a risk to society and need tighter regulation, say Oxford researchers

Written by
Sandra Wachter, Brent Mittelstadt and Chris Russell    

Leading experts in regulation and ethics at the Oxford Internet Institute, part of the University of Oxford, have identified a new type of harm created by LLMs which they believe poses long-term risks to democratic societies and needs to be addressed

Large Language Models pose a risk to society and need tighter regulation, say Oxford researchers

Leading experts in regulation and ethics at the Oxford Internet Institute, part of the University of Oxford, have identified a new type of harm created by LLMs which they believe poses long-term risks to democratic societies and needs to be addressed by creating a new legal duty for LLM providers.

In their new paper ‘Do large language models have a legal duty to tell the truth?’, published by the Royal Society Open Science, the Oxford researchers set out how LLMs produce responses that are plausible, helpful and confident but contain factual inaccuracies, misleading references and biased information.  They term this problematic phenomenon as ‘careless speech’ which they believe causes long-term harms to science, education and society.

continua qui: https://www.oii.ox.ac.uk/news-events/do-large-language-models-have-a-legal-duty-to-tell-the-truth/