Ciao Daniela, ho molto apprezzato il fatto che sposti il problema su un piano epistemologico e condivido che per la clinica il paradigma indiziario sia più indicato dei big data. Sono entrambe scelte rare e preziose. (Junio • Pisa)
Il giorno 10 feb 2026, alle ore 12:00, nexa-request@server-nexa.polito.it ha scritto:
Send nexa mailing list submissions to nexa@server-nexa.polito.it
To subscribe or unsubscribe via the World Wide Web, visit https://server-nexa.polito.it/mailman/listinfo/nexa or, via email, send a message with subject or body 'help' to nexa-request@server-nexa.polito.it
You can reach the person managing the list at nexa-owner@server-nexa.polito.it
When replying, please edit your Subject line so it is more specific than "Re: Contents of nexa digest..."
Today's Topics:
1. Nuovo articolo, The imaginary patient. Fantastic health data and where to find them (Daniela Tafani)
----------------------------------------------------------------------
Message: 1 Date: Mon, 9 Feb 2026 14:01:41 +0000 From: Daniela Tafani <daniela.tafani@unipi.it> To: nexa <nexa@server-nexa.polito.it> Subject: [nexa] Nuovo articolo, The imaginary patient. Fantastic health data and where to find them Message-ID: <PAWPR07MB955898E81E7F5A2B86BFAD35F365A@PAWPR07MB9558.eurprd07.prod.outlook.com>
Content-Type: text/plain; charset="Windows-1252"
Buongiorno,
ho appena pubblicato un articolo sui chatbot e la medicina e vi sarei molto grata di qualsiasi osservazione.
The imaginary patient. Fantastic health data and where to find them <https://btfp.sp.unipi.it/en/2026/01/the-imaginary-patient/>
Abstract In the name of an upcoming AI revolution in healthcare and of chatbots and ‘agentic AI’ allegedly capable of providing health advice, the US, international institutions, and the European Commission are pushing for the digitisation of clinical and health data. Unfortunately, ‘digitisation’ is just a new name for mass surveillance, there is no such a thing as ‘agentic AI’ and those surveilling us are not interested in making healthcare more accessible or personalised, which would simply require hiring more healthcare personnel. At best, a machine learning system capable of tracking correlations can be an auxiliary tool in medicine. An extruder of probable text strings can only make us imagine that we are in a doctor–patient relationship, but what a doctor does is certainly not autocomplete. A chatbot is therefore both useless and dangerous, if the objective is healthcare. However, a private healthcare company whose goal is to increase quarterly dividends may consider a chatbot to be a useful cost-reduction tool. Within an instrumental rationality, a healthcare company may find it rational to replace doctors with chatbots, just as it would find it rational to feed prisoners dog food if it ran a prison. After all, financial analysts who think in terms of business logic are seriously asking themselves whether treating patients is a sustainable business model. Announcements of an AI-based healthcare revolution have no scientific basis. However, they are deeply rooted in sectors that are heavily funded by the pharmaceutical industry and technology companies. Here, the interest in promises of automation, in dehumanising patients, in commodifying health and in destroying public healthcare systems converge with the surveillance, control, manipulation and domination goals of the US military–industrial complex. Even if it cannot cure us, generative AI is perfect for all these purposes.
Un saluto, Daniela
------------------------------
Subject: Digest Footer
_______________________________________________ nexa mailing list nexa@server-nexa.polito.it https://server-nexa.polito.it/mailman/listinfo/nexa
------------------------------
End of nexa Digest, Vol 202, Issue 9 ************************************