The US supreme court case that could bring the tech giants to their knees | John Naughton | The Guardian
<https://www.theguardian.com/commentisfree/2022/oct/22/the-us-supreme-court-c...> Two weeks ago, the US supreme court decided that it would hear Gonzalez v Google, a landmark case that is giving certain social-media moguls sleepless nights for the very good reason that it could blow a large hole in their fabulously lucrative business models. Since this might be good news for democracy, it’s also a reason for the rest of us to sit up and pay attention. First, some background. In 1996, two US lawmakers, Representative Chris Cox from California and Senator Ron Wyden from Oregon, inserted a clause into the sprawling telecommunications bill that was then on its way through Congress. The clause eventually became section 230 of the Communications Decency Act and read: “No provider or user of an interactive computer service shall be treated as the publisher or speaker of any information provided by another information content provider.” The motives of the two politicians were honourable: they had seen how providers of early web-hosting services had been held liable for damage caused by content posted by users over whom they had no control. It’s worth remembering that those were early days for the internet and Cox and Wyden feared that if lawyers had henceforth to crawl over everything hosted on the medium, then the growth of a powerful new technology would be crippled more or less from birth. And in that sense they were right. What they couldn’t have foreseen, though, was that section 230 would turn into a get-out-of-jail card for some of the most profitable companies on the planet – such as Google, Facebook and Twitter, which built platforms enabling their users to publish anything and everything without the owners incurring legal liability for it. So far-reaching was the Cox-Wyden clause that a law professor eventually wrote a whole book about it, The Twenty-Six Words That Created the Internet. A bit hyperbolic, perhaps, but you get the idea. Now spool forward to November 2015 when Nohemi Gonzalez, a young American studying in Paris, was gunned down in a restaurant by the Islamic State terrorists who murdered 129 other people that night. Her family sued Google, arguing that its YouTube subsidiary had used algorithms to push IS videos to impressionable viewers, using the information that the company had collected about them. Their petition seeking a supreme court review argues that “videos that users viewed on YouTube were the central manner in which IS enlisted support and recruits from areas outside the portions of Syria and Iraq which it controlled”. The key thing about the Gonzalez suit, though, is not that YouTube should not be hosting IS videos (section 230 allows that) but that its machine-learning “recommendation” algorithms, which may push other, perhaps more radicalising, videos, renders it liable for the resulting damage. Or, to put it crudely, while YouTube may have legal protection for hosting whatever its users post on it, it does not – and should not – have protection for an algorithm that determines what they should view next. This is dynamite for the social-media platforms because recommendation engines are the key to their prosperity. They are the power tools that increase the user “engagement” – keeping people on the platform to leave the digital trails (viewing, sharing, liking, retweeting, purchasing, etc) – that enable the companies to continually refine user profiles for targeted advertising. And make unconscionable profits from doing so. If the supreme court were to decide that these engines did not enjoy section 230 protection, then social media firms would suddenly find the world a much colder place. And stock-market analysts might be changing their advice to clients from “hold” to “sell”. Legal scholars have been arguing for decades that section 230 needs revision. Freedom of speech fanatics see it as a keystone of liberty, as the “kill switch” of the web. Donald Trump made threatening noises about it. Tech critics (such as this columnist) regard it as an enabler of corporate hypocrisy and irresponsibility. However you look at it, though, it’s more than half a century since it became law, which is about 350 years in internet time. Having such a statute to regulate the contemporary networked world seems a bit like having a man with a red flag walking in front of a driverless car. (Though, come to think of it, that might not be such a bad idea.) Versions of the question posed by the Gonzalez suit – whether section 230 immunises internet platforms when they make targeted recommendations of content posted by other users – have been put to US courts over the last few years. To date, five courts of appeals judges have concluded that the section does provide such immunity. Three appeals judges have ruled that it does not, while one other has concluded only that legal precedent precludes liability for recommendation engines. There’s no legal consensus here, in other words. It’s high time that the supreme court decided the matter. After all, isn’t that what its there for? .
Ci andrei cauto. https://www.techdirt.com/2022/06/07/academic-paper-shows-how-badly-the-mains... Federico
Alberto Cammozzo via nexa <nexa@server-nexa.polito.it> writes: [...]
The key thing about the Gonzalez suit, though, is not that YouTube should not be hosting IS videos (section 230 allows that) but that its machine-learning “recommendation” algorithms, which may push other, perhaps more radicalising, videos, renders it liable for the resulting damage. Or, to put it crudely, while YouTube may have legal protection for hosting whatever its users post on it, it does not – and should not – have protection for an algorithm that determines what they should view next.
Non si tratta di "un algoritmo" ma di software [1] a cui vengono dati in pasto tonnellate di dati personali, quindi faccio fatica a comprendere perché la cosa dovrebbe essere così inutilmente complicata: il software [1] - in questo caso utilizzato per **profilare** e SPINGERE contenuti sulle "piattaforme" con lo scopo di far stare attaccati i propri /consumatori/ - NON ricade nell'ambito di applicazione della sezione 230. A me pare semplicissimo ma mi dicono che io sono quello che la fa sempre facile (mentre io ho il fondato sospetto che ci siano troppe persone che la fanno più complicata di quello che è). [...]
Legal scholars have been arguing for decades that section 230 needs revision.
Già così è difficile oltre ogni livello di sopportazione umana riuscire a stare sul mercato dei servizi erogati via internet, proprio grazie ai colossi che li danno via gratis in cambio dei dati che succhiano... vi immaginate cosa potrebbe succedere se "per caso" il principio della section 230 venisse messo in dicussione o anche solo leggermente "complicato"?!? Pensate se "Wordpress.com" corresse anche solo lontanamente il rischio di essere coinvolta in una causa per qualche stronzata che uno dei suoi clienti scrive su un blog "hostato" attraverso i suoi server... idem per servizi tipo "forum" o... mailing lists! Pensate se il Centro Nexa potesse essere coinvolto in una controversia legale a causa delle stronzate che scrivo io qui!?! Quindi di che diavolo di "riforma" della section 230 stiamo parlando? Non è /infinitamente/ più semplice trattare le cose per quello che sono: ripetiamo tutti insieme "il software non è contenuto", "il software non è contenuto", "il software non è contenuto".... Saluti, 380° [...] [1] ovviamente proprietario, altrimenti il suo funzionamento (compreso l'algoritmo) sarebbe conoscibile... e ci sarebbero già diversi /fork/ per consentire alle persone di utilizzare i servizi come piace a loro (persone) e non a Loro (fornitori) [2] ovviamente /incluse/ le sue interfacce web (scarsamente configurabili, /estremamente/ mascheranti) e le sue API, neh! -- 380° (Giovanni Biscuolo public alter ego) «Noi, incompetenti come siamo, non abbiamo alcun titolo per suggerire alcunché» Disinformation flourishes because many people care deeply about injustice but very few check the facts. Ask me about <https://stallmansupport.org>.
participants (3)
-
380° -
Alberto Cammozzo -
Federico Leva (Nemo)