a proposito dell'articolo di Chomsky sul NYT
Si tratta di questo articolo: https://www.nytimes.com/2023/03/08/opinion/noam-chomsky-chatgpt-ai.html (mi scuso qualora se ne fosse già discusso in lista) Ne stavo discutendo con colleghi di vari paesi in un'altra mailing list e in risposta ad alcune critiche fatte all'articolo ho scritto quanto riporto sotto (scusatemi se lo lascio in inglese ma siccome coinvolge l'uso di ChatGPT nell'analisi di un testo in inglese traducendolo diventava un'altra cosa) ... I find your viewpoint highly interesting, even if I somehow disagree on some of your criticism to Chomsky's article. I find an analogy between the way ChatGPT learned to master the language in comparison to human beings and how Alpha Zero learned to master board games in comparison to human beings. The deep-learning based approach exploited in both cases (almost) all available data, either through their availability on the Internet (in the case of the language) or through sophisticated systematic search through the strategy space (in the case of board games). Human beings become master (in both fields) leveraging a much smaller information space: the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used. I do not know of any estimation of this (which would be an interesting thing to do) but I would say the size is inferior of some orders of magnitude (which to me implies the "language learning algorithm" inside human brain is exponentially more efficient then those used in DL - but that's a different story). Moreover human beings know the "meaning" of their action and are aware of the fact that they know such meaning. (See the note at the end supporting my statements.) This appears to me to be an important difference between the two approaches and I think is the reason why Chomsky wrote that ChatGTP is not going to teach us anything useful about how human beings learn and use language. One more interesting discussion of the Chomsky viewpoint, this time defending it, I suggest (in case you've not already read it) is the piece of Gary Marcus https://garymarcus.substack.com/p/caricaturing-noam-chomsky Best, Enrico NOTE: To test my statements I asked ChatGPT to analyse a paragraph of my mail. You can read the prompt and the answer further below. The answer is at best confusing. ChatGPT fails to identify the text has 2 sentences and not just 1. It does a wrong analysis of the first sentence, which contain two main clauses and not one main and one subordinate. That's particularly interesting since right after the wrong analysis it "explains" that "a main clause is an independent clause that stands alone as a complete sentence". So, it writes that it "knows" what a main clause is right after having shown it doesn't know. And the rest contains a number of additional errors. So, even if the examples ChatGPT has been exposed to surely contain many descriptions of grammar rules, it has not been able to get any understanding of them. A secondary high school student, having received much less data and much less training is able to do a much better job. CVD :-) PROMPT Please perform the analysis of clause and sentence structure of the text enclosed below between «»: «Human beings become master (in both fields) leveraging a much smaller information space: the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used. I do not know of any estimation of this (which would be an interesting thing to do) but I would say the size is inferior of some orders of magnitude (which to me implies the "language learning algorithm" inside human brain is exponentially more efficient then those used in DL - but that's a different story).» ANSWER (ChatGPT release of Feb.13 - tested on Mar.14th) The text contains one sentence: "Human beings become master (in both fields) leveraging a much smaller information space: the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used. I do not know of any estimation of this (which would be an interesting thing to do) but I would say the size is inferior of some orders of magnitude (which to me implies the "language learning algorithm" inside human brain is exponentially more efficient then those used in DL - but that's a different story)." The sentence contains two clauses: Main clause: "Human beings become master leveraging a much smaller information space" Subordinate clause: "the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used" The main clause is an independent clause that stands alone as a complete sentence. It contains a subject ("human beings") and a predicate ("become master leveraging a much smaller information space"). The subordinate clause is a dependent clause that cannot stand alone as a complete sentence. It provides additional information to the main clause by describing the size of the information space that humans use to become masters in both fields. The subordinate clause contains a subject ("the number of examples a child is exposed to while learning language") and a predicate ("is much smaller than what ChatGPT used"). The phrase "just by considering time and space constraints" acts as an adverbial modifier that further clarifies the subordinate clause. The text also contains two embedded clauses: "which would be an interesting thing to do" is an embedded clause that acts as an object of the verb "know" in the main clause. "which to me implies the 'language learning algorithm' inside human brain is exponentially more efficient then those used in DL - but that's a different story" is an embedded clause that acts as the subject of the verb "is" in the subordinate clause. Additionally, the text contains a parenthetical remark: "but that's a different story," which is not syntactically connected to the rest of the sentence. END OF CHATGPT ANSWER -- -- EN https://www.hoepli.it/libro/la-rivoluzione-informatica/9788896069516.html ====================================================== Prof. Enrico Nardelli Presidente di "Informatics Europe" Direttore del Laboratorio Nazionale "Informatica e Scuola" del CINI Dipartimento di Matematica - Università di Roma "Tor Vergata" Via della Ricerca Scientifica snc - 00133 Roma home page: https://www.mat.uniroma2.it/~nardelli blog: https://link-and-think.blogspot.it/ tel: +39 06 7259.4204 fax: +39 06 7259.4699 mobile: +39 335 590.2331 e-mail: nardelli@mat.uniroma2.it online meeting: https://blue.meet.garr.it/b/enr-y7f-t0q-ont ====================================================== --
Ciao Enrico, le principali differenze tra il fanciullo e ChatGPT, di cui notiamo comunque essere il primo robot a raggiungere la performance di un 9nne sulla scala di teoria della mente (era 3enne solo a novembre scorso): 1. il consumo energetico. Il nostro cervello non arriva ad una lampadina, mentre per ogni richiesta di ChatGP consumiamo una quantita' immonda di energia. 2. la creazione. Il bambino crea la spiegazione, e poi la testa immediatamente, validandola o creadone un altra. ChatGPT parte dallo scibile disponibile per inferire e quindi imparare, ovviamente non crea mai nulla di effettivamente nuovo. 3. la plasticita' degli organoidi cerebrali rispetto ai chip sotto ChatGPT e l'utilizzo della chimica nei processi cerebrali. La gestione della "potatura" delle memorie inutili nel nostro cervello e' ordini di grandezza piu' efficiente ed efficace di quanto disponibile con AI. Su Zafferano di sabato Un abbraccio Roberto ________________________________ Da: nexa <nexa-bounces@server-nexa.polito.it> per conto di Enrico Nardelli <nardelli@mat.uniroma2.it> Inviato: martedì 14 marzo 2023 17:30:08 A: Nexa Oggetto: [nexa] a proposito dell'articolo di Chomsky sul NYT Si tratta di questo articolo: https://www.nytimes.com/2023/03/08/opinion/noam-chomsky-chatgpt-ai.html (mi scuso qualora se ne fosse già discusso in lista) Ne stavo discutendo con colleghi di vari paesi in un'altra mailing list e in risposta ad alcune critiche fatte all'articolo ho scritto quanto riporto sotto (scusatemi se lo lascio in inglese ma siccome coinvolge l'uso di ChatGPT nell'analisi di un testo in inglese traducendolo diventava un'altra cosa) ... I find your viewpoint highly interesting, even if I somehow disagree on some of your criticism to Chomsky's article. I find an analogy between the way ChatGPT learned to master the language in comparison to human beings and how Alpha Zero learned to master board games in comparison to human beings. The deep-learning based approach exploited in both cases (almost) all available data, either through their availability on the Internet (in the case of the language) or through sophisticated systematic search through the strategy space (in the case of board games). Human beings become master (in both fields) leveraging a much smaller information space: the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used. I do not know of any estimation of this (which would be an interesting thing to do) but I would say the size is inferior of some orders of magnitude (which to me implies the "language learning algorithm" inside human brain is exponentially more efficient then those used in DL - but that's a different story). Moreover human beings know the "meaning" of their action and are aware of the fact that they know such meaning. (See the note at the end supporting my statements.) This appears to me to be an important difference between the two approaches and I think is the reason why Chomsky wrote that ChatGTP is not going to teach us anything useful about how human beings learn and use language. One more interesting discussion of the Chomsky viewpoint, this time defending it, I suggest (in case you've not already read it) is the piece of Gary Marcus https://garymarcus.substack.com/p/caricaturing-noam-chomsky Best, Enrico NOTE: To test my statements I asked ChatGPT to analyse a paragraph of my mail. You can read the prompt and the answer further below. The answer is at best confusing. ChatGPT fails to identify the text has 2 sentences and not just 1. It does a wrong analysis of the first sentence, which contain two main clauses and not one main and one subordinate. That's particularly interesting since right after the wrong analysis it "explains" that "a main clause is an independent clause that stands alone as a complete sentence". So, it writes that it "knows" what a main clause is right after having shown it doesn't know. And the rest contains a number of additional errors. So, even if the examples ChatGPT has been exposed to surely contain many descriptions of grammar rules, it has not been able to get any understanding of them. A secondary high school student, having received much less data and much less training is able to do a much better job. CVD :-) PROMPT Please perform the analysis of clause and sentence structure of the text enclosed below between «»: «Human beings become master (in both fields) leveraging a much smaller information space: the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used. I do not know of any estimation of this (which would be an interesting thing to do) but I would say the size is inferior of some orders of magnitude (which to me implies the "language learning algorithm" inside human brain is exponentially more efficient then those used in DL - but that's a different story).» ANSWER (ChatGPT release of Feb.13 - tested on Mar.14th) The text contains one sentence: "Human beings become master (in both fields) leveraging a much smaller information space: the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used. I do not know of any estimation of this (which would be an interesting thing to do) but I would say the size is inferior of some orders of magnitude (which to me implies the "language learning algorithm" inside human brain is exponentially more efficient then those used in DL - but that's a different story)." The sentence contains two clauses: Main clause: "Human beings become master leveraging a much smaller information space" Subordinate clause: "the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used" The main clause is an independent clause that stands alone as a complete sentence. It contains a subject ("human beings") and a predicate ("become master leveraging a much smaller information space"). The subordinate clause is a dependent clause that cannot stand alone as a complete sentence. It provides additional information to the main clause by describing the size of the information space that humans use to become masters in both fields. The subordinate clause contains a subject ("the number of examples a child is exposed to while learning language") and a predicate ("is much smaller than what ChatGPT used"). The phrase "just by considering time and space constraints" acts as an adverbial modifier that further clarifies the subordinate clause. The text also contains two embedded clauses: "which would be an interesting thing to do" is an embedded clause that acts as an object of the verb "know" in the main clause. "which to me implies the 'language learning algorithm' inside human brain is exponentially more efficient then those used in DL - but that's a different story" is an embedded clause that acts as the subject of the verb "is" in the subordinate clause. Additionally, the text contains a parenthetical remark: "but that's a different story," which is not syntactically connected to the rest of the sentence. END OF CHATGPT ANSWER -- -- EN https://www.hoepli.it/libro/la-rivoluzione-informatica/9788896069516.html [cid:part1.QxOcBkKn.zM1wtphx@mat.uniroma2.it] ====================================================== Prof. Enrico Nardelli Presidente di "Informatics Europe" Direttore del Laboratorio Nazionale "Informatica e Scuola" del CINI Dipartimento di Matematica - Università di Roma "Tor Vergata" Via della Ricerca Scientifica snc - 00133 Roma home page: https://www.mat.uniroma2.it/~nardelli blog: https://link-and-think.blogspot.it/ tel: +39 06 7259.4204 fax: +39 06 7259.4699 mobile: +39 335 590.2331 e-mail: nardelli@mat.uniroma2.it<mailto:nardelli@mat.uniroma2.it> online meeting: https://blue.meet.garr.it/b/enr-y7f-t0q-ont ====================================================== --
Grazie del commento Rob. Io non sono un cognitivista quindi mi esprimo un'opinione consapevole che possa essere estremamente ingenua. Trovo il confronto con una teoria della mente molto fuorviante, dal momento che le menti che conosco sono menti che posseggono una semantica mentre ChatGPT possiede solo capacità sintattiche approssimando quelle semantiche mediante la statistica. Per non parlare poi degli aspetti di consapevolezza... Ciao, Enrico Per non parlare poi Il 15/03/2023 14:26, Roberto Dolci ha scritto:
Ciao Enrico,
le principali differenze tra il fanciullo e ChatGPT, di cui notiamo comunque essere il primo robot a raggiungere la performance di un 9nne sulla scala di teoria della mente (era 3enne solo a novembre scorso):
1. il consumo energetico. Il nostro cervello non arriva ad una lampadina, mentre per ogni richiesta di ChatGP consumiamo una quantita' immonda di energia.
2. la creazione. Il bambino crea la spiegazione, e poi la testa immediatamente, validandola o creadone un altra. ChatGPT parte dallo scibile disponibile per inferire e quindi imparare, ovviamente non crea mai nulla di effettivamente nuovo.
3. la plasticita' degli organoidi cerebrali rispetto ai chip sotto ChatGPT e l'utilizzo della chimica nei processi cerebrali. La gestione della "potatura" delle memorie inutili nel nostro cervello e' ordini di grandezza piu' efficiente ed efficace di quanto disponibile con AI. Su Zafferano di sabato
Un abbraccio Roberto ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ *Da:* nexa <nexa-bounces@server-nexa.polito.it> per conto di Enrico Nardelli <nardelli@mat.uniroma2.it> *Inviato:* martedì 14 marzo 2023 17:30:08 *A:* Nexa *Oggetto:* [nexa] a proposito dell'articolo di Chomsky sul NYT
Si tratta di questo articolo: https://www.nytimes.com/2023/03/08/opinion/noam-chomsky-chatgpt-ai.html (mi scuso qualora se ne fosse già discusso in lista)
Ne stavo discutendo con colleghi di vari paesi in un'altra mailing list e in risposta ad alcune critiche fatte all'articolo ho scritto quanto riporto sotto (scusatemi se lo lascio in inglese ma siccome coinvolge l'uso di ChatGPT nell'analisi di un testo in inglese traducendolo diventava un'altra cosa)
...
I find your viewpoint highly interesting, even if I somehow disagree on some of your criticism to Chomsky's article.
I find an analogy between the way ChatGPT learned to master the language in comparison to human beings and how Alpha Zero learned to master board games in comparison to human beings.
The deep-learning based approach exploited in both cases (almost) all available data, either through their availability on the Internet (in the case of the language) or through sophisticated systematic search through the strategy space (in the case of board games).
Human beings become master (in both fields) leveraging a much smaller information space: the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used. I do not know of any estimation of this (which would be an interesting thing to do) but I would say the size is inferior of some orders of magnitude (which to me implies the "language learning algorithm" inside human brain is exponentially more efficient then those used in DL - but that's a different story).
Moreover human beings know the "meaning" of their action and are aware of the fact that they know such meaning.
(See the note at the end supporting my statements.)
This appears to me to be an important difference between the two approaches and I think is the reason why Chomsky wrote that ChatGTP is not going to teach us anything useful about how human beings learn and use language.
One more interesting discussion of the Chomsky viewpoint, this time defending it, I suggest (in case you've not already read it) is the piece of Gary Marcus https://garymarcus.substack.com/p/caricaturing-noam-chomsky
Best, Enrico
NOTE: To test my statements I asked ChatGPT to analyse a paragraph of my mail. You can read the prompt and the answer further below.
The answer is at best confusing.
ChatGPT fails to identify the text has 2 sentences and not just 1.
It does a wrong analysis of the first sentence, which contain two main clauses and not one main and one subordinate.
That's particularly interesting since right after the wrong analysis it "explains" that "a main clause is an independent clause that stands alone as a complete sentence". So, it writes that it "knows" what a main clause is right after having shown it doesn't know.
And the rest contains a number of additional errors.
So, even if the examples ChatGPT has been exposed to surely contain many descriptions of grammar rules, it has not been able to get any understanding of them. A secondary high school student, having received much less data and much less training is able to do a much better job.
CVD :-)
PROMPT Please perform the analysis of clause and sentence structure of the text enclosed below between «»: «Human beings become master (in both fields) leveraging a much smaller information space: the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used. I do not know of any estimation of this (which would be an interesting thing to do) but I would say the size is inferior of some orders of magnitude (which to me implies the "language learning algorithm" inside human brain is exponentially more efficient then those used in DL - but that's a different story).»
ANSWER (ChatGPT release of Feb.13 - tested on Mar.14th) The text contains one sentence:
"Human beings become master (in both fields) leveraging a much smaller information space: the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used. I do not know of any estimation of this (which would be an interesting thing to do) but I would say the size is inferior of some orders of magnitude (which to me implies the "language learning algorithm" inside human brain is exponentially more efficient then those used in DL - but that's a different story)."
The sentence contains two clauses:
Main clause: "Human beings become master leveraging a much smaller information space" Subordinate clause: "the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used"
The main clause is an independent clause that stands alone as a complete sentence. It contains a subject ("human beings") and a predicate ("become master leveraging a much smaller information space").
The subordinate clause is a dependent clause that cannot stand alone as a complete sentence. It provides additional information to the main clause by describing the size of the information space that humans use to become masters in both fields. The subordinate clause contains a subject ("the number of examples a child is exposed to while learning language") and a predicate ("is much smaller than what ChatGPT used"). The phrase "just by considering time and space constraints" acts as an adverbial modifier that further clarifies the subordinate clause.
The text also contains two embedded clauses:
"which would be an interesting thing to do" is an embedded clause that acts as an object of the verb "know" in the main clause. "which to me implies the 'language learning algorithm' inside human brain is exponentially more efficient then those used in DL - but that's a different story" is an embedded clause that acts as the subject of the verb "is" in the subordinate clause.
Additionally, the text contains a parenthetical remark: "but that's a different story," which is not syntactically connected to the rest of the sentence.
END OF CHATGPT ANSWER
--
-- EN
https://www.hoepli.it/libro/la-rivoluzione-informatica/9788896069516.html ====================================================== Prof. Enrico Nardelli Presidente di "Informatics Europe" Direttore del Laboratorio Nazionale "Informatica e Scuola" del CINI Dipartimento di Matematica - Università di Roma "Tor Vergata" Via della Ricerca Scientifica snc - 00133 Roma home page: https://www.mat.uniroma2.it/~nardelli blog: https://link-and-think.blogspot.it/ tel: +39 06 7259.4204 fax: +39 06 7259.4699 mobile: +39 335 590.2331 e-mail: nardelli@mat.uniroma2.it online meeting: https://blue.meet.garr.it/b/enr-y7f-t0q-ont ======================================================
--
-- -- EN https://www.hoepli.it/libro/la-rivoluzione-informatica/9788896069516.html ====================================================== Prof. Enrico Nardelli Presidente di "Informatics Europe" Direttore del Laboratorio Nazionale "Informatica e Scuola" del CINI Dipartimento di Matematica - Università di Roma "Tor Vergata" Via della Ricerca Scientifica snc - 00133 Roma home page: https://www.mat.uniroma2.it/~nardelli blog: https://link-and-think.blogspot.it/ tel: +39 06 7259.4204 fax: +39 06 7259.4699 mobile: +39 335 590.2331 e-mail: nardelli@mat.uniroma2.it online meeting: https://blue.meet.garr.it/b/enr-y7f-t0q-ont ====================================================== --
Ciao Enrico, hai perfettamente ragione, il test della teoria della mente e' un approssimazione per vedere - in modo standard e ripetibile - quale grado di formazione di coscienza abbiamo davanti, qui il pre-print https://arxiv.org/abs/2302.02083. Pero' mira a guardare la capacita' di capire il prossimo dal contesto, che e' una delle abilita' che acquisiamo da bimbi fino all'eta' adulta e che ci consente di sviluppare bias/stereotipi/pregiudizi che accelerano il nostro processo stimulo-analisi-reazione. Non butterei via il bimbo con l'acqua sporca, perche' specialmente il miglioramento da 3 a 9 anni in pochi mesi ci da misura della capacita' di inferenza che questi modelli sviluppano. Resta interessante 😎 A presto Rob ________________________________ Da: nexa <nexa-bounces@server-nexa.polito.it> per conto di Enrico Nardelli <nardelli@mat.uniroma2.it> Inviato: mercoledì 15 marzo 2023 14:10:45 A: Nexa Oggetto: Re: [nexa] a proposito dell'articolo di Chomsky sul NYT Grazie del commento Rob. Io non sono un cognitivista quindi mi esprimo un'opinione consapevole che possa essere estremamente ingenua. Trovo il confronto con una teoria della mente molto fuorviante, dal momento che le menti che conosco sono menti che posseggono una semantica mentre ChatGPT possiede solo capacità sintattiche approssimando quelle semantiche mediante la statistica. Per non parlare poi degli aspetti di consapevolezza... Ciao, Enrico Per non parlare poi Il 15/03/2023 14:26, Roberto Dolci ha scritto: Ciao Enrico, le principali differenze tra il fanciullo e ChatGPT, di cui notiamo comunque essere il primo robot a raggiungere la performance di un 9nne sulla scala di teoria della mente (era 3enne solo a novembre scorso): 1. il consumo energetico. Il nostro cervello non arriva ad una lampadina, mentre per ogni richiesta di ChatGP consumiamo una quantita' immonda di energia. 2. la creazione. Il bambino crea la spiegazione, e poi la testa immediatamente, validandola o creadone un altra. ChatGPT parte dallo scibile disponibile per inferire e quindi imparare, ovviamente non crea mai nulla di effettivamente nuovo. 3. la plasticita' degli organoidi cerebrali rispetto ai chip sotto ChatGPT e l'utilizzo della chimica nei processi cerebrali. La gestione della "potatura" delle memorie inutili nel nostro cervello e' ordini di grandezza piu' efficiente ed efficace di quanto disponibile con AI. Su Zafferano di sabato Un abbraccio Roberto ________________________________ Da: nexa <nexa-bounces@server-nexa.polito.it><mailto:nexa-bounces@server-nexa.polito.it> per conto di Enrico Nardelli <nardelli@mat.uniroma2.it><mailto:nardelli@mat.uniroma2.it> Inviato: martedì 14 marzo 2023 17:30:08 A: Nexa Oggetto: [nexa] a proposito dell'articolo di Chomsky sul NYT Si tratta di questo articolo: https://www.nytimes.com/2023/03/08/opinion/noam-chomsky-chatgpt-ai.html (mi scuso qualora se ne fosse già discusso in lista) Ne stavo discutendo con colleghi di vari paesi in un'altra mailing list e in risposta ad alcune critiche fatte all'articolo ho scritto quanto riporto sotto (scusatemi se lo lascio in inglese ma siccome coinvolge l'uso di ChatGPT nell'analisi di un testo in inglese traducendolo diventava un'altra cosa) ... I find your viewpoint highly interesting, even if I somehow disagree on some of your criticism to Chomsky's article. I find an analogy between the way ChatGPT learned to master the language in comparison to human beings and how Alpha Zero learned to master board games in comparison to human beings. The deep-learning based approach exploited in both cases (almost) all available data, either through their availability on the Internet (in the case of the language) or through sophisticated systematic search through the strategy space (in the case of board games). Human beings become master (in both fields) leveraging a much smaller information space: the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used. I do not know of any estimation of this (which would be an interesting thing to do) but I would say the size is inferior of some orders of magnitude (which to me implies the "language learning algorithm" inside human brain is exponentially more efficient then those used in DL - but that's a different story). Moreover human beings know the "meaning" of their action and are aware of the fact that they know such meaning. (See the note at the end supporting my statements.) This appears to me to be an important difference between the two approaches and I think is the reason why Chomsky wrote that ChatGTP is not going to teach us anything useful about how human beings learn and use language. One more interesting discussion of the Chomsky viewpoint, this time defending it, I suggest (in case you've not already read it) is the piece of Gary Marcus https://garymarcus.substack.com/p/caricaturing-noam-chomsky Best, Enrico NOTE: To test my statements I asked ChatGPT to analyse a paragraph of my mail. You can read the prompt and the answer further below. The answer is at best confusing. ChatGPT fails to identify the text has 2 sentences and not just 1. It does a wrong analysis of the first sentence, which contain two main clauses and not one main and one subordinate. That's particularly interesting since right after the wrong analysis it "explains" that "a main clause is an independent clause that stands alone as a complete sentence". So, it writes that it "knows" what a main clause is right after having shown it doesn't know. And the rest contains a number of additional errors. So, even if the examples ChatGPT has been exposed to surely contain many descriptions of grammar rules, it has not been able to get any understanding of them. A secondary high school student, having received much less data and much less training is able to do a much better job. CVD :-) PROMPT Please perform the analysis of clause and sentence structure of the text enclosed below between «»: «Human beings become master (in both fields) leveraging a much smaller information space: the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used. I do not know of any estimation of this (which would be an interesting thing to do) but I would say the size is inferior of some orders of magnitude (which to me implies the "language learning algorithm" inside human brain is exponentially more efficient then those used in DL - but that's a different story).» ANSWER (ChatGPT release of Feb.13 - tested on Mar.14th) The text contains one sentence: "Human beings become master (in both fields) leveraging a much smaller information space: the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used. I do not know of any estimation of this (which would be an interesting thing to do) but I would say the size is inferior of some orders of magnitude (which to me implies the "language learning algorithm" inside human brain is exponentially more efficient then those used in DL - but that's a different story)." The sentence contains two clauses: Main clause: "Human beings become master leveraging a much smaller information space" Subordinate clause: "the number of examples a child is exposed to while learning language is, just by considering time and space constraints, much smaller than what ChatGPT used" The main clause is an independent clause that stands alone as a complete sentence. It contains a subject ("human beings") and a predicate ("become master leveraging a much smaller information space"). The subordinate clause is a dependent clause that cannot stand alone as a complete sentence. It provides additional information to the main clause by describing the size of the information space that humans use to become masters in both fields. The subordinate clause contains a subject ("the number of examples a child is exposed to while learning language") and a predicate ("is much smaller than what ChatGPT used"). The phrase "just by considering time and space constraints" acts as an adverbial modifier that further clarifies the subordinate clause. The text also contains two embedded clauses: "which would be an interesting thing to do" is an embedded clause that acts as an object of the verb "know" in the main clause. "which to me implies the 'language learning algorithm' inside human brain is exponentially more efficient then those used in DL - but that's a different story" is an embedded clause that acts as the subject of the verb "is" in the subordinate clause. Additionally, the text contains a parenthetical remark: "but that's a different story," which is not syntactically connected to the rest of the sentence. END OF CHATGPT ANSWER -- -- EN https://www.hoepli.it/libro/la-rivoluzione-informatica/9788896069516.html [cid:part1.N631ls4Z.ADzWC5so@mat.uniroma2.it] ====================================================== Prof. Enrico Nardelli Presidente di "Informatics Europe" Direttore del Laboratorio Nazionale "Informatica e Scuola" del CINI Dipartimento di Matematica - Università di Roma "Tor Vergata" Via della Ricerca Scientifica snc - 00133 Roma home page: https://www.mat.uniroma2.it/~nardelli blog: https://link-and-think.blogspot.it/ tel: +39 06 7259.4204 fax: +39 06 7259.4699 mobile: +39 335 590.2331 e-mail: nardelli@mat.uniroma2.it<mailto:nardelli@mat.uniroma2.it> online meeting: https://blue.meet.garr.it/b/enr-y7f-t0q-ont ====================================================== -- -- -- EN https://www.hoepli.it/libro/la-rivoluzione-informatica/9788896069516.html [cid:part1.N631ls4Z.ADzWC5so@mat.uniroma2.it] ====================================================== Prof. Enrico Nardelli Presidente di "Informatics Europe" Direttore del Laboratorio Nazionale "Informatica e Scuola" del CINI Dipartimento di Matematica - Università di Roma "Tor Vergata" Via della Ricerca Scientifica snc - 00133 Roma home page: https://www.mat.uniroma2.it/~nardelli blog: https://link-and-think.blogspot.it/ tel: +39 06 7259.4204 fax: +39 06 7259.4699 mobile: +39 335 590.2331 e-mail: nardelli@mat.uniroma2.it<mailto:nardelli@mat.uniroma2.it> online meeting: https://blue.meet.garr.it/b/enr-y7f-t0q-ont ====================================================== --
- GPT-4 has the potential to be used to attempt to identify private individuals when augmented with outside data; - GPT-4 is capable of generating discriminatory content favorable to autocratic governments across multiple languages - it does continue the trend of potentially lowering the cost of certain steps of a successful cyberattack; - GPT-4 has the tendency to “hallucinate,” i.e. “produce content that is nonsensical or untruthful in relation to certain sources.” This tendency can be particularly harmful as models become increasingly convincing and believable, leading to overreliance on them by users. Counterintuitively, hallucinations can become more dangerous as models become more truthful, as users build trust in the model when it provides truthful information in areas where they have some familiarity. Additionally, as these models are integrated into society and used to help automate various systems, this tendency to hallucinate is one of the factors that can lead to the degradation of overall information quality and further reduce veracity of and trust in freely available information; Non lo scrive un luddista, lo scrive - in un lungo elenco di ulteriori effetti nefasti - OpenAI, nella System Card di un documento travestito da articolo arxiv, ma il cui autore è l'azienda stessa: <https://cdn.openai.com/papers/gpt-4.pdf> Nella consapevolezza della fragilità degli interventi mitigatori, OpenAI rileva "a need for effective regulation". Buona serata, Daniela
Possiamo chiamarlo "L'approccio dei Sam" (Altman e Bankman-Fried): non rispettare le regole e poi dire "ma com'è che non mi regolate?" ----- Original Message ----- From: "Daniela Tafani" <daniela.tafani@unipi.it> To: "Nexa" <nexa@server-nexa.polito.it> Sent: Wednesday, March 15, 2023 7:11:09 PM Subject: [nexa] GPT-4 - GPT-4 has the potential to be used to attempt to identify private individuals when augmented with outside data; - GPT-4 is capable of generating discriminatory content favorable to autocratic governments across multiple languages - it does continue the trend of potentially lowering the cost of certain steps of a successful cyberattack; - GPT-4 has the tendency to “hallucinate,” i.e. “produce content that is nonsensical or untruthful in relation to certain sources.” This tendency can be particularly harmful as models become increasingly convincing and believable, leading to overreliance on them by users. Counterintuitively, hallucinations can become more dangerous as models become more truthful, as users build trust in the model when it provides truthful information in areas where they have some familiarity. Additionally, as these models are integrated into society and used to help automate various systems, this tendency to hallucinate is one of the factors that can lead to the degradation of overall information quality and further reduce veracity of and trust in freely available information; Non lo scrive un luddista, lo scrive - in un lungo elenco di ulteriori effetti nefasti - OpenAI, nella System Card di un documento travestito da articolo arxiv, ma il cui autore è l'azienda stessa: < [ https://cdn.openai.com/papers/gpt-4.pdf | https://cdn.openai.com/papers/gpt-4.pdf ] > Nella consapevolezza della fragilità degli interventi mitigatori, OpenAI rileva "a need for effective regulation". Buona serata, Daniela _______________________________________________ nexa mailing list nexa@server-nexa.polito.it https://server-nexa.polito.it/cgi-bin/mailman/listinfo/nexa
Già. O dei multimiliardari che chiedono di pagare più tasse. Qui però c'è anche il convitato di pietra della responsabilità: assomigliano perciò al quel personaggio di Pirandello che girava sempre con le mani avanti... ________________________________ Da: Antonio Casilli <antonio.casilli@telecom-paris.fr> Inviato: mercoledì 15 marzo 2023 19:45 A: Daniela Tafani Cc: Nexa Oggetto: Re: [nexa] GPT-4 Possiamo chiamarlo "L'approccio dei Sam" (Altman e Bankman-Fried): non rispettare le regole e poi dire "ma com'è che non mi regolate?" ----- Original Message ----- From: "Daniela Tafani" <daniela.tafani@unipi.it> To: "Nexa" <nexa@server-nexa.polito.it> Sent: Wednesday, March 15, 2023 7:11:09 PM Subject: [nexa] GPT-4 - GPT-4 has the potential to be used to attempt to identify private individuals when augmented with outside data; - GPT-4 is capable of generating discriminatory content favorable to autocratic governments across multiple languages - it does continue the trend of potentially lowering the cost of certain steps of a successful cyberattack; - GPT-4 has the tendency to “hallucinate,” i.e. “produce content that is nonsensical or untruthful in relation to certain sources.” This tendency can be particularly harmful as models become increasingly convincing and believable, leading to overreliance on them by users. Counterintuitively, hallucinations can become more dangerous as models become more truthful, as users build trust in the model when it provides truthful information in areas where they have some familiarity. Additionally, as these models are integrated into society and used to help automate various systems, this tendency to hallucinate is one of the factors that can lead to the degradation of overall information quality and further reduce veracity of and trust in freely available information; Non lo scrive un luddista, lo scrive - in un lungo elenco di ulteriori effetti nefasti - OpenAI, nella System Card di un documento travestito da articolo arxiv, ma il cui autore è l'azienda stessa: < [ https://es.sonicurlprotection-fra.com/click?PV=2&MSGID=202303151845100105620... | https://es.sonicurlprotection-fra.com/click?PV=2&MSGID=202303151845100105620... ] > Nella consapevolezza della fragilità degli interventi mitigatori, OpenAI rileva "a need for effective regulation". Buona serata, Daniela _______________________________________________ nexa mailing list nexa@server-nexa.polito.it https://es.sonicurlprotection-fra.com/click?PV=2&MSGID=202303151845100105620...
D'altra parte, qualcuno ha mai smesso di fumare perché sul pacchetto di sigarette c'è scritto che il fumo uccide? La verità è che tutti finiremo per usare GPT e le tante cose del genere che stanno già venendo fuori. Vi sono moltissimi impieghi per i quali l'inconsapevolezza di questi sistemi non emerge, o non rileva. Nell'industria del software, ad esempio, GPT accelera moltissimo quello che gli sviluppatori fanno abitualmente quando sono alle prese con un puzzle (cioè sempre): ricerche su Google, che puntualmente conducono a siti specialistici dove tuttavia c'è da perder tempo. GPT, sulla base di informazione attinta da quelle stesse fonti, dà in pochi secondi la soluzione 'out of the box', quasi sempre soddisfacente. Nessuno sviluppatore ne farà a meno. L'effetto devastante per l'ecosistema dello sviluppo software è facilmente intuibile: verrà tutto risucchiato dentro i 'gazillion parameters' di GPT. Ma vi sono anche cose molto interessanti che riguardano le parole. La lessicografia classica, cioè basata fondamentalmente sull'introspezione, già da tempo incalzata dalle analisi quantitative (il lessico di frequenza di De Mauro risale agli anni '80) è finita. I Large Language Model non sono certamente un modello plausibile del linguaggio umano, tuttavia rendono possibile osservare, nei 'giochi linguistici', quelle 'somiglianze di famiglia' di cui parlava Wittgenstein, su una scala che non ha precedenti. Risolta la questione se useremo o meno questi oggetti, la domanda ora è: possiamo costruirli e gestirli in modo che non aumentino le disuguaglianze e i pericoli? Uno spunto può venire dal confronto tra i diversi ordini di grandezza che dividono GPT-3 da GPT-4 e le differenze non altrettanto grandi tra le prestazioni dei due sistemi, come se il rapporto tra dimensioni e prestazioni fosse logaritmico. Se così fosse, allora sapremmo qual è il limite del reale vantaggio dei GAFAM rispetto a 'the rest of us'. Viviamo in tempi interessanti. Guido On Wed, 15 Mar 2023 at 20:01, Daniela Tafani <daniela.tafani@unipi.it> wrote:
Già. O dei multimiliardari che chiedono di pagare più tasse.
Qui però c'è anche il convitato di pietra della responsabilità:
assomigliano perciò al quel personaggio di Pirandello che girava sempre con le mani avanti...
------------------------------ *Da:* Antonio Casilli <antonio.casilli@telecom-paris.fr> *Inviato:* mercoledì 15 marzo 2023 19:45 *A:* Daniela Tafani *Cc:* Nexa *Oggetto:* Re: [nexa] GPT-4
Possiamo chiamarlo "L'approccio dei Sam" (Altman e Bankman-Fried): non rispettare le regole e poi dire "ma com'è che non mi regolate?"
----- Original Message ----- From: "Daniela Tafani" <daniela.tafani@unipi.it> To: "Nexa" <nexa@server-nexa.polito.it> Sent: Wednesday, March 15, 2023 7:11:09 PM Subject: [nexa] GPT-4
- GPT-4 has the potential to be used to attempt to identify private individuals when augmented with outside data;
- GPT-4 is capable of generating discriminatory content favorable to autocratic governments across multiple languages
- it does continue the trend of potentially lowering the cost of certain steps of a successful cyberattack;
- GPT-4 has the tendency to “hallucinate,” i.e. “produce content that is nonsensical or untruthful in relation to certain sources.” This tendency can be particularly harmful as models become increasingly convincing and believable, leading to overreliance on them by users. Counterintuitively, hallucinations can become more dangerous as models become more truthful, as users build trust in the model when it provides truthful information in areas where they have some familiarity. Additionally, as these models are integrated into society and used to help automate various systems, this tendency to hallucinate is one of the factors that can lead to the degradation of overall information quality and further reduce veracity of and trust in freely available information;
Non lo scrive un luddista, lo scrive - in un lungo elenco di ulteriori effetti nefasti - OpenAI, nella System Card di un documento travestito da articolo arxiv, ma il cui autore è l'azienda stessa:
< [ https://es.sonicurlprotection-fra.com/click?PV=2&MSGID=202303151845100105620... | https://es.sonicurlprotection-fra.com/click?PV=2&MSGID=202303151845100105620... ] >
Nella consapevolezza della fragilità degli interventi mitigatori, OpenAI rileva "a need for effective regulation".
Buona serata,
Daniela
_______________________________________________ nexa mailing list nexa@server-nexa.polito.it
https://es.sonicurlprotection-fra.com/click?PV=2&MSGID=202303151845100105620... _______________________________________________ nexa mailing list nexa@server-nexa.polito.it https://server-nexa.polito.it/cgi-bin/mailman/listinfo/nexa
participants (5)
-
Antonio Casilli -
Daniela Tafani -
Enrico Nardelli -
Guido Vetere -
Roberto Dolci