Pausing AI Developments Isn't Enough. We Need to Shut it All Down
BY ELIEZER YUDKOWSKY
MARCH 29, 2023 6:01 PM EDT
Yudkowsky is a decision theorist from the U.S. and leads research
at the Machine Intelligence Research Institute. He's been working
on aligning Artificial General Intelligence since 2001 and is
widely regarded as a founder of the field.
An open letter published today calls for “all AI labs to immediately
pause for at least 6 months the training of AI systems more powerful
than GPT-4.”
This 6-month moratorium would be better than no moratorium. I have
respect for everyone who stepped up and signed it. It’s an
improvement on the margin.
I refrained from signing because I think the letter is understating
the seriousness of the situation and asking for too little to solve
it.
The key issue is not “human-competitive” intelligence (as the open
letter puts it); it’s what happens after AI gets to
smarter-than-human intelligence. Key thresholds there may not be
obvious, we definitely can’t calculate in advance what happens when,
and it currently seems imaginable that a research lab would cross
critical lines without noticing.
continua qui:
https://time.com/6266923/ai-eliezer-yudkowsky-open-letter-not-enough/