Here is another article about this, written by someone who says he is an AI
expert.

https://time.com/6266923/ai-eliezer-yudkowsky-open-letter-not-enough/

QUOTE:

Pausing AI Developments Isn't Enough. We Need to Shut it All Down

An open letter published today calls for “all AI labs to immediately pause
for at least 6 months the training of AI systems more powerful than GPT-4.”

This 6-month moratorium would be better than no moratorium. I have respect
for everyone who stepped up and signed it. It’s an improvement on the
margin. . . .

The key issue is not “human-competitive” intelligence (as the open letter
puts it); it’s what happens after AI gets to smarter-than-human
intelligence. Key thresholds there may not be obvious, we definitely can’t
calculate in advance what happens when, and it currently seems imaginable
that a research lab would cross critical lines without noticing.

Many researchers steeped in these issues, including myself, expect that the
most likely result of building a superhumanly smart AI, under anything
remotely like the current circumstances, is that literally everyone on
Earth will die. Not as in “maybe possibly some remote chance,” but as in
“that is the obvious thing that would happen.”

Reply via email to