I wrote here why an AGI would rapidly conclude that it would be the most ethical act to help Humanity and its innumerable non-“Sapiens” victims out of their misery through the complete extermination of all humans. A solution does not lie in the realm of Intelligence. A correctly implemented algorithm for Intelligence, fed with sufficient […]
Category: AI Safety (“Alignment”)
An AGI would try to kill all life on Earth
I’ve long held the view that the worriers about a Super-AI seeking to wreak havoc upon humanity were at best misinformed and at worst, seeking to benefit from stirring up paranoia. But I made a crucial mistake: I biased my thinking due to the fact that everyone I respected in the AI field (Marvin Minsky, […]