Discussion about this post

User's avatar
Eddy Borremans's avatar

Woah, this article basically covers almost every thinking I have been doing about AGI. I agree with a lot of what you say. The amateur philosopher (the one who didn't read up in detail about Yudkowski, nor on the difference between AGI and ASI) in me has a few things to share.

Unless Yudkowski has somehow mathematically proven there is some kind of universal morality that supersedes anything human-wrought, I think his views as you describe them might contradict themselves. If we talk about "greater good", shouldn't it be always about the greater good for mankind? How does eliminating mankind achieve that? Whichever way you look at it, this is all thought out by and for humans and I doubt whether we can know what is good for "the rest of the universe". That'd be quite arrogant. In short, eliminating mankind would also wipe out all our values and whatever is left would literally have no meaning at all. Well that was just the logical way of looking at it. The human in me, the one that is consciously aware he is subject to all kinds of human mechanism (emotion, social values, etc), 100% agrees about the atrociousness of his view.

As for AGI, again I agree with your estimate where we likely stand. And also with the dangers it poses. But I fear that the Djinn that is "the road to AGI" is out of the bottle. And I cannot conceive of a way to put it back. I can't help but think that the scenarios in which we might be able to put it back, may be too bleak to consider, even though pure logic dictate we should (compare how we deal with climate change). In that respect the future really looks bleak. Things are even bleaker when we look at the number of challenges we currently face, that shouldn't have been existential, but still are due to human stupidity: neglecting vulnerabilities of OpenClaw like architectures, sloppy software engineering practices in the very company that claims to be on the forefront of alignment science (Anthropic), government systems of powerful countries that don't have guardrails against lunatics with genocidal tendencies, BigTech that lie there asses off in the name of profit, climate change, and the list goes on. All problems that are theoretically much easier to solve than the existential problems caused by AGI/ASI.

Despite all of the above, I entertain a view of which I am still pondering whether it is naive (it probably is). AGI is dangerous for exactly the reasons you name. The alignment problem is mathematically unsolvable. The Djinn cannot be put back. So where does that put us? Well, imho, with the only skill humanity has time and time again shown that they can perform near miracles with: Engineering. However small the chances that Engineering will save us, it is the only option we have, so we MUST take it. Engineering what? Well a good enough solution to the alignment problem: heuristics, patches, guardrails, etc etc and probably a huge amount of luck. Again, it is not much to go on, but it is all we have. So we better get going and not only pray for, but build that miracle.

Tony Phillips's avatar

Congrats on the book... A year is a very long turnaround time maybe you could split the language editions or keep the rights for English and self publish but its very long...

projectallende@gmail.com

24 more comments...

No posts

Ready for more?