Discussion about this post

User's avatar
Ali Afroz's avatar

While I actually agree with you that AGI is probably not coming any time soon. I think you are premature in dismissing the current paradigm altogether as a way to achieve it. You admit that the current paradigm can allow an AI to imitate an agent, even if it doesn’t have any actual goals of its own. But fundamentally what we are concerned about is an AI acting like a super intelligent agent, so it doesn’t actually matter whether it has goals of its own only that it behaves as if it has these goals. If it is possible for the current paradigm to enable an AI to imitate a very smart human on a task, there is no reason to assume that eventually it would not be able to imitate someone much smarter than a human. Of course, there are constraints given the kind of data that you can easily feed an AI, but still it appears very premature to assume that it is fundamentally impossible given current techniques.

Your argument against orthogonality doesn’t really address the concerns of most people who don’t believe in moral realism. Sure from a present viewpoint, it appears as if things have been getting better in terms of peoples moral outlook, but firstly, even small deviations like whether or not to give animals moral weight often result in drastically different real world outcomes regarding what is the right thing to do. Secondly, in any case, of course, you would expect people in the recent past to be more similar to us in moral outlook, then people in the distant past. This doesn’t disprove the possibility that it’s just random variation instead of the World getting better due to superior moral understanding. In any case, even if morality is real and the AI understands it much better than us, you just don’t consider the possibility that you can know the right thing and still do evil. If an AI is programmed to maximise utility and conclude that because of animals suffering, it would be better to eliminate humans knowing that this is morally wrong isn’t really going to stop it from doing what it’s program to do, which is maximise utility not do the right thing. And what if the right thing goes against our interests. It seems not at all crazy to suppose that because of our biases we undervalue theories that suggest the right thing to do is something very bad for us humans. What if the right thing to do for example is maximise happiness, and this is best done by filling the universe with wireheaded tiny minds that each require very little matter. If an AI genuinely will be magically motivated to do the right thing, regardless of programming, if it’s smart enough, what’s stopping it in this scenario from doing the right thing, even when we obviously would not want it to do it.

Expand full comment
Leo Laforetakis's avatar

First exposure to you, really incredible writing what a pleasure. Thanks !

Expand full comment
28 more comments...

No posts