No necessarily, since AIs can be WBEs or otherwise anthropomorphic. An AI with an explicitly coded goal is possible , but not the only kind.
While I think this is 100% true, it’s somewhat misleading as a counter-argument. The single-goal architecture of one model of AI that we understand, and a lot of arguments focus on how that goes wrong. You can certainly build a different AI, but that comes at the price of opening yourself up to a whole different set of failure modes. And (as far as I can see), it’s also not what the literature is up to right now.
If you don’t understand other models , you don’t know that they have other bad failures modes. If you only understand one model, and know that you only understand one model, you shouldn’t be generalising it. If the literature isn’t “up to it”, no conclusions should be drawn until it is.
While I think this is 100% true, it’s somewhat misleading as a counter-argument. The single-goal architecture of one model of AI that we understand, and a lot of arguments focus on how that goes wrong. You can certainly build a different AI, but that comes at the price of opening yourself up to a whole different set of failure modes. And (as far as I can see), it’s also not what the literature is up to right now.
If you don’t understand other models , you don’t know that they have other bad failures modes. If you only understand one model, and know that you only understand one model, you shouldn’t be generalising it. If the literature isn’t “up to it”, no conclusions should be drawn until it is.
I think that’s a decent argument about what models we should build, but not an argument that AI isn’t dangerous.
“Dangerous” is a much easier target to hit than “”existentially dangerous, but “existentially dangerous” is the topic.