Constant [sorry for getting the attribution wrong in my previous reply] wrote:
We do not know very well how the human mind does anything at all. But that the the human mind comes to have preferences that it did not have initially, cannot be doubted.
I do not know whether those changes in opinion indicate changes in terminal values, but it doesn’t really matter for the purposes of this discussion, since humans aren’t (capital-F) Friendly. You definitely don’t want an FAI to unpredictably change its terminal values. Figuring out how to reliably prevent this kind of thing from happening, even in a strongly self-modifying mind (which humans aren’t), is one of the sub-problems of the FAI problem.
To create a society of AIs, hoping they’ll prevent each other from doing too much damage, isn’t a viable solution to the FAI problem, even in the rudimentary “doesn’t kill all humans” sense. There’s various problems with the idea, among them:
Any two AIs are likely to have a much vaster difference in effective intelligence than you could ever find between two humans (for one thing, their hardware might be much more different than any two working human brains). This likelihood increases further if (at least) some subset of them is capable of strong self-improvement. With enough difference in power, cooperation becomes a losing strategy for the more powerful party.
The AIs might agree that they’d all be better off if they took the matter currently in use by humans for themselves, dividing the spoils among each other.
Constant [sorry for getting the attribution wrong in my previous reply] wrote:
I do not know whether those changes in opinion indicate changes in terminal values, but it doesn’t really matter for the purposes of this discussion, since humans aren’t (capital-F) Friendly. You definitely don’t want an FAI to unpredictably change its terminal values. Figuring out how to reliably prevent this kind of thing from happening, even in a strongly self-modifying mind (which humans aren’t), is one of the sub-problems of the FAI problem.To create a society of AIs, hoping they’ll prevent each other from doing too much damage, isn’t a viable solution to the FAI problem, even in the rudimentary “doesn’t kill all humans” sense. There’s various problems with the idea, among them:
Any two AIs are likely to have a much vaster difference in effective intelligence than you could ever find between two humans (for one thing, their hardware might be much more different than any two working human brains). This likelihood increases further if (at least) some subset of them is capable of strong self-improvement. With enough difference in power, cooperation becomes a losing strategy for the more powerful party.
The AIs might agree that they’d all be better off if they took the matter currently in use by humans for themselves, dividing the spoils among each other.