AI: requirements for pernicious policies
Some have argued that “tool AIs” are safe(r). Recently, Eric Drexler decomposed AIs into “problem solvers” (eg calculators), “advisors” (eg GPS route planners), and actors (autonomous agents). Both solvers and advisors can be seen as examples of tools.
People have argued that tool AIs are not safe. It’s hard to imagine a calculator going berserk, no matter what its algorithm is, but it’s not too hard to come up with clear examples of dangerous tools. This suggests the solvers vs advisors vs actors (or tools vs agents, or oracles vs agents) is not the right distinction.
Instead, I’ve been asking: how likely is the algorithm to implement a pernicious policy? If we model the AI as having an objective function (or utility function) and algorithm that implements it, a pernicious policy is one that scores high in the objective function but is not at all what is intended. A pernicious function could be harmless and entertaining or much more severe.
I will lay aside, for the moment, the issue of badly programmed algorithms (possibly containing its own objective sub-functions). In any case, to implement a pernicious function, we have to ask these questions about the algorithm:
Do pernicious policies exist? Are there many?
Can the AI find them?
Can the AI test them?
Would the AI choose to implement them?
The answer to 1. seems to be trivially yes. Even a calculator could, in theory, output a series of messages that socially hack us, blah, take over the world, blah, extinction, blah, calculator finishes its calculations. What is much more interesting is some types of agents have many more pernicious policies than others. This seems the big difference between actors and other designs. An actor AI in complete control of the USA or Russia’s nuclear arsenal has all sort of pernicious policies easily to hand; an advisor or oracle has much fewer (generally going through social engineering), a tool typically even less. A lot of the physical protection measures are about reducing the number of sucessfull pernicious policies the AI has a cess to.
The answer to 2. is mainly a function of the power of the algorithm. A basic calculator will never find anything dangerous: its programming is simple and tight. But compare an agent with the same objective function and the ability to do an unrestricted policy search with vast resources… So it seems that the answer to 2. does not depend on any solver vs actor division, but purely on the algorithm used.
And now we come to the big question 3., whether the AI can test these policies. Even if the AI can find pernicious policies that rank high on its objective function, it will never implement them unless it can ascertain this fact. And there are several ways it could do so. Let’s assume that a solver AI has a very complicated objective function—one that encodes many relevant facts about the real world. Now, the AI may not “care” about the real world, but it has a virtual version of that, in which it can virtually test all of its policies. With a detailed enough computing power, it can establish whether the pernicious policy would be effective at achieving its virtual goal. If this is a good approximation of how the pernicious policy would behave in the real world, we could have a problem.
But extremely detailed objective functions are unlikely. But even simple ones can show odd behaviour if the agents gets to interact repeatedly with the real world—this is the issue with reinforcement learning. Suppose that the agent attempts a translation job, and is rewarded on the accuracy of its translation. Depending on the details of what the AI knows and who choose the rewards, the AI could end up manipulating its controllers, similarly to this example. The problem is that one there is any interaction, all the complexity of humanity could potentially show up in the reward function, even if the objective function is simple.
Of course, some designs make this very unlikely—resetting the AI periodically can help to alleviate the problem, as can choosing more objective criteria for any rewards. Lastly on this point, we should mention the possibility that human R&D, by selecting and refining the objective function and the algorithm, could take the roll of testing the policies. This is likely to emerge only in cases where many AI designs are considered, and the best candiates are retained based on human judgement.
Finally we come to the question of whether the AI will implement the policy if it’s found it and tested it. You could say that the point of FAI is to create an AI that doesn’t choose to implement pernicious policies—but, more correctly, the point of FAI is to ensure that very few (or zero) pernicious policies exist in the first place, as they all score low on the utility function. However, there are a variety of more complicated designs—satisficers, agents using crude measures—where the questions of “Do pernicious policies exist?” and “Would the AI choose to implement them?” could become quite distinct.
Conclusion: a more through analysis of AI designs is needed
A calculator is safe, because it is a solver, it has a very simple objective function, with no holes in the algorithm, and it can neither find nor test any pernicious policies. It is the combination of these elements that makes it almost certainly safe. If we want to make the same claim about other designs, neither “it’s just a solver” or “it’s objective function is simple” would be enough; we need a careful analysis.
Though, as usual, “it’s not certainly safe” is a quite distinct claim from “it’s (likely) dangerous”, and they should not be conflated.
I think tool AIs are very dangerous because they allow the application of undigested knowledge
I don’t know if the AI should be taking responsibility for testing its own policies, especially in the initial stages. We should have a range of tests that humans apply that the formative AI runs on each iteration so that we can see how it is progressing.
“testing” means establishing that they are high ranked in the objective function. An algorithm has to be able to do that, somehow, or there’s no point in having an objective function.