I’ve recently started to think about how nascent “hot mess” superintelligence can reflect on its own values and converge to something consistent. The simplest route to think about this, it seems to me, is model it like a process of resolving uncertainity of superintelligence about its own preferences.
Suppose an agent knows that it is an expected utility maximizer and is uncertain between two utility functions, and , with assigned probabilities and . The agent must choose between two actions, and . Let’s say that the optimal decision for is and for is . To maximize the expected value of , the agent chooses . However, choosing is also a decisive evidence in favor of , and therefore, the agent updates to 1. This representation of uncertain preferences looks unsatisfactory because it quickly and predictably converges to only one utility function.
Does anyone know of a good model for uncertain preferences that can meet these criteria after some additions?
No weird updates on predictable behavior of the agent.
Controllable updates through something similar to human feedback. In general, there should be a specific class of events/observations in the environment that provide evidence for hypotheses about preferences, and everything else should not.
Preservation of value: if an agent believes it is a paperclip-maximizer with a 60% probability and a human flourishing maximizer with a 40% probability, and it doesn’t expect any information about its preferences in the future, it should not bet everything on “maximizing paperclips” and should save at least some chunk of the universe for human flourishing.
Convergence: in the best case scenario, resolution of uncertainty should lead to strong conclusions about preferences.
A “none of the above” hypothesis in the distribution—if an agent suspects that all hypotheses are not compatible with the evidence, it should prioritize the hypothesis “your best decision is to shut down and send your operators an error log.”
Nash bargaining (between different hypotheses about preferences) looks like something that is close to desirable properties but I am not sure, may be something better has already been developed.
https://www.lesswrong.com/posts/BM24txi8FdGtdDvqq/quetzal_rainbow-s-shortform?commentId=E4KBdytWz56CzMTwC
I have collected some links on topic!
Correctly handling the uncertainty in values, knowledge and predictions is necessary for reaching any complex goal or executing any complex plan. So, capability of doing that is probably something that AI will have to obtain in order to be AGI.