One of my 30 or so Friendliness-themed thought experiments is called “Implicit goals of ArgMax” or something like that. In general I think this style of reasoning is very important for accurately thinking about universal AI drives. Specifically it is important to analyze highly precise AI architectures like Goedel machines where there’s little wiggle room for a deus ex machina.
One of my 30 or so Friendliness-themed thought experiments is called “Implicit goals of ArgMax” or something like that. In general I think this style of reasoning is very important for accurately thinking about universal AI drives. Specifically it is important to analyze highly precise AI architectures like Goedel machines where there’s little wiggle room for a deus ex machina.