Toy model of human values

This is just a summary via analogy where I human values come from, as far as I understand it. The expanded version would be Eli’s http://​​lesswrong.com/​​lw/​​l3/​​thou_art_godshatter/​​.

The basic analogy is to chess-playing programs (at least the basic ones from 40 years ago, the art has progressed since then, but not much). The way they work is basically by examining the branching tree of possible moves; since chess is “too big” to solve completely (find the branch that always leads to winning) by present hardware what these programs do is go to a certain depth and then use heuristics to decide whether the end state is good, such as how many pieces are on its side vs. the enemy side, weighed by their “power” (queen is worth more than pawn) and position (center positions are worth more).

The analogy mapping is as follows: the goal of the game is winning, of evolution is survival of a gene fragment (such as human DNA). Explicit encoding of the goal is not computationally feasible or worthwhile (in terms of the goal itself), so values of certain non-terminal states (in terms of the goal) are explicitly given to the program or to a human; the human/​program knows no better than these non-terminal values—they are our values—we are Godshatter.

What do you think?