I don’t understand why it would be—it looks like MENACE is just a simple physical algorithm that successfully optimizes for winning tic-tac-toe. I thought the idea of an OD was that a process optimizing for goal A hard enough could produce a consequentialist* agent that cares about a different goal B. What is the goal B here (or am I misunderstanding the concept)?
Would you consider a MENACE Tic-tac-toe matchbox-based optimizer an OD?
I don’t understand why it would be—it looks like MENACE is just a simple physical algorithm that successfully optimizes for winning tic-tac-toe. I thought the idea of an OD was that a process optimizing for goal A hard enough could produce a consequentialist* agent that cares about a different goal B. What is the goal B here (or am I misunderstanding the concept)?
*in the sense Christiano uses “consequentialist”
You are right, it’s not a good example, since the optimization pressure does not result in optimizing for a different goal.