AIXI doesn’t work. My point was that if it did work, it would need a lot of coddling. Someone would need to tend to its utility function utility continually to make sure it was doing what it was supposed to.
If AIXI were interacting dynamically with its environment to a sufficient degee*, then the selected hypothesis motivating AIXI’s next action would come to contain some description of how AIXI is approaching the problem.
If AIXI is consistently making mistakes which would have been averted if it had possessed some model of itself at the time of making the mistake, then it is not selecting the best hypothesis, and it is not AIXI.
I think my use of words like “learning” suggested that I think of AIXI as a neural net or something. I get how AIXI works, but it’s often hard to be both accurate and succinct when talking about complex ideas.
AIXI doesn’t work. My point was that if it did work, it would need a lot of coddling. Someone would need to tend to its utility function utility continually to make sure it was doing what it was supposed to.
If AIXI were interacting dynamically with its environment to a sufficient degee*, then the selected hypothesis motivating AIXI’s next action would come to contain some description of how AIXI is approaching the problem.
If AIXI is consistently making mistakes which would have been averted if it had possessed some model of itself at the time of making the mistake, then it is not selecting the best hypothesis, and it is not AIXI.
I think my use of words like “learning” suggested that I think of AIXI as a neural net or something. I get how AIXI works, but it’s often hard to be both accurate and succinct when talking about complex ideas.
*for some value of “sufficient”