Maybe I missed something but I could never see why there was anything intrinsically good about (say) the short bias in the Solomonoff prior, it seemed like the whole thinking bigger programs were less likely was just a necessary trick to make the infinite sums finite. If we consider the formalism in the technical note that only keeps a finite number of sentences in memory then we don’t have to worry about this issue and can sample uniformly rather than (arbirtrarily?) picking a bias.
If we can avoid assuming simplicity is more likely and treat it as a fact about the world which we might learn after choosing a prior isn’t that better?
I feel that the bigger issue with this proposal is that it doesn’t solve the problem it’s trying to solve as statements in S are biased towards being picked.
Maybe I missed something but I could never see why there was anything intrinsically good about (say) the short bias in the Solomonoff prior, it seemed like the whole thinking bigger programs were less likely was just a necessary trick to make the infinite sums finite. If we consider the formalism in the technical note that only keeps a finite number of sentences in memory then we don’t have to worry about this issue and can sample uniformly rather than (arbirtrarily?) picking a bias.
In your paper http://ict.usc.edu/pubs/Logical%20Prior%20Probability.pdf you have to worry about the infinite series summing to infinity as the probability distribution is over all sentences in a language, so you have to include this bias.
If we can avoid assuming simplicity is more likely and treat it as a fact about the world which we might learn after choosing a prior isn’t that better?
I feel that the bigger issue with this proposal is that it doesn’t solve the problem it’s trying to solve as statements in S are biased towards being picked.