I don’t understand. Based on reading through the passages you referenced in PtLoS, maximum entropy is a way of choosing a distribution out of a family of distributions (which, by the way, is a frequentist technique, not a Bayesian one). Solomonoff induction is a choice of prior. I don’t really understand in what sense these are related to each other, or in what sense Maxent generates priors at all.
Incidentally I’m surprised that there appears to be so much disagreement about this, given that LW is basically a forum brought into existence on the strength of Yudkowsky’s abilities as a thinker, writer and populariser, and he clearly holds frequentism in contempt.
I’ve always felt that the frequentists that Eliezer argues against are straw men. As I said earlier, I’ve never met a frequentist who is guilty of the accusations that you keep making, although I have met Bayesians whose philosophy interfered with their ability to do good statistical modeling / inference. Have you actually run into the people who you seem to be arguing against? If not, then I think you should restrict yourself to arguing against opinions that people are actually trying to support, although I also think that whether or not some very foolish people happen to be frequentists is irrelevant to the discussion (something Eliezer himself discussed in the “Reversed Stupidity is not Intelligence” post).
If you know nothing about a variable except that it’s in the interval [a, b] your probability distribution must be from the class of distributions where p(x) = 0 for x outside of [a, b]. You pick the distribution of maximal entropy from this class as your prior, to encode ignorance of everything except that x ∈ [a,b].
That is one way Maxent may generate a prior, anyway.
I don’t understand. Based on reading through the passages you referenced in PtLoS, maximum entropy is a way of choosing a distribution out of a family of distributions (which, by the way, is a frequentist technique, not a Bayesian one). Solomonoff induction is a choice of prior. I don’t really understand in what sense these are related to each other, or in what sense Maxent generates priors at all.
I’ve always felt that the frequentists that Eliezer argues against are straw men. As I said earlier, I’ve never met a frequentist who is guilty of the accusations that you keep making, although I have met Bayesians whose philosophy interfered with their ability to do good statistical modeling / inference. Have you actually run into the people who you seem to be arguing against? If not, then I think you should restrict yourself to arguing against opinions that people are actually trying to support, although I also think that whether or not some very foolish people happen to be frequentists is irrelevant to the discussion (something Eliezer himself discussed in the “Reversed Stupidity is not Intelligence” post).
If you know nothing about a variable except that it’s in the interval [a, b] your probability distribution must be from the class of distributions where p(x) = 0 for x outside of [a, b]. You pick the distribution of maximal entropy from this class as your prior, to encode ignorance of everything except that
x ∈ [a,b]
.That is one way Maxent may generate a prior, anyway.
We can call dibs on things now? Ooh, I call dibs on approximating a slowly varying function as a constant!