Another way of putting this is that all probability distributions over an infinite enumerated set are biased towards elements that occur earlier, regardless of the principle of enumeration.
If one has any assignment of probabilities to an infinite series of mutually exclusive hypotheses H1, H2, …, then for every epsilon > 0 there is an N such that every hypothesis after the Nth has probability less than epsilon. In fact, there is an N such that the sum of the probabilities of all the hypotheses after the Nth is less than epsilon.
Indeed you could, but that problem is already present in the definition of Kolmogorov complexity. It’s only defined up to an arbitrary additive constant determined by (in one formulation) the choice of a universal Turing machine. The Kolmogorov complexity of a string is the size of the shortest input for that UTM that produces that string as output, but there’s nothing in the definition to prevent the UTM from having any finite set of arbitrary strings on speed-dial.
Kelly deals with this by looking at complexity from other angles. For example, a complex world can give you a long sequence of observations persuading you that it’s a simple world and then suddenly “change its mind”, but a simple world cannot pretend that it’s complex.
Hmm… maybe I was reading your claim as stronger than you intended. I was imagining you were claiming that property would hold for any finite enumerated subset, which clearly isn’t what you meant.
If the sum of every term in a sequence after the Nth one is less than epsilon, then the sum of every term in any subsequence after the Nth one is also less than epsilon.
Right, but that isn’t what I meant—it is not necessarily the case that for every n, every hypothesis after the nth has probability less than that of the the nth hypothesis. Obviously—which is why I should have noticed my confusion and not misread in the first place.
And yet another way is this: every computable enumeration of all programs is equivalent to a UTM, and therefore generates the same measure of Kolmogorov complexity, up to the usual additive constant.
Proof: Computability of the enumeration means that there is a program which takes any integer N and returns the Nth in that enumeration. Represent this program as a Turing machine and cascade it with a second Turing machine that executes the program returned by the first TM.
Is this a novel observation? Does it go any distance towards answering Shalizi’s question at the end of this note?
This assumes countable additivity. Otherwise, each basic event could have probability zero, though the whole has probability 1.
I’m inclined to use countable additivity, but I couldn’t give you arguments for it, other than by sketching models that violate it and pointing out that they are weird.
Another way of putting this is that all probability distributions over an infinite enumerated set are biased towards elements that occur earlier, regardless of the principle of enumeration.
Certainly not regardless of the probability distribution though. Do you have a particular probability distribution in mind?
Regardless of the probability distribution.
If one has any assignment of probabilities to an infinite series of mutually exclusive hypotheses H1, H2, …, then for every epsilon > 0 there is an N such that every hypothesis after the Nth has probability less than epsilon. In fact, there is an N such that the sum of the probabilities of all the hypotheses after the Nth is less than epsilon.
But N could be 3^^^3. Which does an injury to the term early in my book. E.g. I could swap the probabilities of p (x3^^^3) and p(x1).
Indeed you could, but that problem is already present in the definition of Kolmogorov complexity. It’s only defined up to an arbitrary additive constant determined by (in one formulation) the choice of a universal Turing machine. The Kolmogorov complexity of a string is the size of the shortest input for that UTM that produces that string as output, but there’s nothing in the definition to prevent the UTM from having any finite set of arbitrary strings on speed-dial.
Kelly deals with this by looking at complexity from other angles. For example, a complex world can give you a long sequence of observations persuading you that it’s a simple world and then suddenly “change its mind”, but a simple world cannot pretend that it’s complex.
Why not? It would look almost exactly like the complex worlds imitating it, wouldn’t it?
Hmm… maybe I was reading your claim as stronger than you intended. I was imagining you were claiming that property would hold for any finite enumerated subset, which clearly isn’t what you meant.
If the sum of every term in a sequence after the Nth one is less than epsilon, then the sum of every term in any subsequence after the Nth one is also less than epsilon.
Right, but that isn’t what I meant—it is not necessarily the case that for every n, every hypothesis after the nth has probability less than that of the the nth hypothesis. Obviously—which is why I should have noticed my confusion and not misread in the first place.
Yes, regardless of the distribution.
And yet another way is this: every computable enumeration of all programs is equivalent to a UTM, and therefore generates the same measure of Kolmogorov complexity, up to the usual additive constant.
Proof: Computability of the enumeration means that there is a program which takes any integer N and returns the Nth in that enumeration. Represent this program as a Turing machine and cascade it with a second Turing machine that executes the program returned by the first TM.
Is this a novel observation? Does it go any distance towards answering Shalizi’s question at the end of this note?
This assumes countable additivity. Otherwise, each basic event could have probability zero, though the whole has probability 1.
I’m inclined to use countable additivity, but I couldn’t give you arguments for it, other than by sketching models that violate it and pointing out that they are weird.
Yes, exactly.
This is the best way of putting it.
I agree with you that this is a good way of putting it (but don’t believe that ‘it’ is correct).