Why Bet Kelly?
The Kelly criterion is an elegant, but often misunderstood, result in decision theory. To begin with, suppose you have some amount of some resource, which you would like to increase. (For example, the resource might be monetary wealth.) You are given the opportunity to make a series of identical bets. You determine some fraction of your wealth to wager; then, in each bet, you gain a fraction with probability , and lose a fraction with probability .[1]
In other words, suppose is your wealth after bets. We will define , and we will suppose for simplicity that . Then , where is a random variable defined as:
Now suppose that, for some reason, we want to maximize . By linearity of expectation, . Hence, we should simply maximize . This amounts to solving:
This, , is known as the Kelly bet. For example, it says that if you have a 60-40 edge, then you should bet , i.e., bet of your current wealth on each bet.
That all seems pretty reasonable. But why do we want to maximize ? If we were to simply maximize expected wealth, i.e., , then a straightforward calculation shows that we should not bet Kelly—in fact, we should bet (“YOLO”), wagering the entire bankroll on every bet. This seems extremely counterintuitive, since, after bets, our wealth would then be:
In other words, as grows large, we would almost surely go bankrupt! Nevertheless, this would be the way to maximize . Kelly, whatever its merits, does not maximize -- not even in the long run. Especially not in the long run.
We now come to the perennial debate: why does Kelly seem “obviously right”, and YOLO “obviously wrong”? There are many answers usually offered to this question.
First, what we believe to be the correct answer:
Utility is not linear in wealth. As originally observed by Bernoulli, utility tends to be approximately logarithmic in wealth. If utility happens to be exactly logarithmic in wealth, then the Kelly bet is optimal. For most people, in most circumstances, utility is approximately logarithmic in wealth. The Kelly bet is approximately optimal. On the other hand, utility is very far from being linear in wealth, and so YOLO is a very bad idea.
In a certain sense, it is as simple as that. The von Neumann-Morgenstern utility theorem (vNM) tells us that we should be optimizing for some utility function . We know that the Kelly criterion always optimizes . Therefore, if the Kelly criterion is optimal, it is because .
Now, there are many other answers to “why bet Kelly?” that initially seem plausible:
Kelly maximizes the expected growth rate, . This happens to be true, and speaks to the elegance of Kelly’s result.[2] However, unless for some reason you find yourself in a contest where you only win the prize if you have the highest expected growth rate, this is not a good reason to bet Kelly. vNM says we should maximize expected utility, not maximize expected growth rate.
Kelly maximizes the geometric mean of wealth, . This is also evidently true, as is precisely . However, vNM says we should maximize expected utility (i.e., arithmetic mean of utility), not geometric mean of wealth. Again, if utility happens to be approximately logarithmic in wealth, then maximizing the geometric mean of wealth feels right, but it’s because of the logarithmic utility of wealth.
The Kelly bettor, with high probability, ends up with higher wealth than the non-Kelly bettor. This is particularly evident when Kelly is compared with YOLO. But, again, vNM does not say “maximize wealth with high probability”; it says “maximize expected utility”.
We should try to optimize something that has nice properties (e.g., can be time-averaged, or can be optimized myopically [Mossin, 1968], [Hakansson, 1971]). There is certainly an argument that, if our utility function happens to already be approximately logarithmic, then we might want to adopt logarithmic utility as a heuristic, since it has these nice properties. However, ultimately our true utility function is what it is. If we claim that Kelly is optimal, and we claim that our true utility function is not logarithmic in wealth, then we are rejecting vNM.
We should simply reject vNM, and optimize something else as a terminal value (e.g., geometric mean or maximin). This seems quite drastic, as the vNM axioms are very mild assumptions.
So, we claim, if Kelly is optimal then it is because our utility function is . However, this is not the whole story. The utility function refers to the utility of wealth at the moment after the betting experiment, not the terminal utility of wealth in general. We can imagine that this experiment is just the preamble to a much longer game, in which is the ultimate terminal value of wealth (e.g., in number of lives saved), and we are investing over time steps where, in each step, we have the opportunity to place a bet with some statistical edge . We can then use backward induction to determine the utility function that we should adopt for wealth at previous points in the game: . It is this final function, , that we should treat as our “utility function” in the preamble experiment.
Now, suppose we ultimately have something like this as our terminal utility function:
In other words, number-of-lives-saved is linear in money up to a certain point, then flat—an exaggerated version of the phenomenon of diminishing returns. As it turns out, when we apply backward induction for reasonably large values of (e.g., ) and modest statistical edge (e.g., , we obtain a preamble utility function that looks something like this (taking for simplicity):
In general, this function “looks more like a logarithm” than the piecewise-linear function , and falls off sharply as we approach zero. Clearly it is not actually a logarithm, as it is bounded above and below (and is, in fact, equal to for values ). But, for a broad class of terminal utility functions , the resulting function looks surprisingly logarithm-like.
In summary, the Kelly criterion is an elegant, and surprisingly simple, formula for optimizing . As a general strategy, optimizing is appealing in a number of ways:
It has many aesthetically appealing properties: it maximizes geometric growth rate; it maximizes the geometric mean over outcomes; it results in outperforming other bettors with high probability; and it is stable in the sense of Mossin and Hakansson.
Separately, is often a good approximation to the true expected utility of money-after-the-bet, if the scenario specifies a long series of subsequent opportunities to make bets. When we examine the instrumental utility function that arises from applying backward induction to such a series of opportunities, we find that it often “looks like a logarithm”.
However, we should remember that the Kelly bet, ultimately, is only an approximation. The true optimal bet—the one that actually maximizes expected utility -- may be significantly different, in either direction.
Acknowledgements: We would like to thank davidad for many helpful comments on earlier drafts of this article.
I… think this is wrong? It’s late and I should sleep so I’m not going to double check, but this sounds like you’re saying that you can take two sequences, one has a higher value at every element but the other has a higher limit.
If something similar to what you wrote is correct, I think it will be that Kelly maximizes E(limn→∞W1/nn). That feels about right to me, but I’m not confident.
Something I’ve often wondered—if utility for money is logarithmic, AND maximizing expected growth means logarithmic betting in the underlying resource, should we be actually thinking log(log(n))? I think the answer is “no”, because declining marginal utility is irrelevant to this—we still value more over less at all points.
No—you should bet so as to maximize E[U]. If U(W)=logW, and you are wagering W, then bet Kelly, which optimizes E[logW]=E[U]. However, if for some reason you are directly wagering U (which seems very unlikely), then the optimal bet is actually YOLO, not Kelly.
I think the key thing to note here is that “maximizing expected growth” looks the same whether the thing you’re trying to grow is money or log-money or sqrt-money or what. It “just happens” that (at least in this framework) the way one maximizes expected growth is the same as the way one maximizes expected log-money.
I’ve recently written about this myself. My goal was partly to clarify this, though I don’t know if I succeeded.
I think the post confuses things by motivating the Kelly bet as the thing that maximizes expected log-money, and also has other neat properties. To my mind, if you want to maximize expected log-money, you just… do the arithmetic to figure out what that means. It’s not quite trivial, but it’s stats-101 stuff. I don’t think it seems more interesting to do the arithmetic that maximizes expected log-money compared to expected money or expected sqrt-money. Kelly certainly didn’t introduce the criterion as “hey guys, here’s a way to maximize expected log-money”. (Admittedly, I don’t much care about his framing either. The original paper is information-theoretic in a way that seems to be mostly forgotten about these days.)
To my mind, the important thing about the Kelly bet is the “almost certainly win more money than anyone using a different strategy, over a long enough time period” thing. (Which is the same as maximizing expected growth rate, when growth is exponential. If growth is linear you still might care if you’re earning $2/day or $1/day, but the “growth rate” of both is 0 as defined here.) So I prefer to motivate the Kelly bet as being the thing that does that, and then say “and incidentally, turns out this also maximizes expected log-wealth, which is neat because...”.