Good point; haggling is a good example of a fuzzy boundary between threats and trade.
If A is willing to sell a widget for any price above $10, and B is willing to buy a widget for any price below $20, and there are no other buyers or sellers, then for any price X strictly between $10 and $20, A saying “I won’t sell for less than X” and B saying “I won’t sell for more than X” are both threats under my model.
Which means that agents that “naively” precommit to never respond to any threats (the way I understand them) will not reach an agreement when haggling. They’ll also fail at the Ultimatum game.
So there needs to be a better model for threats, possibly one that takes shelling points into account; or maybe there should be a special category for “the kind of threats it’s beneficial to precommit to ignore”.
Hmm the pre-commitment to ignore would depend on other agents and their pre-pre-commitment to ignore pre-commitments. It just goes recursive like Sherlock Holmes vs Moriarty, and when you go meta and try to look for ‘limit’ of recursion, it goes recursive again… i have a feeling that it is inherently a rock-paper-skissors situation where you can’t cheat like this robot. (I.e. I would suggest, at that point, to try to make a bunch of proofs of impossibility to narrow expectations down somewhat).
It’s not possible to coordinate in general against arbitrary opponents, like it’s impossible to predict what an arbitrary program does, but it’s advantageous for players to eventually coordinate their decisions (on some meta-level of precommitment). On one hand, players want to set prices their way, but on the other they want to close the trade eventually, and this tradeoff keeps the outcome from both extremes (“unfair” prices and impossibility of trade). Players have an incentive to setup some kind of Loebian cooperation (as in theseposts), which stops the go-meta regress, although each will try to set the point where cooperation happens in their favor.
I was thinking rather of Halting Problem—like impossibility, along with rock-paper-skissors situation that prevents declaring any one strategy, even the cooperative, as the ‘best’.
If difficulty of selecting and implementing a strategy is part of the tradeoff (so that more complicated strategies count as “worse” because of their difficulty, even if they promise an otherwise superior outcome), maybe there are “best” strategies in some sense, like there is a biggest natural number that you can actually write down in 30 seconds. (Such things would of course have the character of particular decisions, not of decision theory.)
Huh? Just start writing. The rule wasn’t “the number you can define in 30 seconds”, but simply “the number you can write down in 30 seconds”. Like the number of strawberries you can eat in 30 seconds, no paradox there!
Given a fixed state of knowledge about possible opponents and finite number of feasible options for your decision, there will be maximal decisions, even if in an iterated contest the players could cycle their decisions against updated opponents indefinitely.
Good point; haggling is a good example of a fuzzy boundary between threats and trade.
If A is willing to sell a widget for any price above $10, and B is willing to buy a widget for any price below $20, and there are no other buyers or sellers, then for any price X strictly between $10 and $20, A saying “I won’t sell for less than X” and B saying “I won’t sell for more than X” are both threats under my model.
Which means that agents that “naively” precommit to never respond to any threats (the way I understand them) will not reach an agreement when haggling. They’ll also fail at the Ultimatum game.
So there needs to be a better model for threats, possibly one that takes shelling points into account; or maybe there should be a special category for “the kind of threats it’s beneficial to precommit to ignore”.
Hmm the pre-commitment to ignore would depend on other agents and their pre-pre-commitment to ignore pre-commitments. It just goes recursive like Sherlock Holmes vs Moriarty, and when you go meta and try to look for ‘limit’ of recursion, it goes recursive again… i have a feeling that it is inherently a rock-paper-skissors situation where you can’t cheat like this robot. (I.e. I would suggest, at that point, to try to make a bunch of proofs of impossibility to narrow expectations down somewhat).
It’s not possible to coordinate in general against arbitrary opponents, like it’s impossible to predict what an arbitrary program does, but it’s advantageous for players to eventually coordinate their decisions (on some meta-level of precommitment). On one hand, players want to set prices their way, but on the other they want to close the trade eventually, and this tradeoff keeps the outcome from both extremes (“unfair” prices and impossibility of trade). Players have an incentive to setup some kind of Loebian cooperation (as in these posts), which stops the go-meta regress, although each will try to set the point where cooperation happens in their favor.
I was thinking rather of Halting Problem—like impossibility, along with rock-paper-skissors situation that prevents declaring any one strategy, even the cooperative, as the ‘best’.
If difficulty of selecting and implementing a strategy is part of the tradeoff (so that more complicated strategies count as “worse” because of their difficulty, even if they promise an otherwise superior outcome), maybe there are “best” strategies in some sense, like there is a biggest natural number that you can actually write down in 30 seconds. (Such things would of course have the character of particular decisions, not of decision theory.)
There is not a biggest natural number that you can actually write down in thirty seconds—that’s equivalent to Berry’s paradox.
Huh? Just start writing. The rule wasn’t “the number you can define in 30 seconds”, but simply “the number you can write down in 30 seconds”. Like the number of strawberries you can eat in 30 seconds, no paradox there!
I was reading “write down” more generally than “write down each digit of in base ten,” but I guess that’s not how you meant it.
Hmm if it was a programming contest I would expect non-transitive ‘betterness’.
Given a fixed state of knowledge about possible opponents and finite number of feasible options for your decision, there will be maximal decisions, even if in an iterated contest the players could cycle their decisions against updated opponents indefinitely.