For the idea of interesting things, can you give an idea of how you would want random chance inside the program Quirrell shows to be handled? I’ll give a few examples below. In all cases, you can assume that the small positive chance of Quirrell stopping is some very tiny 1/3^^^^^^^3, but as above, you don’t know this exact number. (and clearly, Quirrell has the ability to make random numbers since he does it for this in the first place.)
1: Quirrell shows you a first program which says “Your AI stays exactly the same. Furthermore, I am going to offer a random 1/3^^^^3 chance of self-destruct and a 3^^^^3-1/3^^^^3 chance of doubling your winnings?”
At this point, you can either, accept, or reject.
If you accept, then Quirrell shows you a second program which says “Your AI stays exactly the same. Furthermore, I am going to offer a random 1/3^^^^3 chance of self-destruct and a 3^^^^3-1/3^^^^3 chance of doubling your winnings?” (and Quirrell just shows you this program over and over again.)
Well, you have exactly the same AI, so presumably you accept again. If you accepted, chances are, you’ll self-destruct with zillions of points. Quirrell can be very likely to win against some AI’s without changing the AI at all, by simply offering your AI a massive number of deals until it blows up, if the risk amount is well above the random stop amount. If you rejected, you’re safe. Is this correct?
2: Quirrell offers a slightly different deal: “Your AI stays exactly the same, except that you can remember the number of times I have offered you this deal and you may generate random numbers if desired. Furthermore, I am going to offer a random 1/3^^^^3 chance of self-destruct and a 3^^^^3-1/3^^^^3 chance of doubling your winnings, and I am going to offer you nothing but this deal for the rest of the game.”
Okay, well now you aren’t stuck into being unable to stop accepting the deal once you accept it, so presumably you should accept some number of deals N or stop accepting deals with some random chance O, or you should just reject up front. Let’s say you’re aware that at any point N, it would always look good to accept more deals, so you would eventually blow up much like in point 1. Perhaps you reject in advance again?
3: Let’s say Quirrell offers a slightly different deal: “Your AI stays exactly the same, except that you can remember the number of times I have offered you this deal and you may generate random numbers if desired. Furthermore, I am going to offer a random 1⁄4 chance of self-destruct and a 3⁄4 chance of doubling your winnings, and on the next round I am going to offer you a 1⁄8 chance of self-destruct and a 7⁄8 chance of doubling your winnings, and on each round I am going to half the chance of a self-destruct and increase the chance of doubling your winnings by the remaining amount.”
If you accept all deals in this case, you only have about a 50% chance of self-destruct as opposed to near certainty or tempting near certainty, but that might still be bad. You reject again.
4: So let’s say Quirrell offers a slightly different deal: “Your AI stays exactly the same, except that you can remember the number of times I have offered you this deal and you may generate random numbers if desired. Furthermore, I am going to offer a random 1⁄8 chance of self-destruct and a 7⁄8 chance of doubling your winnings, and if you take or double, on the next round I am going to offer you a 1⁄16 chance of self-destruct and a 15⁄16 chance of doubling your winnings, and on each round I am going to half the chance of a self-destruct and increase the chance of doubling your winnings by the remaining amount.”
If you accept all deals in this case, you only have about a cumulative 25% chance of self-destruct, but that might still be bad, and it will be safer next round.
So you wait 1 deal, and then you only have about a cumulative 12.5% chance of self-destruct, but that might still be bad, and it will be safer next round.
So you wait 2 deals, and then you only have about a cumulative 6.25% chance of self-destruct, but that might still be bad, and it will be safer next round.
Except, you can always use THAT logic. Quirrell always offers a safer deal on the next round anyway, so you hold off on accepting any of the deals at all, ever, and never start doubling. Quirrell eventually randomly stops and hands you a single Quirrell point, saying nothing. Did you win?
If self destructing is infinitely bad, it seems like you did, but if it is only finitely bad, you almost certainly lost unless the finite number was utterly gigantic.and while the described self-destruct was bad, I don’t think it was infinitely bad or gigantically finitely bad. It was a fairly well defined amount of badness.
But even if you say “Self destructing is about as bad as −10000 Quirrel points” it still doesn’t seem to prevent Quirrel from either offering you increasingly tantalizingly high rewards if you accept just a bit more risk, or offering you increasingly safe bets (which you don’t take, because you know a safer one is coming around) if you want less risk.
I mean, in the case of some of the other proposed problems, Quirrell offers you a problem with unknown or tricky to figure out risk. In these problems, Quirrell is more up front about the risks, and the damages, and it still seems like it’s very hard to say what you kind of risk management you should do to win, even though Quirrell hasn’t even bothered to use his power of recoding your values yet.
Except, you can always use THAT logic. Quirrell always offers a safer deal on the next round anyway, so you hold off on accepting any of the deals at all, ever, and never start doubling. Quirrell eventually randomly stops and hands you a single Quirrell point, saying nothing.
At some point the cumulative probability of self-destruction drops below the probability of accidentally cashing out this round. If you’d trade off a probability of self-destruction against an equal probability of a bajillion points, you start doubling then.
Hmm. That is a good point, but there is a slight complication: You don’t know how frequently the cash out occurs. From earlier:
In all cases, you can assume that the small positive chance of Quirrell stopping is some very tiny 1/3^^^^^^^3, but as above, you don’t know this exact number.
It’s true that if you continually half the chance of destruction every round it will eventually be the right thing to do to start doubling, but if you don’t know what the cash out chance is I’m not sure how you would calculate a solid approach, (I think there IS a method though, I believe I’ve seen these kinds of things calculated before.) even though Quirrell has removed all unknowns from the problem EXCEPT the uncertain cash out chance.
It sounds like one approach I should consider is to try to run some numbers with a concrete cash-out (of a higher chance) and then run more numbers (for lower chances) and see what kind of slope develops. Then I could attempt to figure out what kind of values I should use for your attempt to determine the cash out chance (it seems tricky, since every time a cash out does not occur, you would presumably become slightly more convinced a cash out is less likely.)
If I did enough math, it does seem like it would be possible to at least get some kind of estimate for it.
For the idea of interesting things, can you give an idea of how you would want random chance inside the program Quirrell shows to be handled? I’ll give a few examples below. In all cases, you can assume that the small positive chance of Quirrell stopping is some very tiny 1/3^^^^^^^3, but as above, you don’t know this exact number. (and clearly, Quirrell has the ability to make random numbers since he does it for this in the first place.)
1: Quirrell shows you a first program which says “Your AI stays exactly the same. Furthermore, I am going to offer a random 1/3^^^^3 chance of self-destruct and a 3^^^^3-1/3^^^^3 chance of doubling your winnings?”
At this point, you can either, accept, or reject.
If you accept, then Quirrell shows you a second program which says “Your AI stays exactly the same. Furthermore, I am going to offer a random 1/3^^^^3 chance of self-destruct and a 3^^^^3-1/3^^^^3 chance of doubling your winnings?” (and Quirrell just shows you this program over and over again.)
Well, you have exactly the same AI, so presumably you accept again. If you accepted, chances are, you’ll self-destruct with zillions of points. Quirrell can be very likely to win against some AI’s without changing the AI at all, by simply offering your AI a massive number of deals until it blows up, if the risk amount is well above the random stop amount. If you rejected, you’re safe. Is this correct?
2: Quirrell offers a slightly different deal: “Your AI stays exactly the same, except that you can remember the number of times I have offered you this deal and you may generate random numbers if desired. Furthermore, I am going to offer a random 1/3^^^^3 chance of self-destruct and a 3^^^^3-1/3^^^^3 chance of doubling your winnings, and I am going to offer you nothing but this deal for the rest of the game.”
Okay, well now you aren’t stuck into being unable to stop accepting the deal once you accept it, so presumably you should accept some number of deals N or stop accepting deals with some random chance O, or you should just reject up front. Let’s say you’re aware that at any point N, it would always look good to accept more deals, so you would eventually blow up much like in point 1. Perhaps you reject in advance again?
3: Let’s say Quirrell offers a slightly different deal: “Your AI stays exactly the same, except that you can remember the number of times I have offered you this deal and you may generate random numbers if desired. Furthermore, I am going to offer a random 1⁄4 chance of self-destruct and a 3⁄4 chance of doubling your winnings, and on the next round I am going to offer you a 1⁄8 chance of self-destruct and a 7⁄8 chance of doubling your winnings, and on each round I am going to half the chance of a self-destruct and increase the chance of doubling your winnings by the remaining amount.”
If you accept all deals in this case, you only have about a 50% chance of self-destruct as opposed to near certainty or tempting near certainty, but that might still be bad. You reject again.
4: So let’s say Quirrell offers a slightly different deal: “Your AI stays exactly the same, except that you can remember the number of times I have offered you this deal and you may generate random numbers if desired. Furthermore, I am going to offer a random 1⁄8 chance of self-destruct and a 7⁄8 chance of doubling your winnings, and if you take or double, on the next round I am going to offer you a 1⁄16 chance of self-destruct and a 15⁄16 chance of doubling your winnings, and on each round I am going to half the chance of a self-destruct and increase the chance of doubling your winnings by the remaining amount.”
If you accept all deals in this case, you only have about a cumulative 25% chance of self-destruct, but that might still be bad, and it will be safer next round. So you wait 1 deal, and then you only have about a cumulative 12.5% chance of self-destruct, but that might still be bad, and it will be safer next round. So you wait 2 deals, and then you only have about a cumulative 6.25% chance of self-destruct, but that might still be bad, and it will be safer next round.
Except, you can always use THAT logic. Quirrell always offers a safer deal on the next round anyway, so you hold off on accepting any of the deals at all, ever, and never start doubling. Quirrell eventually randomly stops and hands you a single Quirrell point, saying nothing. Did you win?
If self destructing is infinitely bad, it seems like you did, but if it is only finitely bad, you almost certainly lost unless the finite number was utterly gigantic.and while the described self-destruct was bad, I don’t think it was infinitely bad or gigantically finitely bad. It was a fairly well defined amount of badness.
But even if you say “Self destructing is about as bad as −10000 Quirrel points” it still doesn’t seem to prevent Quirrel from either offering you increasingly tantalizingly high rewards if you accept just a bit more risk, or offering you increasingly safe bets (which you don’t take, because you know a safer one is coming around) if you want less risk.
I mean, in the case of some of the other proposed problems, Quirrell offers you a problem with unknown or tricky to figure out risk. In these problems, Quirrell is more up front about the risks, and the damages, and it still seems like it’s very hard to say what you kind of risk management you should do to win, even though Quirrell hasn’t even bothered to use his power of recoding your values yet.
At some point the cumulative probability of self-destruction drops below the probability of accidentally cashing out this round. If you’d trade off a probability of self-destruction against an equal probability of a bajillion points, you start doubling then.
Hmm. That is a good point, but there is a slight complication: You don’t know how frequently the cash out occurs. From earlier:
It’s true that if you continually half the chance of destruction every round it will eventually be the right thing to do to start doubling, but if you don’t know what the cash out chance is I’m not sure how you would calculate a solid approach, (I think there IS a method though, I believe I’ve seen these kinds of things calculated before.) even though Quirrell has removed all unknowns from the problem EXCEPT the uncertain cash out chance.
It sounds like one approach I should consider is to try to run some numbers with a concrete cash-out (of a higher chance) and then run more numbers (for lower chances) and see what kind of slope develops. Then I could attempt to figure out what kind of values I should use for your attempt to determine the cash out chance (it seems tricky, since every time a cash out does not occur, you would presumably become slightly more convinced a cash out is less likely.)
If I did enough math, it does seem like it would be possible to at least get some kind of estimate for it.
Thanks! I’ll try to look into this further.