Elliot_Olds
Hi Ben. Are you still planning to post a transcript sometime?
Thanks. I had one question about your Toward Idealized Decision Theory paper.
I can’t say I fully understand UDT, but the ‘updateless’ part does seem very similar to the “act as if you had precommitted to any action that you’d have wanted to precommit to” core idea of NDT. It’s not clear to me that the super powerful UDT would make the wrong decision in the game where two players pick numbers between 0-10 and get payouts based on their pick and the total sum.
Wouldn’t the UDT reason as follows? “If my algorithm were such that I wouldn’t just pick 1 when the human player forced me into it by picking 9 (for instance maybe I always pick 5 in this game), then I may still have a reputation as a powerful predictor but it’s much more likely that I’d also have a reputation as an entity that can’t be bullied like this, so the human would be less likely to pick 9. That state of the world is better for me, so I shouldn’t be the type of agent that makes the greedy choice to pick 1 when I predict the human will pick 9.”
The argument in your paper seems to rely on the human assuming the UDT will reason like a CDT once it knows the human will pick 9.
An AI should certainly cooperate if it discovered that by chance its opposing AI had identical source code.
I read your paper and the two posts in your short sequence. Thanks for the links. I still think it’s very unlikely that one of the AIs in your original hypothetical (when they don’t examine each other’s source code) would do better by defecting.
I accept that if an opposing AI had a model of you that was just decent but not great, then there is some amount of logical connection there. What I haven’t seen is any argument about the shape of the graph of logical connection strength vs similarity of entities. I hypothesize that for any two humans who exist today, if you put them in a one shot PD, the logical connection is negligible.
Has anyone written specifically on how exactly to give weights to logical connections between similar but non-identical entities?
I think defect is the right answer in your AI problem and therefore that NDT gets it right, but I’m aware lots of LWers think otherwise. I haven’t researched this enough to want to argue it, but is there a discussion you’d recommend I read that spells out the reasoning? Otherwise I’ll just look through LW posts on prisoner’s dilemmas.
Secondly, I’d like to try to somehow incorporate logical effects into NDT. I agree they’re important. Any suggestions for where I could find lots of examples of decision problems where logical effects matter, to help me think about the general case?
In the retro blackmail, CDT does not precommit to refusing even if it’s given the opportunity to do so before the researcher gets its source code.
To clarify: you mean that CDT doesn’t precommit at time t=1 even if the researcher hasn’t gotten the code representing CDT’s state at time t=0 yet. The CDT doesn’t think precommitting will help because it knows the code the researcher will get will be from before its precommitment. I agree that this is true, and a CDT won’t want to precommit.
I guess my definition even after my clarification is ambiguous, as it’s not clear that what a CDT wishes it could have precomitted to at an earlier time should take precedence over what it would wish to precommit to at a later time. NDT seems to be best when you always prefer the earliest precommitment. The intuition is something like:
You should always make the decision that a CDT-agent would have wished he had precommitted to, if he had magically had the opportunity to costlessly precommit to to a decision at a time before the beginning of the universe.
This would allow you to act is if you had precommitted to things before you existed.
But two agents written in spacelike separation from each other might have decision algorithms that are logically correlated, despite there being no causal connection no matter how far back you go.
Can you give an example of this? Similar to the calculator example in the TDT paper, I’m imagining some scenario where one AI takes instructions for creating you to another galaxy, and another AI keeps a copy of the instructions for creating you on Earth. At some point, both AIs read the instructions and create identical beings, one of which is you. The AI that created you says that you’ll be playing a prisoner’s dilemma game with the other entity created in the same way, and asks for your decision.
In some sense, there is only a logical connection between these two entities because they’ve only existed for a short time and are too far away to have a causal effect on each other. However they are very causally related, and I could probably make an argument that they are replicas of the same person.
Do you have an example of a logical connection that has no causal connection at all (or as minimal a causal connection as possible)?
For example, a decision algorithm based on precommitment is unable to hold selfish preferences (valuing a cookie for me more than a cookie for a copy of me) in anthropic situations
I disagree that it makes sense to talk about one of the future copies of you being “you” whereas the other isn’t. They’re both you to the same degree (if they’re exact copies).
Eliezer talked about this in his TDT paper. It is possible to hypothesize scenarios where agents get punished or rewarded for arbitrary reasons. For instance an AI could punish agents who made decisions based on the idea of their choices determining the results of abstract computations (as in TDT). This wouldn’t show that TDT is a bad decision theory or even that it’s no better than any other theory.
If we restrict ourselves to action-determined and decision-determined problems (see Eliezer’s TDT paper) we can say that TDT is better than CDT, because it gets everything right that CDT gets right, plus it gets right some things that CDT gets wrong.
Can you think of any way that a situation could be set up that punishes an NDT agent, that doesn’t reduce to an AI just not liking NDT agents and arbitrarily trying to hurt them?
I think my definition of NDT above was worded badly. The problematic part is “if he had previously known he’d be in his currently situation.” Consider this definition:
You should always make the decision that a CDT-agent would have wished he had precommitted to, if he previously considered the possibility of his current situation and had the opportunity to costlessly precommit to a decision.
The key is that the NDT agent isn’t behaving as if he knew for sure that he’d end up blackmailed when he made his precommitment (since his precommitment affects the probability of his being blackmailed), but rather he’s acting “as if” he precommitted to some behavior based on reasonable estimates of the likelihood of his being kidnapped in various cases.
I believe that NDT gets this problem right.
The paper you link to shows that a pure CDT agent would not self modify into an NDT agent, because a CDT agent wouldn’t really have the concept of “logical” connections between agents. The understanding that both logical and causal connections are real things is what would compel an agent to self-modify to NDT.
However, if there was some path by which an agent started out as pure CDT and then became NDT, the NDT agent would still choose correctly on Retro Blackmail even if the researcher had its original CDT source code. The NDT agent’s decision procedure explicitly tells it to behave as if it had precommitted before the researcher got its source code.
So even if the CDT --> NDT transition is impossible, since I don’t think any of us here are pure CDT agents, we can still adopt NDT and profit.
You’re right, I think poker chips is too messy. Since you already have pens right there, it’d be better for people to just write a little mark / star / [their name] on any card that they were interested in.
Pretty much. The fact that NDT is so obvious is why I’m puzzled as to why TDT needed to be created, and why Eliezer didn’t end his paper shortly after the discussion of Gloria. NDT seems to get all the tricky decision problems right, even at least one that TDT gets wrong, so what am I missing?
Idea related to the clipboard, but combined with poker chips:
There is a stack of blank note cards on the table, and several pens/markers. If there’s an existing discussion and you want to talk about an unrelated topic, you grab a notecard, write down the topic, and place it face up on the table. At any time, there may be several note cards on the table representing topics people want to talk about. Each person also has a poker chip (or a few) that they may place near a particular card, expressing their interest in talking about that topic. Poker chips are basically upvotes.
Why isn’t the following decision theory optimal?
Help create an instrumental rationality “stack ranking”?
I stutter and have done a lot of research on stuttering. It’s rare that adult stutterers ever completely stop stuttering, but these two ebooks are the best resources I know of for dealing with it:
http://www.stutteringhelp.org/Portals/English/Book_0012_tenth_ed.pdf http://www.scribd.com/doc/23283047/Easy-Stuttering-Avoidance-Reduction-Therapy
The short version is that the less you try to suppress or conceal your stuttering the less severe it will become in the long run.
“the plan that lets you save money in the US is a life-engulfing minefield of time-consuming bargin-hunting, self-denial, and tax evasion.”
I work as a software developer in the US, have never made a ‘budget’ for myself or tried to analyze my finaces before now, I pay taxes normally, eat out often, and have no trouble saving lots of money. I’m going to substitute my expenses and pretend I only make 100k and see how much I’d still be able to save (living in Seattle).
Rent: 16.8k instead of 23.2k Utilities: 2k instead of 7k (how can you spend 7k on utilities if you’re a single person in an apartment?) Misc house expenses: 0.5k instead of 6.8k (what are these misc expenses that other people supposedly spend so much on?) Food: The estimate of 13.3k is reasonable for food, although it’s easy to spend a lot less without hardship. Transportation: 4.6k instead of 16.5k (who spends 16.5k per year on transportation? Just don’t buy a new BMW every 5 years and you should be set. I bought my car for $9k, 5 years ago).
Apparently it’s pretty easy live well in a large US city and save 33.9k per year without really paying attention to your finances. If you’re a good software developer you should be able to make a lot more than 100k and therefore save much more per year.
Steve Omohundro says:
”1) Nobody powerful wants to create unsafe AI but they do want to take advantage of AI capabilities.
2) None of the concrete well-specified valuable AI capabilities require unsafe behavior”
I think a lot of powerful people / organizations do want take advantage of possibly unsafe AI capabilities, such as ones that would allow them to be the emperors of the universe for all time. Especially if not doing so means that their rivals have a higher chance of becoming the emperors of the universe.