Median estimate for when they’ll start working on a serious code project (i.e., not just toy code to illustrate theorems) is 2017.
This will not necessarily be development of friendly AI—maybe a component of friendly AI, maybe something else. (I have no strong estimates for what that other thing would be, but just as an example—a simulated-world sandbox).
Everything I say above (and elsewhere), is my opinion, not MIRIs.
Median estimate for when they’ll start working on friendly AI, if they get started with that before the Singularity, and if their direction doesn’t shift away from their apparent current long-term plans to do so: 2025.
Median estimate for when they’ll start working on friendly AI, if they get started with that before the Singularity, and if their direction doesn’t shift away from their apparent current long-term plans to do so: 2025.
Yes, but not because of MIRI. Along with FHI, they are doing more than anyone to improve our odds. As to whether writing code or any other strategy is the right one—I don’t know, but I trust MIRI more than anyone to get that right.
I am pretty sure, that as soon as first AGI’s arrive on the market, people would start to take possible dangers more seriously.
And it will be quite likely at that point that we are much closer to having an AGI that will foom than to having an AI that won’t kill us and that it is too late.
I know it is a local trope that death and destruction is apparent and necessary logical conclusion of creating an intelligent machine capable of self improvement and goal modification, but I certainly don’t share those sentiments.
How do you estimate the probability that AGI’s won’t take over the world (people who constructed them may use them for that purpose, but it is a different story), and would be used as simple tools and advisors in the same way boring, old fashioned and safe way 100% of our current technology is used?
I am explicitly saying that MRI or FAI are pointless, or anything like that. I just want to point out that they posture as if they were saving the world from imminent destruction, while it is no where certain weather said danger is really the case.
How do you estimate the probability that AGI’s won’t take over the world (people who constructed them may use them for that purpose, but it is a different story), and would be used as simple tools and advisors in the same way boring, old fashioned and safe way 100% of our current technology is used?
1%? I believe that it is nearly impossible to use a foomed AI in a safe manner without explicitly trying to do so. That’s kind of why I am worried about the threat of any uFAI developed before it is proven that we can develop a Friendly one and without using whatever the proof entails.
Anyway,
...would be used as simple tools and advisors in the same way boring, old fashioned and safe way 100% of our current technology is used?
I wasn’t aware that we use a 100% of our current technology in a safe way.
You may have a different picture of current technology than I do, or you may be extrapolating different aspects. We’re already letting software optimize the external world directly, with slightly worrying results. You don’t get from here to strictly and consistently limited Oracle AI without someone screaming loudly about risks. In addition, Oracle AI has its own problems (tell me if the LW search function doesn’t make this clear).
Some critics appear to argue that the direction of current tech will automatically produce CEV. But today’s programs aim to maximize a behavior, such as disgorging money. I don’t know in detail how Google filters its search results, but I suspect they want to make you feel more comfortable with links they show you, thus increasing clicks or purchases from sometimes unusually dishonest ads. They don’t try to give you whatever information a smarter, better informed you would want your current self to have. Extrapolating today’s Google far enough doesn’t give you a Friendly AI, it gives you the making of a textbook dystopia.
The distribution is asymmetric for obvious reasons. The probability for 2014 is pretty close to zero. This means that there is a 50% probability that a serious code project will start after 2020.
This is inconsistent with 2017 being a median estimate.
Over the years, I have seen a few online comments about toy programs written by MIRI people, e.g., this, search for “Haskell”. But I don’t know anything more about these programs that those brief reports.
When do you estimate that MIRI will start writing the code for a friendly AI?
Median estimate for when they’ll start working on a serious code project (i.e., not just toy code to illustrate theorems) is 2017.
This will not necessarily be development of friendly AI—maybe a component of friendly AI, maybe something else. (I have no strong estimates for what that other thing would be, but just as an example—a simulated-world sandbox).
Everything I say above (and elsewhere), is my opinion, not MIRIs. Median estimate for when they’ll start working on friendly AI, if they get started with that before the Singularity, and if their direction doesn’t shift away from their apparent current long-term plans to do so: 2025.
This is not a MIRI official estimate and you really should have disclaimed that.
OK, I will edit this one as well to say that.
We’re so screwed, aren’t we?
Yes, but not because of MIRI. Along with FHI, they are doing more than anyone to improve our odds. As to whether writing code or any other strategy is the right one—I don’t know, but I trust MIRI more than anyone to get that right.
Oh yes, I know that. It just says a lot that our best shot is still decades away from achieving it’s goal.
Which, to be fair, isn’t saying much.
Seeing as we are talking about speculative dangers coming from a speculative technology that has yet to be developed, it seems pretty understandable.
I am pretty sure, that as soon as first AGI’s arrive on the market, people would start to take possible dangers more seriously.
And it will be quite likely at that point that we are much closer to having an AGI that will foom than to having an AI that won’t kill us and that it is too late.
I know it is a local trope that death and destruction is apparent and necessary logical conclusion of creating an intelligent machine capable of self improvement and goal modification, but I certainly don’t share those sentiments.
How do you estimate the probability that AGI’s won’t take over the world (people who constructed them may use them for that purpose, but it is a different story), and would be used as simple tools and advisors in the same way boring, old fashioned and safe way 100% of our current technology is used?
I am explicitly saying that MRI or FAI are pointless, or anything like that. I just want to point out that they posture as if they were saving the world from imminent destruction, while it is no where certain weather said danger is really the case.
1%? I believe that it is nearly impossible to use a foomed AI in a safe manner without explicitly trying to do so. That’s kind of why I am worried about the threat of any uFAI developed before it is proven that we can develop a Friendly one and without using whatever the proof entails.
Anyway,
I wasn’t aware that we use a 100% of our current technology in a safe way.
You may have a different picture of current technology than I do, or you may be extrapolating different aspects. We’re already letting software optimize the external world directly, with slightly worrying results. You don’t get from here to strictly and consistently limited Oracle AI without someone screaming loudly about risks. In addition, Oracle AI has its own problems (tell me if the LW search function doesn’t make this clear).
Some critics appear to argue that the direction of current tech will automatically produce CEV. But today’s programs aim to maximize a behavior, such as disgorging money. I don’t know in detail how Google filters its search results, but I suspect they want to make you feel more comfortable with links they show you, thus increasing clicks or purchases from sometimes unusually dishonest ads. They don’t try to give you whatever information a smarter, better informed you would want your current self to have. Extrapolating today’s Google far enough doesn’t give you a Friendly AI, it gives you the making of a textbook dystopia.
What are the error bars around these estimates?
The first estimate: 50% probability between 2015 and 2020.
The second estimate: 50% probability between 2020 and 2035. (again, taking into account all the conditioning factors).
Um.
The distribution is asymmetric for obvious reasons. The probability for 2014 is pretty close to zero. This means that there is a 50% probability that a serious code project will start after 2020.
This is inconsistent with 2017 being a median estimate.
Unless he thinks it’s very unlikely the project will start between 2017 and 2020 for some reason.
Good point. I’ll have to re-think that estimate and improve it.
If some rich individual were to donate 100 million USD to MIRI today, how would you revise your estimate (if at all)?
Can you elaborate on the types of toy code that you (or others) have tried in terms of illustrating theoreoms?
I have not tried any.
Over the years, I have seen a few online comments about toy programs written by MIRI people, e.g., this, search for “Haskell”. But I don’t know anything more about these programs that those brief reports.