I’m still forming my views and I don’t think I’m well calibrated to state any probability with authority yet. My uncertainty still feels so high that I think my error bars would be too wide for my actual probability estimates to be useful. Some things I’m thinking about:
Forecasters are not that great at making forecasts greater than 5 years out according to Superforecasting IIRC and I don’t think AGI is going to happen within the next 5 years.
AGI has not been created yet and its possible that AI development gets derailed due to other factors e.g.:
Political and economic conditions change such that investment in AI slows down.
Global conflict exacerbates which slows down AI (maybe this speeds it up but I think there would be other pressing needs when a lot of resources has to be diverted to war)
Other global catastrophic risks could happen before AGI is developed i.e. should I be more scared of AGI than say nuclear war or GCBRs at this point (not that great but could still happen)
On the path to AGI there could be a catastrophic failure that kills a few people but can be contained but gets people really afraid of AI.
Maybe some of the work on AI safety ends up helping produce mostly aligned AI. I’m not sure if everyone dies if an AI is 90% aligned.
Maybe the AGI systems that are built don’t have instrumental convergence maybe if we get AGI through CAIS which seems to me like the most likely way we’ll get there.
Maybe like physics once the low hanging fruit has been plucked then it takes a while to make breakthroughs which extends the timelines
For me to be personally afraid I’d have to think this was the primary way I would die which seems unlikely given all the other ways I could die between now and if/when AGI is developed.
AI researchers, who are the people that most likely believe that AGI is possible more than anyone else, don’t have consensus when it comes to this issue. I know experts can be wrong about their own fields but I’d expect them to be more split on the issue(I don’t know what the current status is now just know what it was in the Grace et. al survey). I know very little about AGI, should I be more concerned than AI researchers are?
I still think it’s important to work on AI Safety since even a small chance that AGI could go wrong would still have a high expected value in terms of the negative outcome. I think most of my thinking comes from the fact that I think it is more probable that there will be a slow take off instead of a fast take off. I may also just be bad at being scared or feeling doomed.
What are some relatively-likely examples of future possible observations that would make you think AGI is every likely to kill everyone?
People start building AI that is agentic and open ended in its actions.
Would you expect to make observations like that well in advance of AGI (if doom is in fact likely), such that we can expect to have plenty of time to prepare if we ever have to make that future update?
Yes, because I think the most likely scenario is a slow take off. This is because it costs money to scale compute and we actually need to validate and the more complex a system the harder it is to build correctly, probably takes a few iterations to get things to work well enough that it can be tested against a benchmark before moving on to trying to get a system to have more capability. I think this process will have to happen many times before getting to AI that is dangerous and on the way I’d expect to start seeing some interesting agentic behavior with short-horizon planning.
Or do you think we’re pretty screwed, evidentially speaking, and can probably never update much toward ‘this is likely to kill us’ until it’s too late to do anything about it?
I think the uncertainty will be pretty high until we start seeing sophisticated agentic behavior. Though I don’t think we should wait that long to try come up with solutions since I think a small chance that this could happen still warrants concern.
I’m still forming my views and I don’t think I’m well calibrated to state any probability with authority yet. My uncertainty still feels so high that I think my error bars would be too wide for my actual probability estimates to be useful. Some things I’m thinking about:
Forecasters are not that great at making forecasts greater than 5 years out according to Superforecasting IIRC and I don’t think AGI is going to happen within the next 5 years.
AGI has not been created yet and its possible that AI development gets derailed due to other factors e.g.:
Political and economic conditions change such that investment in AI slows down.
Global conflict exacerbates which slows down AI (maybe this speeds it up but I think there would be other pressing needs when a lot of resources has to be diverted to war)
Other global catastrophic risks could happen before AGI is developed i.e. should I be more scared of AGI than say nuclear war or GCBRs at this point (not that great but could still happen)
On the path to AGI there could be a catastrophic failure that kills a few people but can be contained but gets people really afraid of AI.
Maybe some of the work on AI safety ends up helping produce mostly aligned AI. I’m not sure if everyone dies if an AI is 90% aligned.
Maybe the AGI systems that are built don’t have instrumental convergence maybe if we get AGI through CAIS which seems to me like the most likely way we’ll get there.
Maybe like physics once the low hanging fruit has been plucked then it takes a while to make breakthroughs which extends the timelines
For me to be personally afraid I’d have to think this was the primary way I would die which seems unlikely given all the other ways I could die between now and if/when AGI is developed.
AI researchers, who are the people that most likely believe that AGI is possible more than anyone else, don’t have consensus when it comes to this issue. I know experts can be wrong about their own fields but I’d expect them to be more split on the issue(I don’t know what the current status is now just know what it was in the Grace et. al survey). I know very little about AGI, should I be more concerned than AI researchers are?
I still think it’s important to work on AI Safety since even a small chance that AGI could go wrong would still have a high expected value in terms of the negative outcome. I think most of my thinking comes from the fact that I think it is more probable that there will be a slow take off instead of a fast take off. I may also just be bad at being scared or feeling doomed.
People start building AI that is agentic and open ended in its actions.
Yes, because I think the most likely scenario is a slow take off. This is because it costs money to scale compute and we actually need to validate and the more complex a system the harder it is to build correctly, probably takes a few iterations to get things to work well enough that it can be tested against a benchmark before moving on to trying to get a system to have more capability. I think this process will have to happen many times before getting to AI that is dangerous and on the way I’d expect to start seeing some interesting agentic behavior with short-horizon planning.
I think the uncertainty will be pretty high until we start seeing sophisticated agentic behavior. Though I don’t think we should wait that long to try come up with solutions since I think a small chance that this could happen still warrants concern.