it feels wrong to call other research dangerous, especially given its enormous potential for good.
I agree that calling 99.9% of AI research “dangerous” and AI Safety research “safe” is not an useful dichotomy. However, I consider AGI companies/labs and people focusing on implementing self-improving AI/code synthesis extremely dangerous. Same for any breakthrough in general AI, or things that greatly shorten the AGI timeline.
Do you mean that some AI research have positive expected utility (e.g. in medecine) and should not be called dangerous because the good they produce compensates for the increased AI-risk?
Just to return for a moment to what I wrote, I don’t mean to be making an assessment here on “dangerous”, but instead to provide this service for things people themselves think are dangerous. Figuring out where to draw the line in what capabilities research is so dangerous it should not be published is a thing I have only very weak opinions on. For example, if you figured out how to make recursive self improvement work in a way that doesn’t immediately result in wild divergence and could stablely produce better results over many iterations I’d say that’s dangerous, but less than that I’m not sure where you might draw the line.
I agree that calling 99.9% of AI research “dangerous” and AI Safety research “safe” is not an useful dichotomy. However, I consider AGI companies/labs and people focusing on implementing self-improving AI/code synthesis extremely dangerous. Same for any breakthrough in general AI, or things that greatly shorten the AGI timeline.
Do you mean that some AI research have positive expected utility (e.g. in medecine) and should not be called dangerous because the good they produce compensates for the increased AI-risk?
Just to return for a moment to what I wrote, I don’t mean to be making an assessment here on “dangerous”, but instead to provide this service for things people themselves think are dangerous. Figuring out where to draw the line in what capabilities research is so dangerous it should not be published is a thing I have only very weak opinions on. For example, if you figured out how to make recursive self improvement work in a way that doesn’t immediately result in wild divergence and could stablely produce better results over many iterations I’d say that’s dangerous, but less than that I’m not sure where you might draw the line.