It is speculative in the sense that any new technology being developed is speculative—but closely related approaches are already used for assurance in practice, so provable safety isn’t actually just speculative, there are concrete benefits in the near term. And I would challenge you to name a different and less speculative framework that actually deals with any issues of ASI risks that isn’t pure hopium.
Uncharitably, but I think not entirely inaccurately, these include: “maybe AI can’t be that much smarter than humans anyways,” “let’s get everyone to stop forever,” “we’ll use AI to figure it out, even though we have no real ideas,” “we just will trust that no-one makes it agentic,” “the agents will be able to be supervised by other AI which will magically be easier to align,” “maybe multiple AIs will compete in ways that isn’t a disaster,” “maybe we can just rely on prosaic approaches forever and nothing bad happens,” “maybe it will be better than humans at having massive amounts of unchecked power by default.” These all certainly seem to rely far more on speculative claims, with far less concrete ideas about how to validate or ensure them.
I’m not saying that it’s not worth pursuing as an agenda, but I also am not convinced it is promising enough to justify pursuing math related AI capabilities, compared to e.g. creating safety guarantees into which you can plug in AI capabilities once they arise anyway.
But “creating safety guarantees into which you can plug in AI capabilities once they arise anyway” is the point, and it requires at least some non-trivial advances in AI capabilities.
It is speculative in the sense that any new technology being developed is speculative—but closely related approaches are already used for assurance in practice, so provable safety isn’t actually just speculative, there are concrete benefits in the near term. And I would challenge you to name a different and less speculative framework that actually deals with any issues of ASI risks that isn’t pure hopium.
Uncharitably, but I think not entirely inaccurately, these include: “maybe AI can’t be that much smarter than humans anyways,” “let’s get everyone to stop forever,” “we’ll use AI to figure it out, even though we have no real ideas,” “we just will trust that no-one makes it agentic,” “the agents will be able to be supervised by other AI which will magically be easier to align,” “maybe multiple AIs will compete in ways that isn’t a disaster,” “maybe we can just rely on prosaic approaches forever and nothing bad happens,” “maybe it will be better than humans at having massive amounts of unchecked power by default.” These all certainly seem to rely far more on speculative claims, with far less concrete ideas about how to validate or ensure them.
I’m not saying that it’s not worth pursuing as an agenda, but I also am not convinced it is promising enough to justify pursuing math related AI capabilities, compared to e.g. creating safety guarantees into which you can plug in AI capabilities once they arise anyway.
But “creating safety guarantees into which you can plug in AI capabilities once they arise anyway” is the point, and it requires at least some non-trivial advances in AI capabilities.
You should probably read the current programme thesis.