If we were “doomed” in this way, would you agree that the thing to do—for those who could do it—is to keep trying to solve the problem of alignment? i.e. trying to identify an AI design that could be autonomous, and smarter than human, and yet still safe?
Let me articulate my intuitions in a little bit more of a refined way: “If we ever get to a point where there are few secrets left, or that it’s common knowledge one can solve AGI with ~1000-10,000 million dollars, then delaying tactics probably wouldn’t work, because there’s nothing left for DeepMind to publish that speeds up the timeline.”
Inside those bounds, yes. I still think that people should keep working on alignment today, I just think other dumber people like me should try the delaying tactics I articulated in addition to funding alignment research.
If we were “doomed” in this way, would you agree that the thing to do—for those who could do it—is to keep trying to solve the problem of alignment? i.e. trying to identify an AI design that could be autonomous, and smarter than human, and yet still safe?
Let me articulate my intuitions in a little bit more of a refined way: “If we ever get to a point where there are few secrets left, or that it’s common knowledge one can solve AGI with ~1000-10,000 million dollars, then delaying tactics probably wouldn’t work, because there’s nothing left for DeepMind to publish that speeds up the timeline.”
Inside those bounds, yes. I still think that people should keep working on alignment today, I just think other dumber people like me should try the delaying tactics I articulated in addition to funding alignment research.