I’m not quite sure if ‘hard action like effectively monitoring it or controlling a necessary input to dangerous AI’ is realistic to implement everywhere there’s a concentration of researchers.
How do you envision this being realizable outside of extreme scenarios such as a world dictatorship?
I largely agree! Maybe we can get a stable policy regime of tracking hardware and auditing all large training runs with model evals that can identify unsafe systems. Maybe the US government can do intermediate stuff like tracking hardware and restricting training compute.
But mostly this frame is about raising questions or suggesting orientations or helping you notice if something appears.
(By the way, I roughly endorse this frame less than the others, which is why it’s at the end.)
Maybe we can get a stable policy regime of tracking hardware and auditing all large training runs with model evals that can identify unsafe systems. Maybe the US government can do intermediate stuff like tracking hardware and restricting training compute.
Wouldn’t this need to be done worldwide, near simultaneously, to be effective?
I’m not sure doing it in one country will move the needle much.
I’m still skeptical it’ll be anywhere near that easy. India, Japan, Korea, and some other countries are also coming on to the scene and will likely need to be included in any future deal.
Plus even when they’re at the negotiating table, the parties are incentivized to stall as long as possible to cut the best possible deal, because they won’t all be identically worried to the same degree.
And there’s nothing to stop them from walking away at any time. Even the folks in Europe may want to hold out if they feel like they can extract the maximum concessions elsewhere from both U.S. and China.
Thanks for the answer. Your frame is interesting.
I’m not quite sure if ‘hard action like effectively monitoring it or controlling a necessary input to dangerous AI’ is realistic to implement everywhere there’s a concentration of researchers.
How do you envision this being realizable outside of extreme scenarios such as a world dictatorship?
I largely agree! Maybe we can get a stable policy regime of tracking hardware and auditing all large training runs with model evals that can identify unsafe systems. Maybe the US government can do intermediate stuff like tracking hardware and restricting training compute.
But mostly this frame is about raising questions or suggesting orientations or helping you notice if something appears.
(By the way, I roughly endorse this frame less than the others, which is why it’s at the end.)
Wouldn’t this need to be done worldwide, near simultaneously, to be effective?
I’m not sure doing it in one country will move the needle much.
To some extent, yes, it would need US + Europe (including UK) + China. A strong treaty is necessary for some goals.
I’d guess that the US alone could buy a year.
One Western government doing something often causes other Western governments to do it.
(Edit in response to reply: I don’t think we have important disagreements here, so ending the conversation.)
I’m still skeptical it’ll be anywhere near that easy. India, Japan, Korea, and some other countries are also coming on to the scene and will likely need to be included in any future deal.
Plus even when they’re at the negotiating table, the parties are incentivized to stall as long as possible to cut the best possible deal, because they won’t all be identically worried to the same degree.
And there’s nothing to stop them from walking away at any time. Even the folks in Europe may want to hold out if they feel like they can extract the maximum concessions elsewhere from both U.S. and China.