Strong arguments of this kind? I sure hope not, that’d make it easier for more people to find insights for how to build an AI that causes doom.
Tarnish
Note that some of the best arguments are of the shape “AI will cause doom because it’s not that hard to build the following...” followed by insights about how to build an AI that causes doom. Those arguments are best rederived privately rather than shared publicly, and by asking publicly you’re filtering the strength of arguments you might get exposed to.
Unfortunately, that does not appear to be a stable solution. Even if the US paused its AI development, China or other countries could gain an advantage by accelerating their own work.
Arguing-for-pausing does not need to be a stable solution to help. If it buys time, that’s already helpful. If the US pauses AI development, but China doesn’t, that’s still less many people working on AI that might kill everyone.
Mu. The most basic rationalist precept is to not forcibly impose your values onto another mind.
It is? Last I heard, the two most basic precepts of rationality were:
Epistemic rationality: systematically improving the accuracy of your beliefs.
Instrumental rationality: systematically achieving your values.
(Typically with a note saying “ultimately, when at odds, the latter trumps the former”)
As far as I know, there is unfortunately no system for this. I think what people typically do is contact MIRI leadership, but I don’t know MIRI leadership to have particularly put silent people in touch with other silent people as a result.