I’m just throwing around ideas here. I see this as obvious and not infohazards. I’m not comfortable with the totalitarian nature of the idea either, and I offer it as an alternative to the leading AGI literally controlling the whole world, which is the best viable solution to a multilateral AGI scenario I’ve heard to date.
We can’t take a slapdash solution forever. There are already open-source, locally runnable LLMs with no ethical training, and those will get better over time. Defense is harder than offense. We can hope to stop a few clever attempts at world destruction. But one will eventually slip through if it is possible for anyone to build and use an AGI for malevolent purposes.
I do have thoughts on how this might not lead to totalitarianism, but it definitely needs more thought.
I agree that superintelligent AIs running around is bad and can do damage, but why would the near-term risk posed by this particular development seem like itself cause for updating any particular way? Your post makes it sound like you are frightened by AutoGPT in particular and I don’t understand why.
I’m not frightened by autoGPT in particular. It’s the coming improvements to that approach that frighten me. But I’m even more encouraged by the alignment upsides if this becomes the dominant approach.
I’m just throwing around ideas here. I see this as obvious and not infohazards. I’m not comfortable with the totalitarian nature of the idea either, and I offer it as an alternative to the leading AGI literally controlling the whole world, which is the best viable solution to a multilateral AGI scenario I’ve heard to date.
We can’t take a slapdash solution forever. There are already open-source, locally runnable LLMs with no ethical training, and those will get better over time. Defense is harder than offense. We can hope to stop a few clever attempts at world destruction. But one will eventually slip through if it is possible for anyone to build and use an AGI for malevolent purposes.
I do have thoughts on how this might not lead to totalitarianism, but it definitely needs more thought.
I agree that superintelligent AIs running around is bad and can do damage, but why would the near-term risk posed by this particular development seem like itself cause for updating any particular way? Your post makes it sound like you are frightened by AutoGPT in particular and I don’t understand why.
I’m not frightened by autoGPT in particular. It’s the coming improvements to that approach that frighten me. But I’m even more encouraged by the alignment upsides if this becomes the dominant approach.