This doesn’t really solve your problem, but it’s interesting that humans are also trying to create subagents. The whole AI problem is us trying to create subagents. It turns out that that is very very hard. And if we want to solve FAI, making subagents that actually follow our utility function, that’s even harder.
So humans are an existence proof for minds which are very powerful, but unable to make subagents. Controlling true superintelligences is a totally different issue of course. But maybe in some cases we can restrict them from being superintelligent?
Well it’s not impossible to restrict the AIs from accessing their own source code. Especially if they are implemented in specialized hardware like we are.
It’s not impossible, no. But it’s another failure point. And the AI might deduce stuff about itself by watching how it’s run. And a world that has built an AI is a world where there will be lots of tools for building AIs around...
This doesn’t really solve your problem, but it’s interesting that humans are also trying to create subagents. The whole AI problem is us trying to create subagents. It turns out that that is very very hard. And if we want to solve FAI, making subagents that actually follow our utility function, that’s even harder.
So humans are an existence proof for minds which are very powerful, but unable to make subagents. Controlling true superintelligences is a totally different issue of course. But maybe in some cases we can restrict them from being superintelligent?
We’d be considerably better at subagent creation if we could copy our brains and modify them at will...
Well it’s not impossible to restrict the AIs from accessing their own source code. Especially if they are implemented in specialized hardware like we are.
It’s not impossible, no. But it’s another failure point. And the AI might deduce stuff about itself by watching how it’s run. And a world that has built an AI is a world where there will be lots of tools for building AIs around...