Why wouldn’t it be sufficient to solve the alignment problem by just figuring out exactly how the human brain works, and copying that? The result would at worst be no less aligned to human values than an average human. (Presuming of course that a psychopath’s brain was not the model used.)
The first plane didn’t emulate birds. The first AGI probably won’t be based on a retro engineering of the brain. The blue brain project is unlikely to finish reproducing the brain before DeepMind finds the right architecture.
But I agree that being able to retro engineer the brain is very valuable for alignment, this is one of the path described here, in the final post of intro-to-brain-like-agi-safety, section Reverse-engineer human social instincts.
Why wouldn’t it be sufficient to solve the alignment problem by just figuring out exactly how the human brain works, and copying that? The result would at worst be no less aligned to human values than an average human. (Presuming of course that a psychopath’s brain was not the model used.)
The first plane didn’t emulate birds. The first AGI probably won’t be based on a retro engineering of the brain. The blue brain project is unlikely to finish reproducing the brain before DeepMind finds the right architecture.
But I agree that being able to retro engineer the brain is very valuable for alignment, this is one of the path described here, in the final post of intro-to-brain-like-agi-safety, section Reverse-engineer human social instincts.