Humans got trained via evolution alongside a bunch of dumber animals. Then we killed a lot of them.
Evolution doesn’t align with anything other than differential reproduction rates, so you’d somehow have to make the only way to reproduce to be aligned with human values, which basically sounds like solving alignment and then throwing evolution on top for funsies.
From my very spotty info on evolution: Humans got ‘trained’ to maximise reproducibility and in doing so maximised a bunch of other stuff along the way- including resource acquisition.
What I spoke about here is creating an environment where a more intelligent+fast agent is put in an environment that is deliberately crafted such that it can only survive by helping much dumber, slower agents. Training to act co-operatively.
Writing this out, I may have just made an overcomplicated version of reinforcement learning.
Humans got trained via evolution alongside a bunch of dumber animals. Then we killed a lot of them.
Evolution doesn’t align with anything other than differential reproduction rates, so you’d somehow have to make the only way to reproduce to be aligned with human values, which basically sounds like solving alignment and then throwing evolution on top for funsies.
From my very spotty info on evolution:
Humans got ‘trained’ to maximise reproducibility and in doing so maximised a bunch of other stuff along the way- including resource acquisition.
What I spoke about here is creating an environment where a more intelligent+fast agent is put in an environment that is deliberately crafted such that it can only survive by helping much dumber, slower agents. Training to act co-operatively.
Writing this out, I may have just made an overcomplicated version of reinforcement learning.