Human values evolve in human ways. A priori, an AI’s value drift would almost surely take it in alien, worthless-to-us directions. A non-evolving AI sounds easier to align—we only need to hit the human-aligned region of valuespace once instead of needing to keep hitting it.
Human values evolve in human ways. A priori, an AI’s value drift would almost surely take it in alien, worthless-to-us directions. A non-evolving AI sounds easier to align—we only need to hit the human-aligned region of valuespace once instead of needing to keep hitting it.