Value drift (for uploads), misunderstanding of what we mean (for standard AI), and a whole host of other problems (a lot about this on Less Wrong). This seems to be a satisfacing solution that doesn’t seem immediately to blow up.
Gray as charged, but I also committed the fallacy of misreading the OP as saying “objective time”, which makes things look very close to black. Given OP’s record I should have read it over twice.
We can estimate, based on experience, that values are are unlikely to change much in one hour, for humans with stable, thoughtful personalities in stable environments with moderate stimuli.
Value drift (for uploads), misunderstanding of what we mean (for standard AI), and a whole host of other problems (a lot about this on Less Wrong). This seems to be a satisfacing solution that doesn’t seem immediately to blow up.
Interesting proposition, but how could one possibly know whether 1hr is enough to prevent value drift?
You have committed the fallacy of gray. Even where certainty in unattainable, one hour has less expected opportunity for moral drift than 1000 years.
Gray as charged, but I also committed the fallacy of misreading the OP as saying “objective time”, which makes things look very close to black. Given OP’s record I should have read it over twice.
We can estimate, based on experience, that values are are unlikely to change much in one hour, for humans with stable, thoughtful personalities in stable environments with moderate stimuli.