I see what you mean, and I don’t think I disagree.
I think one more question will clarify. If your / our preferences were different, would the mathematical set / utility function you consider to be morality be different also? Namely, is the set of “rules that constitute morality” contingent upon what an agent already values (suitably extrapolated)?
No. On the other hand, me!pebble-sorter would have no interest in morality at all, and go on instead about how p-great p-morality is. But I wouldn’t mix up p-morality with morality.
So, you’re defining “morality” as an extrapolation from your preferences now, and if your preferences change in the future, that future person would care about what your present self might call futureYou-morality, even if future you insists on calling it “morality”?
I see what you mean, and I don’t think I disagree.
I think one more question will clarify. If your / our preferences were different, would the mathematical set / utility function you consider to be morality be different also? Namely, is the set of “rules that constitute morality” contingent upon what an agent already values (suitably extrapolated)?
No. On the other hand, me!pebble-sorter would have no interest in morality at all, and go on instead about how p-great p-morality is. But I wouldn’t mix up p-morality with morality.
So, you’re defining “morality” as an extrapolation from your preferences now, and if your preferences change in the future, that future person would care about what your present self might call futureYou-morality, even if future you insists on calling it “morality”?