I understood the “computation” theory as: there’s this abstract algorithm, approximately embedded in the unreliable hardware of my brain, and the morality judgments are its results, which are normally produced in the form of quick intuitions. But the algorithm is able to flexibly respond to arguments, etc. Then the observation of my brain thinking certain thoughts is how the algorithm feels from the inside.
I think it is at least a useful metaphor. You disagree? Do you have an exposition of your views on this?
Then the observation of my brain thinking certain thoughts is how the algorithm feels from the inside.
It’s some evidence about what the algorithm judges, but not the algorithm itself. Humans make errors, while morality is the criterion of correctness of judgment, which can’t be reliably observed by unaided eye, even if that’s the best we have.
I understood the “computation” theory as: there’s this abstract algorithm, approximately embedded in the unreliable hardware of my brain, and the morality judgments are its results, which are normally produced in the form of quick intuitions. But the algorithm is able to flexibly respond to arguments, etc. Then the observation of my brain thinking certain thoughts is how the algorithm feels from the inside.
I think it is at least a useful metaphor. You disagree? Do you have an exposition of your views on this?
It’s some evidence about what the algorithm judges, but not the algorithm itself. Humans make errors, while morality is the criterion of correctness of judgment, which can’t be reliably observed by unaided eye, even if that’s the best we have.