I think they can inspire you to change your values.
Taken at face value, this statement doesn’t make much sense because it immediately begs the question of change according to what, and in what sense isn’t that change part of your value already. My guess here is that your mental model says something like “there’s a set of primal drives inside my head like eating pizza that I call ‘values’, and then there are my ‘true’ values like a healthy lifestyle which my conscious, rational mind posits, and I should change my primal drives to match my ‘true’ values” (pardon for straw-manning your position, but I need it to make my point).
A much better model in my opinion would be that all these values belong to the same exact category. These “values” or “drives” then duke it out amongst each other, and your conscious mind merely observes and makes up a plausible-sounding socially-acceptable story about your motivations (this is, after all, the evolutionary function of human intelligence in the first place as far as I know), like a press secretary sitting silently in the corner while generals are having a heated debate.
At best, your conscious mind might act as a mediator between these generals, coming up with clever ideas that pushes the Pareto boundary of these competing values so that they can all be satisfied to a greater degree at the same time. Things like “let’s try e-cigarettes instead of regular tobacco—maybe it satisfies both our craving for nicotine and our long-term health!”.
Even high-falutin values like altruism or long-term health are induced by basic drives like empathy and social status. They are no different to, say, food cravings, not even in terms of inferential distance. Compare for instance “I ate pizza, it was tasty and I felt good” with “I was chastised for eating unhealthily, it felt bad”. Is there really any important difference here?
You could of course deny this categorization and insist that only a part of this value set represents your true values. The danger here isn’t that you’ll end up optimizing for the wrong set of values since who’s to tell you what “wrong” is, it’s that you’ll be perpetually confused about why you keeping failing to act upon your declared “true” values—why your revealed preferences through behavior keep diverging from the stated preferences, and you end up making bad decisions. Decisions that are suboptimal even when judged only against your “true” values, because you have not been leveraging your conscious, rational mind properly by giving it bad epistemics.
As an example, consider an immature teenager who doesn’t care at all about his future self and just wants to have fun right now. Would you say, “Well, he values what he values.”?
Haha, unfortunately you posed the question to the one guy out of 100 who would gladly answer “Absolutely”, followed by “What’s wrong with being an immature teenager?”
On a more serious note, it is true that our values often shift over time, but it’s unclear to me why that makes regret minimization the correct heuristic. Regret can occur in two ways: One is that we have better information later in life, along the lines of “Oh I should have picked these numbers in last week’s lottery instead of the numbers I actually picked”. But this is just hindsight and useless to your current self because you don’t have access to that knowledge.
The other is through value shift, along the lines of “I just ate a whole pizza and now that my food-craving brain-subassembly has shut up my value function consists mostly of concerns for my long-term health”. Even setting temporal discounting aside, I fail to see why your post-dinner-values should take precedence over your pre-dinner-values, or for that matter why deathbed-values should take precedence over teenage-values. They are both equally real moments of conscious experience.
But, since we only ever live and make decisions in the present moment, if you happen to have just finished a pizza, you now have the opportunity to manipulate your future values to match your current values by taking actions that makes the salad option more available the next time pizza-craving comes around by e.g. shopping for ingredients. In AI lingo, you’ve just made yourself subagent-stable.
My personal anecdote is that as a teenager I did listen to the “mature adults” to study more and spend less time having fun. It was a bad decision according to both my current values and teenage-values, made out of ignorance about how the world operates.
As a final thought, I would give the meta-advice of not trying to think too deeply about normative ethics. Take AlphaGo as a cautionary tale: after 2000 years of pondering, the deepest truths of Go are revealed to be just a linear combination of a bunch of feature vectors. Quite poetic, if you ask me.
I think I may have lead us down the wrong path here. The ultimate question is the one of temporal discounting, and that question depends on how much we do/should value those post-singularity life years. If values can’t shift, then there isn’t really anything to talk about; you just ask yourself how much you value those years, and then move on. But if they can shift, and you acknowledge that they can, then we can discuss some thought experiments and stuff. It doesn’t seem important to discuss whether those shifts are due to discovering more about your pre-existing values, or due to actually changing those pre-existing values.
Haha, unfortunately you posed the question to the one guy out of 100 who would gladly answer “Absolutely”, followed by “What’s wrong with being an immature teenager?”
Ah, I see. You and I probably just have very different intuitions regarding what to value then, and I sense that thought experiments won’t bring us much closer.
Actually, I wonder what you think of this. Are you someone who sees death as a wildlyterriblething (I am)? If so, isn’t it because you place a correspondingly high value on the years of life you’d be losing?
The other is through value shift, along the lines of “I just ate a whole pizza and now that my food-craving brain-subassembly has shut up my value function consists mostly of concerns for my long-term health”. Even setting temporal discounting aside, I fail to see why your post-dinner-values should take precedence over your pre-dinner-values, or for that matter why deathbed-values should take precedence over teenage-values. They are both equally real moments of conscious experience.
In the pizza example, I think the value shift would moreso be along the lines of “I was prioritizing my current self too much relative to my future selves”. Presumably, post-dinner-values would be incorporating pre-dinner-self. Eg. it wouldn’t just say, “Screw my past self, my values are only about the present moment and onwards.” So I see your current set of values as being the most “accurate”, in which case regret minimization seems like it makes sense.
The ultimate question is the one of temporal discounting, and that question depends on how much we do/should value those post-singularity life years. If values can’t shift, then there isn’t really anything to talk about; you just ask yourself how much you value those years, and then move on. But if they can shift, and you acknowledge that they can, then we can discuss some thought experiments and stuff.
I think we’re getting closer to agreement as I’m starting to see what you’re getting at. My comment here would be that yes, your values can shift, and they have shifted after thinking hard about what post-Singularity life will be like and getting all excited. But the shift it has caused is a larger multiplier in front of the temporal discounted integral, not the disabling of temporal discounting altogether.
Actually, I wonder what you think of this. Are you someone who sees death as a wildlyterriblething (I am)?
Yes, but I don’t think there is any layer of reasoning beneath that preference. Evading death is just something that is very much hard-coded into us by evolution.
In the pizza example, I think the value shift would moreso be along the lines of “I was prioritizing my current self too much relative to my future selves”. Presumably, post-dinner-values would be incorporating pre-dinner-self.
I don’t think that’s true. Crucially, there is no knowledge being gained over the course of dinner, only value shift. It’s not like you didn’t know beforehand that pizza was unhealthy, or that you will regret your decision. And if post-dinner self does not take explicit steps to manipulate future value, the situation will repeat itself the next day, and the day after, and so on for hundreds of times.
I think we’re getting closer to agreement as I’m starting to see what you’re getting at. My comment here would be that yes, your values can shift, and they have shifted after thinking hard about what post-Singularity life will be like and getting all excited. But the shift it has caused is a larger multiplier in front of the temporal discounted integral, not the disabling of temporal discounting altogether.
I’m in agreement here! Some follow up questions: what are your thoughts on how much discounting should be done? Relatedly, what are your thoughts on how much we should value life? Is it obvious that past eg. 500 years, it’s far enough into the future that it becomes negligible? If not, why aren’t these things that are discussed? Also, do you share my impression that people (on LW) largely assume that life expectancy is something like 80 years and life is valued at something like $10M?
Yes, but I don’t think there is any layer of reasoning beneath that preference. Evading death is just something that is very much hard-coded into us by evolution.
Regardless of whether it stems from a layer of reasoning or whether it is hard-coded, doesn’t it imply that you aren’t doing too much temporal discounting? If you did a lot of temporal discounting and didn’t value the years beyond eg. 250 years old very much, then death wouldn’t be that bad, right?
Taken at face value, this statement doesn’t make much sense because it immediately begs the question of change according to what, and in what sense isn’t that change part of your value already. My guess here is that your mental model says something like “there’s a set of primal drives inside my head like eating pizza that I call ‘values’, and then there are my ‘true’ values like a healthy lifestyle which my conscious, rational mind posits, and I should change my primal drives to match my ‘true’ values” (pardon for straw-manning your position, but I need it to make my point).
A much better model in my opinion would be that all these values belong to the same exact category. These “values” or “drives” then duke it out amongst each other, and your conscious mind merely observes and makes up a plausible-sounding socially-acceptable story about your motivations (this is, after all, the evolutionary function of human intelligence in the first place as far as I know), like a press secretary sitting silently in the corner while generals are having a heated debate.
At best, your conscious mind might act as a mediator between these generals, coming up with clever ideas that pushes the Pareto boundary of these competing values so that they can all be satisfied to a greater degree at the same time. Things like “let’s try e-cigarettes instead of regular tobacco—maybe it satisfies both our craving for nicotine and our long-term health!”.
Even high-falutin values like altruism or long-term health are induced by basic drives like empathy and social status. They are no different to, say, food cravings, not even in terms of inferential distance. Compare for instance “I ate pizza, it was tasty and I felt good” with “I was chastised for eating unhealthily, it felt bad”. Is there really any important difference here?
You could of course deny this categorization and insist that only a part of this value set represents your true values. The danger here isn’t that you’ll end up optimizing for the wrong set of values since who’s to tell you what “wrong” is, it’s that you’ll be perpetually confused about why you keeping failing to act upon your declared “true” values—why your revealed preferences through behavior keep diverging from the stated preferences, and you end up making bad decisions. Decisions that are suboptimal even when judged only against your “true” values, because you have not been leveraging your conscious, rational mind properly by giving it bad epistemics.
Haha, unfortunately you posed the question to the one guy out of 100 who would gladly answer “Absolutely”, followed by “What’s wrong with being an immature teenager?”
On a more serious note, it is true that our values often shift over time, but it’s unclear to me why that makes regret minimization the correct heuristic. Regret can occur in two ways: One is that we have better information later in life, along the lines of “Oh I should have picked these numbers in last week’s lottery instead of the numbers I actually picked”. But this is just hindsight and useless to your current self because you don’t have access to that knowledge.
The other is through value shift, along the lines of “I just ate a whole pizza and now that my food-craving brain-subassembly has shut up my value function consists mostly of concerns for my long-term health”. Even setting temporal discounting aside, I fail to see why your post-dinner-values should take precedence over your pre-dinner-values, or for that matter why deathbed-values should take precedence over teenage-values. They are both equally real moments of conscious experience.
But, since we only ever live and make decisions in the present moment, if you happen to have just finished a pizza, you now have the opportunity to manipulate your future values to match your current values by taking actions that makes the salad option more available the next time pizza-craving comes around by e.g. shopping for ingredients. In AI lingo, you’ve just made yourself subagent-stable.
My personal anecdote is that as a teenager I did listen to the “mature adults” to study more and spend less time having fun. It was a bad decision according to both my current values and teenage-values, made out of ignorance about how the world operates.
As a final thought, I would give the meta-advice of not trying to think too deeply about normative ethics. Take AlphaGo as a cautionary tale: after 2000 years of pondering, the deepest truths of Go are revealed to be just a linear combination of a bunch of feature vectors. Quite poetic, if you ask me.
I think I may have lead us down the wrong path here. The ultimate question is the one of temporal discounting, and that question depends on how much we do/should value those post-singularity life years. If values can’t shift, then there isn’t really anything to talk about; you just ask yourself how much you value those years, and then move on. But if they can shift, and you acknowledge that they can, then we can discuss some thought experiments and stuff. It doesn’t seem important to discuss whether those shifts are due to discovering more about your pre-existing values, or due to actually changing those pre-existing values.
Ah, I see. You and I probably just have very different intuitions regarding what to value then, and I sense that thought experiments won’t bring us much closer.
Actually, I wonder what you think of this. Are you someone who sees death as a wildly terrible thing (I am)? If so, isn’t it because you place a correspondingly high value on the years of life you’d be losing?
In the pizza example, I think the value shift would moreso be along the lines of “I was prioritizing my current self too much relative to my future selves”. Presumably, post-dinner-values would be incorporating pre-dinner-self. Eg. it wouldn’t just say, “Screw my past self, my values are only about the present moment and onwards.” So I see your current set of values as being the most “accurate”, in which case regret minimization seems like it makes sense.
I think we’re getting closer to agreement as I’m starting to see what you’re getting at. My comment here would be that yes, your values can shift, and they have shifted after thinking hard about what post-Singularity life will be like and getting all excited. But the shift it has caused is a larger multiplier in front of the temporal discounted integral, not the disabling of temporal discounting altogether.
Yes, but I don’t think there is any layer of reasoning beneath that preference. Evading death is just something that is very much hard-coded into us by evolution.
I don’t think that’s true. Crucially, there is no knowledge being gained over the course of dinner, only value shift. It’s not like you didn’t know beforehand that pizza was unhealthy, or that you will regret your decision. And if post-dinner self does not take explicit steps to manipulate future value, the situation will repeat itself the next day, and the day after, and so on for hundreds of times.
I’m in agreement here! Some follow up questions: what are your thoughts on how much discounting should be done? Relatedly, what are your thoughts on how much we should value life? Is it obvious that past eg. 500 years, it’s far enough into the future that it becomes negligible? If not, why aren’t these things that are discussed? Also, do you share my impression that people (on LW) largely assume that life expectancy is something like 80 years and life is valued at something like $10M?
Regardless of whether it stems from a layer of reasoning or whether it is hard-coded, doesn’t it imply that you aren’t doing too much temporal discounting? If you did a lot of temporal discounting and didn’t value the years beyond eg. 250 years old very much, then death wouldn’t be that bad, right?