Could be or could be not—the original example is quite barebones and we can read different interpretations into it. But in any case that seems irrelevant: we are not talking about the difference between instrumental and terminal goals, we are talking about the choice between two agents/proxies one of which has a closer value system and the other is more effective at achieving his goals.
we are talking about the choice between two agents/proxies one of which has a closer value system and the other is more effective at achieving his goals.
If the two environmentalists had a debate about this subject, each could start the debate by saying they want to do whatever is best for the environment. And then each could present a series of facts suggesting that nuclear power either is or is not good for the environment—a factual disagreement about what the right instrumental goal is for achieving the terminal goal of helping the environment.
If you think anti-nuclear environmentalists possess lack of nuclear plants as a terminal value, imagine what would happen if one was convinced of the factual belief that nuclear power is actually good for the environment. If your model is correct, we can imagine that they would continue to be anti-nuclear environmentalists because that’s their terminal goal (while acknowledging that nuclear power is actually the best option for the environment). But we have counterexamples like Stewart Brand who switched from anti-nuclear to pro-nuclear after doing research & having their beliefs change.
Actual human beings’ goals don’t divide neatly into instrumental and terminal, and actual humans can be inconsistent. So you can have someone who has instrumental goals (that can be changed with evidence showing that they don’t meet a terminal goal), terminal goals (which cannot), and inbetween goals like nuclear power that are harder to change than the former category, but easier to change than the latter.
inbetween goals like nuclear power that are harder to change than the former category, but easier to change than the latter
Yep, this is kinda one of the things LW specializes in—helping people become better at changing their minds regarding things they are stubbornly wrong about.
I agree that human beings’ goals don’t neatly divide in to instrumental and terminal. This is just a model we use. I think Lumifer is using the model in a way that’s harmful—labeling stubborn incorrect beliefs as “terminal goals” amounts to throwing up your hands and saying it’s impossible to help people become better at changing their minds. Based on the what I’ve seen, this isn’t the case—although it’s difficult, it is possible to help people become better at changing their minds, and accomplishing this is highly valuable.
If the two environmentalists had a debate about this subject
This is not what this subthread is about. It started with me saying
What you want above all in a political leader is that his value system be aligned to yours. If it is not, the fact that he is effective at reaching his goals becomes a threat, not a benefit.
and skeptical_lurker pointing out that
Only if the goals are actually opposed to yours.
and me continuing with
I still think that similar values are MUCH more important than the capability to execute.
I don’t see how trying to tease apart terminal and instrumental goals is relevant to this issue. I also think that in practice many theoretically-instrumental goals are, in fact, terminal. Stewart Brand changed his mind, but a great deal more people didn’t and I am willing to argue that for at least some and probably many of them the opposition to the nuclear effectively became a terminal goal (along the “when you forget your goal you redouble your efforts” lines).
So there are two models I can have of politicians who advocate policies different from mine. The first is that we have different terminal goals—even though our model of the world is quite similar, in the sense that we agree about which policies would create which outcomes, we differ on which outcomes we prefer to create. The second is that we have different beliefs—for example, you think raising the minimum wage would be on net beneficial for the working class, whereas I think it’s likely to increase unemployment.
These two models suggest different strategies for people who have political disagreements. The first model suggests all-out war: take down the people who have different values from you at any cost through rhetoric, dirty tactics, etc. The second model suggests trying to improve your rationality and their rationality so your beliefs are less stubborn, you can see the world more accurately, and you can better achieve your collective values.
I also think that in practice many theoretically-instrumental goals are, in fact, terminal.
I don’t think this is the right model for something like an incorrect belief in nuclear power being bad. I think it’s more accurate to say that someone has a visceral disgust for nuclear power, or all their friends think nuclear power is bad, or whatever. Labeling incorrect beliefs as terminal goals basically makes them in to black boxes where investigating how the incorrect belief formed is a waste of time. The advantage of investigating how the incorrect belief formed is that we can learn how to prevent incorrect beliefs from forming in ourselves and others. That’s basically the project of this site.
I tend to think of it as a religion, but let’s be charitable and call it a set of (often inconsistent) preferences. For example, some people prefer not to live near a nuclear plant. How is it a goal?
Well, to a large extent you are right and much of environmentalism is quasi-religious. However, being charitable, consider the specific goal of maintaining an environment suitable for human habitation. Having a government that knows whether nuclear or coal power is more dangerous is very important here.
Could be or could be not—the original example is quite barebones and we can read different interpretations into it. But in any case that seems irrelevant: we are not talking about the difference between instrumental and terminal goals, we are talking about the choice between two agents/proxies one of which has a closer value system and the other is more effective at achieving his goals.
If the two environmentalists had a debate about this subject, each could start the debate by saying they want to do whatever is best for the environment. And then each could present a series of facts suggesting that nuclear power either is or is not good for the environment—a factual disagreement about what the right instrumental goal is for achieving the terminal goal of helping the environment.
If you think anti-nuclear environmentalists possess lack of nuclear plants as a terminal value, imagine what would happen if one was convinced of the factual belief that nuclear power is actually good for the environment. If your model is correct, we can imagine that they would continue to be anti-nuclear environmentalists because that’s their terminal goal (while acknowledging that nuclear power is actually the best option for the environment). But we have counterexamples like Stewart Brand who switched from anti-nuclear to pro-nuclear after doing research & having their beliefs change.
Actual human beings’ goals don’t divide neatly into instrumental and terminal, and actual humans can be inconsistent. So you can have someone who has instrumental goals (that can be changed with evidence showing that they don’t meet a terminal goal), terminal goals (which cannot), and inbetween goals like nuclear power that are harder to change than the former category, but easier to change than the latter.
Yep, this is kinda one of the things LW specializes in—helping people become better at changing their minds regarding things they are stubbornly wrong about.
I agree that human beings’ goals don’t neatly divide in to instrumental and terminal. This is just a model we use. I think Lumifer is using the model in a way that’s harmful—labeling stubborn incorrect beliefs as “terminal goals” amounts to throwing up your hands and saying it’s impossible to help people become better at changing their minds. Based on the what I’ve seen, this isn’t the case—although it’s difficult, it is possible to help people become better at changing their minds, and accomplishing this is highly valuable.
This is not what this subthread is about. It started with me saying
and skeptical_lurker pointing out that
and me continuing with
I don’t see how trying to tease apart terminal and instrumental goals is relevant to this issue. I also think that in practice many theoretically-instrumental goals are, in fact, terminal. Stewart Brand changed his mind, but a great deal more people didn’t and I am willing to argue that for at least some and probably many of them the opposition to the nuclear effectively became a terminal goal (along the “when you forget your goal you redouble your efforts” lines).
So there are two models I can have of politicians who advocate policies different from mine. The first is that we have different terminal goals—even though our model of the world is quite similar, in the sense that we agree about which policies would create which outcomes, we differ on which outcomes we prefer to create. The second is that we have different beliefs—for example, you think raising the minimum wage would be on net beneficial for the working class, whereas I think it’s likely to increase unemployment.
These two models suggest different strategies for people who have political disagreements. The first model suggests all-out war: take down the people who have different values from you at any cost through rhetoric, dirty tactics, etc. The second model suggests trying to improve your rationality and their rationality so your beliefs are less stubborn, you can see the world more accurately, and you can better achieve your collective values.
I don’t think this is the right model for something like an incorrect belief in nuclear power being bad. I think it’s more accurate to say that someone has a visceral disgust for nuclear power, or all their friends think nuclear power is bad, or whatever. Labeling incorrect beliefs as terminal goals basically makes them in to black boxes where investigating how the incorrect belief formed is a waste of time. The advantage of investigating how the incorrect belief formed is that we can learn how to prevent incorrect beliefs from forming in ourselves and others. That’s basically the project of this site.
I agree with hg00 - I meant that environmentalism is a goal and nuclear power is a means to an end, not a value in itself.
In which sense environmentalism is a goal?
I tend to think of it as a religion, but let’s be charitable and call it a set of (often inconsistent) preferences. For example, some people prefer not to live near a nuclear plant. How is it a goal?
Well, to a large extent you are right and much of environmentalism is quasi-religious. However, being charitable, consider the specific goal of maintaining an environment suitable for human habitation. Having a government that knows whether nuclear or coal power is more dangerous is very important here.