A used car salesperson convincing themselves that what they’re selling isn’t a piece of crud is an example of where irrationality is a “good” (effective) strategy. I don’t think that’s what we are trying to encourage here. That’s why I say instrumental truthiness—the truth part is important too.
I also maintain that focus on “winning” is psychologically in conflict with truth seeking. Politics = mind killer is best example.
I think the orthodox LW view would be that this used car salesperson might have an immoral utility function but that he isn’t irrational.
I also maintain that focus on “winning” is psychologically in conflict with truth seeking.
That basically means that sometimes the person who seeks the truth doesn’t win. That outcome isn’t satisfactory to Eliezer.
In Rationality is Systematized Winning he writes:
If the “irrational” agent is outcompeting you on a systematic and predictable basis, then it is time to reconsider what you think is “rational”.
Of course you can define rationality for yourself differently but it’s a mistake to project your own goals on others.
I am suprised that a significant group of people think that rationality is inclusive of useful false beliefs. Wouldn’t we call LW an effectiveness forum, rather than a rationalist forum in that case?
That basically means that sometimes the person who seeks the truth doesn’t win.
I think you’re reading too much into that one quite rhetorical article, but I acknowledge he prioritises “winning” quite highly. I think he ought to revise that view. Trying to win with false beliefs risks not achieving your goals, but being oblivious to that fact. Like a mad person killing their friends because he/she thinks they’ve turned into evil dog-headed creatures or some such (exaggeration to illustrate my point).
Of course you can define rationality for yourself differently but it’s a mistake to project your own goals on others.
Fair point. And maybe you’re right I’m in the minority… I’m still not yet certain. I do note that upvotes does not indicate agreement, only a feeling that the article is an interesting read etc. Also, I note many comments disagree with article. It warrants further investigation for me though.
I am suprised that a significant group of people think that rationality is inclusive of useful false beliefs.
Often they use “instrumental rationality” for that meaning and “epistemic rationality” for the other one. Searching this site for epistemic instrumental returns some relevant posts.
A used car salesperson convincing themselves that what they’re selling isn’t a piece of crud is an example of where irrationality is a “good” (effective) strategy. I don’t think that’s what we are trying to encourage here. That’s why I say instrumental truthiness—the truth part is important too.
I also maintain that focus on “winning” is psychologically in conflict with truth seeking. Politics = mind killer is best example.
I think the orthodox LW view would be that this used car salesperson might have an immoral utility function but that he isn’t irrational.
That basically means that sometimes the person who seeks the truth doesn’t win. That outcome isn’t satisfactory to Eliezer. In Rationality is Systematized Winning he writes:
Of course you can define rationality for yourself differently but it’s a mistake to project your own goals on others.
A recent article title Truth, it’s not that great got 84% upvotes on LW.
I am suprised that a significant group of people think that rationality is inclusive of useful false beliefs. Wouldn’t we call LW an effectiveness forum, rather than a rationalist forum in that case?
I think you’re reading too much into that one quite rhetorical article, but I acknowledge he prioritises “winning” quite highly. I think he ought to revise that view. Trying to win with false beliefs risks not achieving your goals, but being oblivious to that fact. Like a mad person killing their friends because he/she thinks they’ve turned into evil dog-headed creatures or some such (exaggeration to illustrate my point).
Fair point. And maybe you’re right I’m in the minority… I’m still not yet certain. I do note that upvotes does not indicate agreement, only a feeling that the article is an interesting read etc. Also, I note many comments disagree with article. It warrants further investigation for me though.
Often they use “instrumental rationality” for that meaning and “epistemic rationality” for the other one. Searching this site for
epistemic instrumental
returns some relevant posts.