For my own views, I think it’s often the case that people bias towards not changing their mind. This feels like part of the impetus for the push-back early on in stressing the benefit in changing your mind. (Duncan Sabien has a neat simple analogy of a pendulum getting counterbalanced back and forth on this).
However, I don’t think I’ve even really figured out the trick of changing my mind yet. There have been a few instances recently where nothing felt very sacred, and that sort of fluidity felt like it could be a powerful asset when trying to updated towards certain things in reality.
Hi, thanks for the comment! Yep, I agree that people often bias towards not changing their minds, and that this means it’s probably better to focus on improving in that direction for the moment (I think Duncan’s pendulum analogy is spot-on). When I was thinking about this originally, I framed it more as “why do settled beliefs feel so appealing to one part of me when they’re so obviously a bad idea”, I.e. looking more for an explanation than a (partial) justification. I was originally going to frame it like that here too, but it felt like a bit of a just-so story so I didn’t want to speculate about whether it was the actual explanation (e.g. evolutionarily).
Hurrah for taking the exam!
For my own views, I think it’s often the case that people bias towards not changing their mind. This feels like part of the impetus for the push-back early on in stressing the benefit in changing your mind. (Duncan Sabien has a neat simple analogy of a pendulum getting counterbalanced back and forth on this).
However, I don’t think I’ve even really figured out the trick of changing my mind yet. There have been a few instances recently where nothing felt very sacred, and that sort of fluidity felt like it could be a powerful asset when trying to updated towards certain things in reality.
Hi, thanks for the comment! Yep, I agree that people often bias towards not changing their minds, and that this means it’s probably better to focus on improving in that direction for the moment (I think Duncan’s pendulum analogy is spot-on). When I was thinking about this originally, I framed it more as “why do settled beliefs feel so appealing to one part of me when they’re so obviously a bad idea”, I.e. looking more for an explanation than a (partial) justification. I was originally going to frame it like that here too, but it felt like a bit of a just-so story so I didn’t want to speculate about whether it was the actual explanation (e.g. evolutionarily).