Rationality is—or should be—for regular people, and very few regular people need to worry about the curvature of the universe in an average day.
Requiring rationality to be restricted to an aversion to edge-cases limits its usefulness to the point of being almost entirely without value.
To relate this more directly: that flat-spacetime thing is very relevant to understanding how “something” can come from “nothing”. Which touches on how we all got here—a very important, existentially speaking, question. One that can have an impact on even the ‘ordinary’ person’s ‘average day’. After all; if it turns out there’s no reason for anyone to believe in a God, then many of the things many people do or say on a daily basis become… extraneous at best.
Furthermore: one of the things that instrumental rationality as an approach needs to have in its “toolkit” is the ability to deeply examine thoughts, ideas, and events in advance and from those examinations create heuristics (“rules of thumb”) that enable us to make better decisions. That requires the use of sometimes very ‘technical’ turns of phrase. It’s simply unavoidable.
That gets all the more true when you’re trying to convey a very precise thought about a very nuanced topic. The thing is, regardless of where one looks in life there are more levels of complexity than we normally pay attention to. But that doesn’t make those levels of complexity irrelevant; it just means that we abstract that complexity away in our ‘average’ lives. Enter said heuristics.
Part of instrumental rationality as an approach, I believe, is the notion of at least occassionally breaking down into their constituent parts the various forms of complexity we usually ignore, in order to try to come up with better abstractions with which to ignore said complexity when it shouldn’t be a focus of our attention. I’ve gotten in “trouble” here on lesswrong for making similar statements before, however -- (though to add nuance that was more about whether generalizations are appropriate in a given ‘depth’ of conversation.)
“ever”: within the projected remaining longevity of anyone currently alive.
“average person”: A sufficient portion of people who are no more than 1 standard deviation away from the mode of any given manner of behavior as to be representative of the whole.
-- that being said: no, no I do not.
A different set of definitions:
“ever”: throughout the remainder of history
“an average person”: at least one person who is validly described as ‘average’ at the time it happens
-- Yes, yes I do.
Even explaining that took more nuance than you’d like, I suspect. Please note how radically different the two statements are, even though they both conform very closely to what you said. THIS is why nuance is sometimes indispensable.
Within our lifetimes, conversational speech will not resemble a legal document.
Not all conversations, no—but if an average person is unprepared for legalese then he’d better always have a lawyer with him when he signs anything, ever. This has an unhappy context for our conversatoin: is there a rationality-equivalent of a lawyer?
.
Requiring rationality to be restricted to an aversion to edge-cases limits its usefulness to the point of being almost entirely without value.
To relate this more directly: that flat-spacetime thing is very relevant to understanding how “something” can come from “nothing”. Which touches on how we all got here—a very important, existentially speaking, question. One that can have an impact on even the ‘ordinary’ person’s ‘average day’. After all; if it turns out there’s no reason for anyone to believe in a God, then many of the things many people do or say on a daily basis become… extraneous at best.
Furthermore: one of the things that instrumental rationality as an approach needs to have in its “toolkit” is the ability to deeply examine thoughts, ideas, and events in advance and from those examinations create heuristics (“rules of thumb”) that enable us to make better decisions. That requires the use of sometimes very ‘technical’ turns of phrase. It’s simply unavoidable.
That gets all the more true when you’re trying to convey a very precise thought about a very nuanced topic. The thing is, regardless of where one looks in life there are more levels of complexity than we normally pay attention to. But that doesn’t make those levels of complexity irrelevant; it just means that we abstract that complexity away in our ‘average’ lives. Enter said heuristics.
Part of instrumental rationality as an approach, I believe, is the notion of at least occassionally breaking down into their constituent parts the various forms of complexity we usually ignore, in order to try to come up with better abstractions with which to ignore said complexity when it shouldn’t be a focus of our attention. I’ve gotten in “trouble” here on lesswrong for making similar statements before, however -- (though to add nuance that was more about whether generalizations are appropriate in a given ‘depth’ of conversation.)
.
… Defining a few terms:
“ever”: within the projected remaining longevity of anyone currently alive.
“average person”: A sufficient portion of people who are no more than 1 standard deviation away from the mode of any given manner of behavior as to be representative of the whole.
-- that being said: no, no I do not.
A different set of definitions:
“ever”: throughout the remainder of history
“an average person”: at least one person who is validly described as ‘average’ at the time it happens
-- Yes, yes I do.
Even explaining that took more nuance than you’d like, I suspect. Please note how radically different the two statements are, even though they both conform very closely to what you said. THIS is why nuance is sometimes indispensable.
.
Not all conversations, no—but if an average person is unprepared for legalese then he’d better always have a lawyer with him when he signs anything, ever. This has an unhappy context for our conversatoin: is there a rationality-equivalent of a lawyer?
Relevant
The Order of Silent Confessors, maybe?
.
It is by my will alone that I set my mind in motion.
.
So far.
.