There’s a point intermediate between “completely new” and “just being difficult”.
Fair enough. To me, your previous words pattern matched very strongly to ‘being difficult because they think this is dumb but don’t want to say why because it seems like too much work’ (or something). My mistake.
I didn’t mean new to LW, I meant new to the questions you were posing and the answers you got.
Back on the topic at hand,
In order to do that I would have to assume that I know what questions are the right ones and that he does not. Assuming this would amount to assuming that I am right about the subject and he is wrong.
Consider the following: you meet a friend of a friend who seems reasonable enough, and they start telling you about their startup. They go on and on for a long time but try as you might, you can’t figure out how on earth they’re going to make money. Finally, you delicately ask “how do you intend to make money?”. They give some wishy washy answer.
Here they have failed to ask a question that you know to be important. You know this quite definitely. Even if they thought that the question were somehow not relevant, if they knew it was usually relevant, they would probably explain why its not in this particular case. Much more likely that they are just not very good at thinking about startups.
Similarly, if they anticipate all of your objections and questions, you will probably think they are being pretty reasonable and be inclined to take them more seriously. And rightfully so, that’s actually decent evidence.
in which case I am again assuming I am right about the subject
There’s a middle ground between ‘assuming I am right’ and ‘assuming they are right’. You can instead be unsure how likely they are to be right, and try to figure it out. One way you can figure it out is by trying to assess whether they seem like they are doing good epistemic things (like do they actually pause to think about things, do they try to understand people’s points, do they respond to the actual question, do they make arguments that later turn out to be convincing, do they base things on believable numbers, do they present actual evidence for their views, etc. etc.)
Are you familiar with the idea of ‘latent variables’ from Bayesian statistics? Are you used to thinking about it in the context of people and the real world? The basic idea is that you can infer hidden properties of things by observing many things it affects (even if it only noisily affects them).
For example, you go to a small school and observe many students doing very impressive science experiments, you might then infer some hidden cause that causes the school to have smart students. Thus you might also guess that in several years, different students at the same school will do well on their SATs, even though that’s not directly related to your actual observations.
I suspect thinking a bunch about latent variables in the real world might be useful for you. Especially as it relates to inferring where people are reasonable and how much they are. Especially the idea of using data from different topics to improve your estimate for a given topic (say using test scores from different students to improve your quality estimate for a specific student).
Here they have failed to ask a question that you know to be important.
How do I know that the question is important, though? I can’t just assume it to be so, or we get the same problem I pointed out—my conclusion that he is worse than me is just being forced by my assumptions.
Of course, if my friend says “yeah, that’s important—why didn’t I think of that?” then my conclusion is fine. But I think that’s going to be pretty rare among creationists, homeopaths, and people who think Jews eat babies.
...thinking a bunch about latent variables in the real world...
That’s why I distinguished not being sensible on a topic and not being sensible in general.
If someone is generally not sensible, I can use facts from outside a particular area to conclude that he won’t be sensible within a particular area. That’s basically using latent variables.
If someone compartmentalizes his lack of sense, so he’s only not sensible in one area (for instance, a creationist who is perfectly fine at calculating restaurant tips), this isn’t going to work.
Are you primarily thinking of this as applying to creationists etc?
The part of the reason I put the caveat ‘people about as reasonable as you’ in the first place was to exclude that category of people from what I was talking about.
That is not the central category of people I’m suggesting this for. Also, I’m not clear on why you would think it was.
I’m not using creationists as an example because it’s central; I’m using it as an example because it’s unambiguous. It’s really hard to sidetrack the argument by suggesting that maybe the creationists are right after all, or that I’m being arrogant by thinking the creationists are mistaken, etc. so creationists work well as an example.
(And an idea that works for central examples but fails for edge cases is an idea that fails.)
The part of the reason I put the caveat ‘people about as reasonable as you’ in the first place was to exclude that category of people from what I was talking about.
But if you add that exception, it swallows the rule. Most people think their opponents are more unreasonable than themselves.
(And an idea that works for central examples but fails for edge cases is an idea that fails.)
Ironically, this is not a universal criteria for the success of ideas. Sometimes its a very useful criteria (think mathematical proofs). Other times, its not a very useful idea (think ‘choosing friends’ or ‘mathematical intuitions’).
For example the idea of ‘cat’ fails for edge cases. Is this a cat? Sort of. Sort of not. But ‘cat’ is still a useful concept.
Concepts are clusters in thing space, and the concept that I am pointing at is also a cluster.
Fair enough. To me, your previous words pattern matched very strongly to ‘being difficult because they think this is dumb but don’t want to say why because it seems like too much work’ (or something). My mistake.
I didn’t mean new to LW, I meant new to the questions you were posing and the answers you got.
Back on the topic at hand,
Consider the following: you meet a friend of a friend who seems reasonable enough, and they start telling you about their startup. They go on and on for a long time but try as you might, you can’t figure out how on earth they’re going to make money. Finally, you delicately ask “how do you intend to make money?”. They give some wishy washy answer.
Here they have failed to ask a question that you know to be important. You know this quite definitely. Even if they thought that the question were somehow not relevant, if they knew it was usually relevant, they would probably explain why its not in this particular case. Much more likely that they are just not very good at thinking about startups.
Similarly, if they anticipate all of your objections and questions, you will probably think they are being pretty reasonable and be inclined to take them more seriously. And rightfully so, that’s actually decent evidence.
There’s a middle ground between ‘assuming I am right’ and ‘assuming they are right’. You can instead be unsure how likely they are to be right, and try to figure it out. One way you can figure it out is by trying to assess whether they seem like they are doing good epistemic things (like do they actually pause to think about things, do they try to understand people’s points, do they respond to the actual question, do they make arguments that later turn out to be convincing, do they base things on believable numbers, do they present actual evidence for their views, etc. etc.)
Are you familiar with the idea of ‘latent variables’ from Bayesian statistics? Are you used to thinking about it in the context of people and the real world? The basic idea is that you can infer hidden properties of things by observing many things it affects (even if it only noisily affects them).
For example, you go to a small school and observe many students doing very impressive science experiments, you might then infer some hidden cause that causes the school to have smart students. Thus you might also guess that in several years, different students at the same school will do well on their SATs, even though that’s not directly related to your actual observations.
I suspect thinking a bunch about latent variables in the real world might be useful for you. Especially as it relates to inferring where people are reasonable and how much they are. Especially the idea of using data from different topics to improve your estimate for a given topic (say using test scores from different students to improve your quality estimate for a specific student).
This might be a good starting point: http://www.stat.columbia.edu/~gelman/research/published/multi2.pdf (read until sec 2.3).
How do I know that the question is important, though? I can’t just assume it to be so, or we get the same problem I pointed out—my conclusion that he is worse than me is just being forced by my assumptions.
Of course, if my friend says “yeah, that’s important—why didn’t I think of that?” then my conclusion is fine. But I think that’s going to be pretty rare among creationists, homeopaths, and people who think Jews eat babies.
That’s why I distinguished not being sensible on a topic and not being sensible in general.
If someone is generally not sensible, I can use facts from outside a particular area to conclude that he won’t be sensible within a particular area. That’s basically using latent variables.
If someone compartmentalizes his lack of sense, so he’s only not sensible in one area (for instance, a creationist who is perfectly fine at calculating restaurant tips), this isn’t going to work.
Ahhhh, maybe I see what you’re complaining about
Are you primarily thinking of this as applying to creationists etc?
The part of the reason I put the caveat ‘people about as reasonable as you’ in the first place was to exclude that category of people from what I was talking about.
That is not the central category of people I’m suggesting this for. Also, I’m not clear on why you would think it was.
I’m not using creationists as an example because it’s central; I’m using it as an example because it’s unambiguous. It’s really hard to sidetrack the argument by suggesting that maybe the creationists are right after all, or that I’m being arrogant by thinking the creationists are mistaken, etc. so creationists work well as an example.
(And an idea that works for central examples but fails for edge cases is an idea that fails.)
But if you add that exception, it swallows the rule. Most people think their opponents are more unreasonable than themselves.
(Sorry for the long delay)
Ah, I see why you’re arguing now.
Ironically, this is not a universal criteria for the success of ideas. Sometimes its a very useful criteria (think mathematical proofs). Other times, its not a very useful idea (think ‘choosing friends’ or ‘mathematical intuitions’).
For example the idea of ‘cat’ fails for edge cases. Is this a cat? Sort of. Sort of not. But ‘cat’ is still a useful concept.
Concepts are clusters in thing space, and the concept that I am pointing at is also a cluster.
This comment on that post is especially relevant.
Maybe I’m still misunderstanding.