From your first comment to my post on you were really agressive. Arguments are fine, but why always the personal attacks?
I tell you what might be going on here: You saw the post, couldn’t make sense of it after a quick glance and decided it was junk and an easy way to gain reputation and boost your ego by bashing.
And you are not alone. There are lots of haters, and nobody who just said, Ok, I don’t believe it, but let’s discuss it, and stop hitting the guy over the head.
The theory is highly counterintuitive, I said as much, but it is worth at least a few minutes of discussion, and i discussed it with quite a few eminent philosophers already. None was convinced (which is hardly surprising), but they found the discussion interesting and the theory consistent.
So something has gone wrong here. Maybe all this talk of “winning” and “bayesian conspiracy” and whatever really does not do a favor to the principle goal of the site of being as unbiased as possible.
First, none of us are being as rude to you as you are to us in this comment alone. If you can’t stand the abuse you’re getting here, then quit commenting on this post.
Second, we’ve given this well more than a few minutes’ discussion, and you’ve given us no reason to believe that we misunderstand your theory—you just object to our categorical dismissal of it. I am perfectly willing to believe that the philosophers you discussed this with gave you credit for making an interesting argument—philosophers are generous like that—and for all its faults, your theory is consistent. But around here, interesting is a matter of writing style, and consistent is a sub-minimal requirement: we demand useful. None of us are rationalists just for the lulz—if a theory doesn’t help us get what we actually want, it really is of no use to us. And by that standard, any skeptical hypothesis is a waste of time, including your proposed Humeiform worldview, when other hypotheses actually work.
First, none of us are being as rude to you as you are to us in this comment alone. If you can’t stand the abuse you’re getting here, then quit commenting on this post.
Oh, I can take the abuse, I’m just wondering.
Second, we’ve given this well more than a few minutes’ discussion, and you’ve given us no reason to believe that we misunderstand your theory
At least at first, I’ve been given just accusations and incredulous stares.
if a theory doesn’t help us get what we actually want, it really is of no use to us
If you want the truth, you have to consider being wrong even about your darlings, say, prediction.
Do you actually believe this theory that you have proposed? Because we aren’t arguing that it’s logically impossible, we’re explaining why we don’t believe it.
Your theory says you can’t cause our beliefs to change and you shouldn’t be surprised about it. It also implies that you defend it by accident, not because it’s true.
The good news is that you have an obvious upgrade right ahead. Not all of us are so lucky.
No such assumption required. For example, if you have 10% credence in your theory, the same 10% says you’re defending it by accident. Viewed another way, we have no reason to listen to you if your theory is false and no reason to listen if it’s true either. Please apply this logic to your beliefs and update.
Seems to me you’re conflating different concepts:
“being the reason for” and “being the cause of”:
compare what an enemy of determinism could say: “we have no reason to listen to you if your theory is false and no reason to listen if it’s true either”. Now what?
Let’s drop abstract truth-seeking for a moment and talk about instrumental values instead.
Believing in causality is useful in a causal world and neutral in an acausal one. Disbelieving in causality is harmful in a causal world and likewise neutral in an acausal one. So, if you assign nonzero credence to the existence of causality (as you implied in a comment above: “why does everybody assume I’m a die-hard believer?”), you’d do better by increasing this credence to 100%, because doing so has positive utility in the causal world (to which you have assigned nonzero credence) and doesn’t matter in the acausal one.
I would say, “increasing this credence toward 100%”—without mathematical proof that the familiar sort of causation is the only such scheme that is feasible, absolute certainty is (slightly) risky. (Even with such proof, it is risky—proofs aren’t perfect guarantees.)
I can’t parse your comment. Are you saying that, conditioned on your theory being true, our beliefs “should” somehow causally update in response to your arguments? That’s obviously false.
From your first comment to my post on you were really agressive. Arguments are fine, but why always the personal attacks? I tell you what might be going on here: You saw the post, couldn’t make sense of it after a quick glance and decided it was junk and an easy way to gain reputation and boost your ego by bashing. And you are not alone. There are lots of haters, and nobody who just said, Ok, I don’t believe it, but let’s discuss it, and stop hitting the guy over the head.
The theory is highly counterintuitive, I said as much, but it is worth at least a few minutes of discussion, and i discussed it with quite a few eminent philosophers already. None was convinced (which is hardly surprising), but they found the discussion interesting and the theory consistent. So something has gone wrong here. Maybe all this talk of “winning” and “bayesian conspiracy” and whatever really does not do a favor to the principle goal of the site of being as unbiased as possible.
Spuckblase, two things.
First, none of us are being as rude to you as you are to us in this comment alone. If you can’t stand the abuse you’re getting here, then quit commenting on this post.
Second, we’ve given this well more than a few minutes’ discussion, and you’ve given us no reason to believe that we misunderstand your theory—you just object to our categorical dismissal of it. I am perfectly willing to believe that the philosophers you discussed this with gave you credit for making an interesting argument—philosophers are generous like that—and for all its faults, your theory is consistent. But around here, interesting is a matter of writing style, and consistent is a sub-minimal requirement: we demand useful. None of us are rationalists just for the lulz—if a theory doesn’t help us get what we actually want, it really is of no use to us. And by that standard, any skeptical hypothesis is a waste of time, including your proposed Humeiform worldview, when other hypotheses actually work.
Edit circa 2014: the Slacktivist blog moved (mostly) to a new website—this is the new link to the “sub-minimal requirement” post.
Oh, I can take the abuse, I’m just wondering.
At least at first, I’ve been given just accusations and incredulous stares.
If you want the truth, you have to consider being wrong even about your darlings, say, prediction.
Do you actually believe this theory that you have proposed? Because we aren’t arguing that it’s logically impossible, we’re explaining why we don’t believe it.
Your theory says you can’t cause our beliefs to change and you shouldn’t be surprised about it. It also implies that you defend it by accident, not because it’s true.
The good news is that you have an obvious upgrade right ahead. Not all of us are so lucky.
Why does everybody assume I’m a die-hard believer in this theory?
No such assumption required. For example, if you have 10% credence in your theory, the same 10% says you’re defending it by accident. Viewed another way, we have no reason to listen to you if your theory is false and no reason to listen if it’s true either. Please apply this logic to your beliefs and update.
Seems to me you’re conflating different concepts: “being the reason for” and “being the cause of”:
compare what an enemy of determinism could say: “we have no reason to listen to you if your theory is false and no reason to listen if it’s true either”. Now what?
Let’s drop abstract truth-seeking for a moment and talk about instrumental values instead.
Believing in causality is useful in a causal world and neutral in an acausal one. Disbelieving in causality is harmful in a causal world and likewise neutral in an acausal one. So, if you assign nonzero credence to the existence of causality (as you implied in a comment above: “why does everybody assume I’m a die-hard believer?”), you’d do better by increasing this credence to 100%, because doing so has positive utility in the causal world (to which you have assigned nonzero credence) and doesn’t matter in the acausal one.
Well, if you stipulate that “abstract truth-seeking” has nothing whatsoever to do with my getting along in the world, then you’re right I guess.
I would say, “increasing this credence toward 100%”—without mathematical proof that the familiar sort of causation is the only such scheme that is feasible, absolute certainty is (slightly) risky. (Even with such proof, it is risky—proofs aren’t perfect guarantees.)
I can’t parse your comment. Are you saying that, conditioned on your theory being true, our beliefs “should” somehow causally update in response to your arguments? That’s obviously false.
We don’t need to assume that. If you have 10% credence for your theory, my reasoning applies for that 10%.