It would help if you could sum up the merely mechanical things that are ‘just what consciousness is’ in Dennett’s (or your!) sense.
I don’t think it would, actually.
The merely mechanical things that are ‘just what consciousness is’ in Dennett’s sense are the “soft problem of consciousness” in Chalmers’ sense; I don’t expect any amount of summarizing or detailing the former to help anyone feel like the “hard problem of consciousness” has been addressed, any more than I expect any amount of explanation of materials science or topology to help A feel like the “hard problem of containment” has been addressed.
But, since you asked: I’m not denying that we have subjective experiences (nor do I believe Dennett is), and I am saying that those experiences are a consequence of our neurobiology (as I believe Dennett does). If you’re looking for more details of things like how certain patterns of photons trigger increased activation levels of certain neural structures, there are better people to ask than me, but I don’t think that’s what you’re looking for.
As for whether they are an inherent quality or an incidental byproduct of that neurobiology, I’m not sure I even understand the question. Is being a container an inherent quality of being composed of certain materials and having certain shape, or an incidental byproduct? How would I tell?
I may not remember Chalmer’s soft problem well enough either for reference of that to help!
If experiences are a consequence of our neurobiology, fine. Presumably a consequence that itself has consequences: experiences can be used in causal explanations? But it seems to me that we could explain how a bat uses echolocation without knowing what echolocation looks like (sounds like? feels like?) to a bat. And that we could distinguish how well people distinguish wavelengths of light etc. without knowing what the colour looks like to them.
It seems subjective experience is just being ignored: we could identify that an AI could carry out all sorts of tasks that we associate with consciousness, but I have no idea when we’d say ‘it now has conscious experiences’. Or whether we’d talk about degrees of conscious experience, or whatever. This is obviously ethically quite important, if not that directly pertinent to me, and it bothers me that I can’t respond to it.
With a container, you describe various qualities and that leaves the question ‘can it contain things’: do things stay in it when put there. You’re adding a sort of purpose-based functional classification to a physical object. When we ask ‘is something conscious’, we’re not asking about a function that it can perform. On a similar note, I don’t think we’re trying to reify something (as with the case where we have a sense of objects having ongoing identity, which we then treat as a fundamental thing and end up asking if a ship is the same after you replace every component of it one by one). We’re not chasing some over-abstracted ideal of consciousness, we’re trying to explain an experienced reality.
So to answer A, I’d say ‘there is no fundamental property of ‘containment’. It’s just a word we use to describe one thing surrounded by another in circumstances X and Y. You’re over-idealising a useful functional concept’. The same is not true of consciousness, because it’s not (just) a function.
It might help if you could identify what, in light of a Dennett-type approach, we can identify as conscious or not. I.e. plants, animals, simple computers, top-level computers, theoretical super-computers of various kinds, theoretical complex networks divided across large areas so that each signal from one part to another takes weeks…
I’m splitting up my response to this into several pieces because it got long. Some other stuff:
what, in light of a Dennett-type approach, we can identify as conscious or not.
The process isn’t anything special, but OK, since you ask.
Let’s assert for simplicity that “I” has a relatively straightforward and consistent referent, just to get us off the ground. Given that, I conclude that am at least sometimes capable of subjective experience, because I’ve observed myself subjectively experiencing.
I further observe that my subjective experiences reliably and differentially predict certain behaviors. I do certain things when I experience pain, for example, and different things when I experience pleasure. When I observe other entities (E2) performing those behaviors, that’s evidence that they, too, experience pain and pleasure. Similar reasoning applies to other kinds of subjective experience.
I look for commonalities among E2 and I generalize across those commonalities. I notice certain biological structures are common to E2 and that when I manipulate those structures, I reliably and differentially get changes in the above-referenced behavior. Later, I observe additional entities (E3) that have similar structures; that’s evidence that E3 also demonstrates subjective experience, even though E3 doesn’t behave the way I do.
Later, I build an artificial structure (E4) and I observe that there are certain properties (P1) of E2 which, when I reproduce them in E4 without reproducing other properties (P2), reproduce the behavior of E2. I conclude that P1 is an important part of that behavior, and P2 is not.
I continue this process of observation and inference and continue to draw conclusions based on it. And at some point someone asks “is X conscious?” for various Xes:
I.e. plants, animals, simple computers, top-level computers, theoretical super-computers of various kinds, theoretical complex networks divided across large areas so that each signal from one part to another takes weeks...
If I interpret “conscious” as meaning having subjective experience, then for each X I observe it carefully and look for the kinds of attributes I’ve attributed to subjective experience… behaviors, anatomical structures, formal structures, etc… and compare it to my accumulated knowledge to make a decision.
Isn’t that how you answer such questions as well?
If not, then I’ll ask you the same question: what, in light of whatever non-Dennett-type approach you prefer, can we identify as conscious or not?
Ok, well given that responses to pain/pleasure can equally be explained by more direct evolutionary reasons, I’m not sure that the inference from action to experience is very useful. Why would you ever connect these things with expereince rather than other, more directly measurable things?
But the point is definitely not that I have a magic bullet or easy solution: it’s that I think there’s a real and urgent question—are they conscious—which I don’t see how information about responses etc. can answer. Compare to the cases of containment, or heat, or life—all the urgent questions are already resolved before those issues are even raised.
As I say, the best way I know of to answer “is it conscious?” about X is to compare X to other systems about which I have confidence-levels about its consciousness and look for commonalities and distinctions.
If there are alternative approaches that you think give us more reliable answers, I’d love to hear about them.
I have no reliable answers! And I have low meta-confidence levels (in that it seems clear to me that people and most other creatures are conscious but I have no confidence in why I think this)
If the Dennett position still sees this as a complete bafflement but thinks it will be resolved with the so-called ‘soft’ problem, I have less of an issue than I thought I did. Though I’d still regard the view that the issue will become clear one of hope rather than evidence.
I’m splitting up my response to this into several pieces because it got long. Some other stuff:
Presumably a consequence that itself has consequences: experiences can be used in causal explanations?
I expect so, sure. For example, I report having experiences; one explanation of that (though hardly the only possible one) starts with my actually having experiences and progresses forward in a causal fashion.
But it seems to me that we could explain how a bat uses echolocation without knowing what echolocation looks like (sounds like? feels like?) to a bat.
Sure, there are many causal explanations of many phenomena, including but not limited to how bats use echolocation, that don’t posit subjective experience as part of their causal chain. For example, humas do all kinds of things without the subjective experience of doing them.
And that we could distinguish how well people distinguish wavelengths of light etc. without knowing what the colour looks like to them.
Certainly.
It seems subjective experience is just being ignored
In the examples you give, yes, it is being ignored. So? Lots of things are being ignored in those examples… mass, electrical conductivity, street address, level of fluency in Russian, etc. If these things aren’t necessary to explain the examples, there’s nothing wrong with ignoring these things.
On the other hand, if we look at an example for which experience ought to be part of the causal chain (for example, as I note above, reporting having those experiences), subjective experience is not ignored. X happens, as a consequence of X a subjective experience Y arises, as a consequence of Y a report Z arises, and so forth. (Of course, for some reports we do have explanations that don’t presume Y… e.g., confabulation, automatic writing, etc. But that needn’t be true for all reports. Indeed, it would be surprising if it were.)
“But we don’t know what Xes give rise to the Y of subjective experience, so we don’t fully understand subjective experience!” Well, yes, that’s true. We don’t fully understand fluency in Russian, either. But we don’t go around as a consequence positing some mysterious essence of Russian fluency that resists neurobiological explanation… though two centuries ago, we might have done so. Nor should we. Neither should we posit some mysterious essence of subjective experience.
“But subjective experience is different! I can imagine what a mechanical explanation of Russian fluency would be like, but I can’t imagine what a mechanical explanation of subjective experience would be like.” Sure, I understand that. Two centuries ago, the notion of a mechanical explanation of Russian fluency would raise similar incredulity… how could a machine speak Russian? I’m not sure how I could go about answering such incredulity convincingly, but I don’t thereby conclude that machines can’t speak Russian. My incredulity may be resistant to my reason, but it doesn’t therefore compel or override my reason.
I have a lot of sympathy for this. The most plausible position of reductive materialism is simply that at some future scientific point this will become clear. But this is inevitably a statement of faith, rather than acknoweldging current acheivement. It’s very hard to compare current apparent mysteries to solved mysteries—I do get that. Having said that, I can’t even see what the steps on the way to explaining consciousness would be, and claiming there is not such thing seems not to be an option (unlike ‘life’, ‘free will’ etc.) whereas most other cases you rely on saying that you can’t see how the full extent could be achieved: a machine might speak crap Russian in some circumstances etc.
Also, if a machine can speak Russian, you can check that. I don’t know how we’d check a machine was conscious.
BTW, when I said ‘it seems subjective experience is just being ignored’, I meant ignored in your and Dennett’s arguments, not in specific explanations. I have nothing against analysing things in ways that ignore consciousness, if they work.
I don’t know what the mechanical explanation would look like, either. But I’m sufficiently aware of how ignorant my counterparts two centuries ago would have been of what a mechanical explanation for speaking Russian would look like that I don’t place too much significance on my ignorance.
I agree that testing whether a system is conscious or not is a tricky problem. (This doesn’t just apply to artificial systems.)
Indeed: though artificial systems are more intuitively difficult as we don’t have as clear an intuitive expectation.
You can take an outside view and say ‘this will dissolve like the other mysteries’. I just genuinely find this implausible, if only because you can take steps towards the other mysteries (speaking bad Russian occasionally) and because you have a clear empirical standard (russians). Whereas for consciousness I don’t have any standard for identifying another’s consciousness: I do it only by analogy with myself and by the implausibility of me having an apparently causal element that others who act similarly to me lack.
I agree that the “consciousness-detector” problem is a hard problem. I just can’t think of a better answer than the generalizing-from-commonalities strategy I discussed previously, so that’s the approach I go with. It seems capable of making progress for now.
And I understand that you find it implausible. That said, I suspect that if we solve the “soft” problem of consciousness well enough that a typical human is inclined to treat an artificial system as though it were conscious, it will start to seem more plausible.
Perhaps it will be plausible and incorrect, and we will happily go along treating computers as conscious when they are no such thing. Perhaps we’re currently going along treating dogs and monkeys and 90% of humans as conscious when they are no such thing.
Perhaps not.
Either way, plausibility (or the absence of it) doesn’t really tell us much.
Yes. This is what worries me: I can see more advances making everyone sure that computers are conscious, but my suspicion is that this will not be logical. Take the same processor and I suspect the chances of it being seen as conscious will rise sharply if it’s put in a moving machine, rise sharply again for a humanoid, again for face/voice and again for physically indistinguishable.
The problem with generalising from commonalities is that I have precisely one direct example of consciousness. Although having said that, I don’t find epiphenomenal accounts convincing, so it’s reasonable for me to think that as my statements about qualia seem to follow causally from experiencing said qualia, that other people don’t have a totally separate framework for their statements about qualia. I wouldn’t be that confident though, and it gets harder with artificial consciousness.
Sure. By the same token, if you take me, remove my ability to communicate, and encase me in an opaque cylinder, nobody will recognize me as a being with subjective experience. Or, for that matter, as a being with the ability to construct English sentences.
We are bounded intellects reasoning under uncertainty in a noisy environment. We will get stuff wrong. Sometimes it will be important stuff.
it’s reasonable for me to think that as my statements about qualia seem to follow causally from experiencing said qualia, that other people don’t have a totally separate framework for their statements about qualia.
I agree. And, as I said initially, I apply the same reasoning not only to the statements I make in English, but to all manner of behaviors that “seem to rise from my qualia,” as you put it… all of it is evidence in favor of other organisms also having subjective experience, even organisms that don’t speak English.
I wouldn’t be that confident though,
How confident are you that I possess subjective experience? Would that confidence rise significantly if we met in person and you verified that I have a typical human body?
Consciousness does seem different in that we can have a better and better understanding of all the various functional elements but that we’re
1) left with a sort of argument from analogy for others having qualia
2) even if we can resolve(1), I can’t see how we can start to know whether my green is your red etc. etc.
I can’t think of many comparable cases: certainly I don’t think containership is comparable. You and I could end up looking at the AI in the moment before it destroys/idealises/both the world and say ‘gosh, I wonder if it’s conscious’. This is nothing like the casuistic ‘but what about this container gives it its containerness’. I think we’re on the same point here, though?
I’m intuitively very confident you’re conscious: and yes, seeing you were human would help (in that one of the easiest ways I can imagine you weren’t conscious is that you’re actually a computer designed to post about things on less wrong. This would also explain why you like Dennett—I’ve always suspected he’s a qualia-less robot too! ;-)
Yes, I agree that we’re much more confused about subjective experience than we are about containership.
We’re also more confused about subjective experience than we are about natural language, about solving math problems, about several other aspects of cognition. We’re not _un_confused about those things, but we’re less confused than we used to be. I expect us to grow still less confused over time.
I disagree about the lack of comparable cases. I agree about containers; that’s just an intuition pump. But the issues that concern you here arise for any theoretical construct for which we have only indirect evidence. The history of science is full of such things. Electrons. Black holes. Many worlds. Fibromyalgia. Phlogiston. Etc.
What makes subjective experience different is not that we lack the ability to perceive it directly; that’s pretty common. What makes it different is that we can perceive it directly in one case, as opposed to the other stuff where we perceive it directly in zero cases.
Of course, it’s also different from many of them in that it matters to our moral reasoning in many cases. I can’t think of a moral decision that depends on whether phlogiston exists, but I can easily think of a moral decision that depends on whether cows have subjective experiences. OTOH, it still isn’t unique; some people make moral decisions that depend on the actuality of theoretical constructs like many worlds and PTSD.
Fair enough. As an intuition pump, for me at least, it’s unhelpful: it gave the impression that you thought that consciousness was merely a label being mistaken for a thing (like ‘life’ as something beyond its parts).
Only having indirect evidence isn’t the problem. For a black hole, I care about the observable functional parts. I wouldn’t be being sucked towards it and being crushed while going ‘but is it really a black hole?’ A black hole is like a container here: what matter are the functional bits that make it up. For consciousness, I care if a robot can reason and can display conscious-type behaviour, but I also care if it can experience and feel.
Many worlds could be comparable if there is evidence that means that there are ‘many worlds’ but people are vague about if these worlds actually exist. And you’re right, this is also a potentially morally relevant point.
Insofar as people infer from the fact of subjective experience that there is some essence of subjective experience that is, as you say, “beyond its parts” (and their patterns of interaction), I do in fact think they are mistaking a label for a thing.
I dunno about essences. The point is that you can observe lots of interactions of neurons and behaviours and be left with an argument from analogy to say “they must be conscious because I am and they are really similar, and the idea that my consciousness is divorced from what I do is just wacky”.
You can observe all the externally observable, measurable things that a black hole or container can do, and then if someone argues about essences you wonder if they’re actually referring to anything: it’s a purely semantic debate. But you can observe all the things a fish, or tree, or advanced computer can do, predict it for all useful purposes, and still not know if it’s conscious. This is bothersome. But it’s not to do with essences, necessarily.
Insofar as people don’t infer something else beyond the parts of a (for example) my body and their pattern of interactions that account for (for example) my subjective experience, I don’t think they are mistaking a label for a thing.
Well, until we know how to identify if something/someone is conscious, it’s all a bit of a mystery: I couldn’t rule out consciousness being some additional thing. I have an inclination to do so because it seems unparsimonious, but that’s it.
I’m splitting up my response to this into several pieces because it got long.
The key bit, IMHO:
So to answer A, I’d say ‘there is no fundamental property of ‘containment’. It’s just a word we use to describe one thing surrounded by another in circumstances X and Y.
And I would agree with you.
and that leaves the question ‘can it contain things’: [..] The same is not true of consciousness, because it’s not (just) a function.
“No,” replies A, “you miss the point completely. I don’t ask whether a container can contain things; clearly it can, I observe it doing so. I ask how it contains things. What is the explanation for its demonstrated ability to contain things? Containership is not just a function,” A insists, “though I understand you want to treat it as one. No, containership is a fundamental essence. You can’t simply ignore the hard question of “is X a container?” in favor of thinking about simpler, merely functional questions like “can X contain Y?”. And, while we’re at it,” A coninues, “what makes you think that an artificial container, such as we build all the time, is actually containing anything rather than merely emulating containership? Sure, perhaps we can’t tell the difference, but that doesn’t mean there isn’t a difference.”
I take it you don’t find A’s argument convincing, and neither do I, but it’s not clear to me what either of us could say to A that A would find at all compelling.
Maybe we couldn’t, but A is simply asserting that containership is a concept beyond its parts, whereas I’m appealing directly to experience: the relevance of this is that whether something has experience matters. Ultimately for any case, if others just express bewilderment in your concepts and apparently don’t get what you’re talking about, you can’t prove it’s an issue. But at any rate, most people seem to have subjective experience.
Being conscious isn’t a label I apply to certain conscious-type systems that I deem ‘valuable’ or ‘true’ in some way. Rather, I want to know what systems should be associated with the clearly relevant and important category of ‘conscious’
My thoughts about how I go about associating systems with the expectation of subjective experience are elsewhere and I have nothing new to add to it here.
As regards you and A… I realize that you are appealing directly to experience, whereas A is merely appealing to containment, and I accept that it’s obvious to you that experience is importantly different from containment in a way that makes your position importantly non-analogous to A’s.
I have no response to A that I expect A to find compelling… they simply don’t believe that containership is fully explained by the permeability and topology of containers. And, you know, maybe they’re right… maybe some day someone will come up with a superior explanation of containerhood that depends on some previously unsuspected property of containers and we’ll all be amazed at the realization that containers aren’t what we thought they were. I don’t find it likely, though.
I also have no response to you that I expect you to find compelling. And maybe someday someone will come up with a superior explanation of consciousness that depends on some previously unsuspected property of conscious systems, and I’ll be amazed at the realization that such systems aren’t what I thought they were, and that you were right all along.
Are you saying you don’t experience qualia and find them a bit surprising (in a way you don’t for containerness)? I find it really hard to not see arguments of this kind as a little disingenous: is the issue genuinely not difficult for some people, or is this a rhetorical stance intended to provoke better arguments, or awareness of the weakness of current arguments?
I have subjective experiences. If that’s the same thing as experiencing qualia, then I experience qualia.
I’m not quite sure what you mean by “surprising” here… no, it does not surprise me that I have subjective experiences, I’ve become rather accustomed to it over the years. I frequently find the idea that my subjective experiences are a function of the formal processes my neurobiology implements a challenging idea… is that what you’re asking?
Then again, I frequently find the idea that my memories of my dead father are a function of the formal processes my neurobiology implements a challenging idea as well. What, on your view, am I entitled to infer from that?
Yes, I meant surprising in light of other discoveries/beliefs.
On memory: is it the conscious experience that’s challenging (in which case it’s just a sub-set of the same issue) or do you find the functional aspects of memory challenging? Even though I know almost nothing about how memory works, I can see plausible models and how it could work, unlike consciousness.
Isn’t our objection to A’s position that it doesn’t pay rent in anticipated experience? If one thinks there is a “hard problem of consciousness” such that different answers would cause one to behave differently, then one must take up the burden of identifying what the difference would look like, even if we can’t create a measuring device to find it just now.
Sure, perhaps we can’t tell the difference, but that doesn’t mean there isn’t a difference.
If A means that we cannot determine the difference in principle, then there’s nothing we should do differently. If A means that a measuring device does not currently exist, he needs to identify the range of possible outputs of the device.
Isn’t our objection to A’s position that it doesn’t pay rent in anticipated experience?
This may be a situation where that’s a messy question. After all, qualia are experience. I keep expecting experiences, and I keep having experiences. Do experiences have to be publicly verifiable?
If two theories both lead me to anticipate the same experience, the fact that I have that experience isn’t grounds for choosing among them.
So, sure, the fact that I keep having experiences is grounds for preferring a theory of subjective-experience-explaining-but-otherwise-mysterious qualia over a theory that predicts no subjective experience at all, but not necessarily grounds for preferring it to a theory of subjective-experience-explaining-neural-activity.
different answers to the HP would undoubtedly change our behaviour, because they would indicate different classes of entity have feelings impacting morality. Indeed, it is pretty hard to think of anything more impactive.
The measuring device for conscious experience is consciousness, which is the whole problem.
Sure. But in this sense, false believed answers to the HP are no different from true believed answers.… that is, they would both potentially change our behavior the way you describe.
Sure. But in this sense, false believed answers to the HP are no different from true believed answers.… that is, they would both potentially change our behavior the way you describe.
That is the case for most any belief you hold (unless you mean “in the exact same way”, not as “change behavior”). You may believe there’s a burglar in your house, and that will impact your actions, whether it be false or true. Say you believe that it’s more likely there is a burglar, you are correct in acting upon that belief even if it turns out to incorrect. It’s not AIXI’s fault if it believes in the wrong thing for the right reasons.
In that sense, you can choose an answer for example based on complexity considerations. In the burglar example, the answer you choose (based on data such as crime rate, cat population etc.) can potentially be further experimentally “verified” (the probability increased) as true or false, but even before such verification, your belief can still be strong enough to act upon.
After all, you do act upon your belief that “I am not living in a simulation which will eventually judge and reward me only for the amount of cheerios I’ve eaten”. It also doesn’t lead to different expected experiences at the present time, yet you also choose to act thus as if it were true. Prior based on complexity considerations alone, yet strong enough to act upon. Same when thinking about whether the sun has qualia (“hot hot hot hot hot”).
(Bit of a hybrid fusion answer also meant to refer to our neighboring discussion branch.)
In the absence of any suggestion of what might be evidence one way or the other, in the absence of any notion of what I would differentially expect to observe in one condition over the other, I don’t see any value to asking the question. If it makes you feel better if I don’t deny their existence, well, OK, I don’t deny their existence, but I really can’t see why anyone should care one way or the other.
Well, in the case of “do landslides have qualia”, Occam’s Razor could be used to assign probabilities just the same as we assign probabilities in the “cheerio simulation” example. So we’ve got methodology, we’ve got impact, enough to adopt a stance on the “psychic unity of the cosmos”, no?
My best guess is that you’re suggesting that, with respect to systems that do not manifest subjective experience in any way we recognize or understand, Occam’s Razor provides grounds to be more confident that they have subjective experience than that they don’t. If that’s what you mean, I don’t see why that should be. If that’s not what you mean, can you rephrase the question?
I think it’s conceivable if not likely that Occam’s Razor would favor or disfavor qualia as a property of more systems than just those that seem to show or communicate them in terms we’re used to. I’m not sure which, but it is a question worth pondering, with an impact on how we view the world, and accessible through established methodology, to a degree.
I’m not advocating assigning a high probability to “landslides have raw experience”, I’m advocating that it’s an important question, the probability of which can be argued. I’m an advocate of the question, not the answer, so to speak. And as such opposed to “I really can’t see why anyone should care one way or the other”.
So, I stand by my assertion that in the absence of evidence one way or the other, I really can’t see why anyone should care.
But I agree that to the extent that Occam’s Razor type reasoning provides evidence, that’s a reason to care.
And if it provided strong evidence one way or another (which I don’t think it does, and I’m not sure you do either) that would provide a strong reason to care.
I stand by my assertion that in the absence of evidence one way or the other,
I have evidence in the form of by personal experience of qualia. Granted, I have no way of showing you that evidence, but that doesn’t mean I don’t have it.
Agreed that the ability to share evidence with others is not a necessary condition of having evidence. And to the extent that I consider you a reliable evaluator of (and reporter of) evidence, your report is evidence, and to that extent I have a reason to care.
Moral implications of a proposition in the absence of evidence one way or another for that proposition are insufficient to justify caring. If I actually care about the experiences of minds capable of experiences, I do best to look for evidence for the presence or absence of such experiences. Failing such evidence, I do best to concentrate my attention elsewhere.
It’s possible to have both a strong reason to care, and weak evidence, ie due the moral hazard dependent on some doubtful proposition. People often adopt precautionary principles in such scenarios.
It’s possible to have both a strong reason to care, and weak evidence, ie due the moral hazard dependent on some doubtful proposition.
I don’t think that’s the situation here though. That sounds like a description of this situation: (imagine) we have weak evidence that 1) snakes are sapient, and we grant that 2) sapience is morally significant. Therefore (perhaps) we should avoid wonton harm to snakes.
Part of why this argument might make sense is that (1) and (2) are independent. Our confidence in (2) is not contingent on the small probability that (1) is true: whether or not snakes are sapient, we’re all agreed (lets say) that sapience is morally significant.
On the other hand, the situation with qualia is one where we have weak evidence (suppose) that A) qualia are real, and we grant that B) qualia are morally significant.
The difference here is that (B) is false if (A) is false. So the fact that we have weak evidence for (A) means that we can have no stronger (and likely, we must have yet weaker) evidence for (B).
Does the situation change significantly if “the situation with qualia” is instead framed as A) snakes have qualia and B) qualia are morally significant?
Yes, if the implication of (A) is that we’re agreed on the reality of qualia but are now wondering whether or not snakes have them. No, if (A) is just a specific case of the general question ‘are qualia real?’. My point was probably put in a confusing way: all I mean to say was that Juno seemed to be arguing as if it were possible to be very confident about the moral significance of qualia while being only marginally confident about their reality.
What I think of the case for qualia is beside the point, I was just commenting on your ‘moral hazard’ argument. There you said that even if we assume that we have only weak evidence for the reality of qualia, we should take the possibility seriously, since we can be confident that qualia are morally significant. I was just pointing out that this argument is made problematic by the fact that our confidence in the moral significance of qualia can be no stronger than our confidence in their reality, and therefore by assumption must be weak.
I am guessing that Juno_Watt means that strong evidence for our own perception of qualia makes them real enough to seriously consider their moral significance, whether or not they are “objectively real”.
Yes, they often do. On your view, is there a threshold of doubtfulness of a proposition below which it is justifiable to not devote resources to avoiding the potential moral hazard of that proposition being true, regardless of the magnitude of that moral hazard?
i don’t think ir’s likely my house will catch fire, but i take out fire insurance. OTOH, if i don’t set a lower bound I will be susceptible to pascal’s muggings.
He may have meant something like “Qualiaphobia implies we would have no expereinces at all”. However, that all depends on what you mean by experience. I don’t think the Expected Experience criterion is useful here (or anywhere else)
I don’t think it would, actually.
The merely mechanical things that are ‘just what consciousness is’ in Dennett’s sense are the “soft problem of consciousness” in Chalmers’ sense; I don’t expect any amount of summarizing or detailing the former to help anyone feel like the “hard problem of consciousness” has been addressed, any more than I expect any amount of explanation of materials science or topology to help A feel like the “hard problem of containment” has been addressed.
But, since you asked: I’m not denying that we have subjective experiences (nor do I believe Dennett is), and I am saying that those experiences are a consequence of our neurobiology (as I believe Dennett does). If you’re looking for more details of things like how certain patterns of photons trigger increased activation levels of certain neural structures, there are better people to ask than me, but I don’t think that’s what you’re looking for.
As for whether they are an inherent quality or an incidental byproduct of that neurobiology, I’m not sure I even understand the question. Is being a container an inherent quality of being composed of certain materials and having certain shape, or an incidental byproduct? How would I tell?
And: how would you reply to A?
That’s such a broad statement, it could cover some forms of dualism.
Agreed.
I may not remember Chalmer’s soft problem well enough either for reference of that to help!
If experiences are a consequence of our neurobiology, fine. Presumably a consequence that itself has consequences: experiences can be used in causal explanations? But it seems to me that we could explain how a bat uses echolocation without knowing what echolocation looks like (sounds like? feels like?) to a bat. And that we could distinguish how well people distinguish wavelengths of light etc. without knowing what the colour looks like to them.
It seems subjective experience is just being ignored: we could identify that an AI could carry out all sorts of tasks that we associate with consciousness, but I have no idea when we’d say ‘it now has conscious experiences’. Or whether we’d talk about degrees of conscious experience, or whatever. This is obviously ethically quite important, if not that directly pertinent to me, and it bothers me that I can’t respond to it.
With a container, you describe various qualities and that leaves the question ‘can it contain things’: do things stay in it when put there. You’re adding a sort of purpose-based functional classification to a physical object. When we ask ‘is something conscious’, we’re not asking about a function that it can perform. On a similar note, I don’t think we’re trying to reify something (as with the case where we have a sense of objects having ongoing identity, which we then treat as a fundamental thing and end up asking if a ship is the same after you replace every component of it one by one). We’re not chasing some over-abstracted ideal of consciousness, we’re trying to explain an experienced reality.
So to answer A, I’d say ‘there is no fundamental property of ‘containment’. It’s just a word we use to describe one thing surrounded by another in circumstances X and Y. You’re over-idealising a useful functional concept’. The same is not true of consciousness, because it’s not (just) a function.
It might help if you could identify what, in light of a Dennett-type approach, we can identify as conscious or not. I.e. plants, animals, simple computers, top-level computers, theoretical super-computers of various kinds, theoretical complex networks divided across large areas so that each signal from one part to another takes weeks…
I’m splitting up my response to this into several pieces because it got long. Some other stuff:
The process isn’t anything special, but OK, since you ask.
Let’s assert for simplicity that “I” has a relatively straightforward and consistent referent, just to get us off the ground. Given that, I conclude that am at least sometimes capable of subjective experience, because I’ve observed myself subjectively experiencing.
I further observe that my subjective experiences reliably and differentially predict certain behaviors. I do certain things when I experience pain, for example, and different things when I experience pleasure. When I observe other entities (E2) performing those behaviors, that’s evidence that they, too, experience pain and pleasure. Similar reasoning applies to other kinds of subjective experience.
I look for commonalities among E2 and I generalize across those commonalities. I notice certain biological structures are common to E2 and that when I manipulate those structures, I reliably and differentially get changes in the above-referenced behavior. Later, I observe additional entities (E3) that have similar structures; that’s evidence that E3 also demonstrates subjective experience, even though E3 doesn’t behave the way I do.
Later, I build an artificial structure (E4) and I observe that there are certain properties (P1) of E2 which, when I reproduce them in E4 without reproducing other properties (P2), reproduce the behavior of E2. I conclude that P1 is an important part of that behavior, and P2 is not.
I continue this process of observation and inference and continue to draw conclusions based on it. And at some point someone asks “is X conscious?” for various Xes:
If I interpret “conscious” as meaning having subjective experience, then for each X I observe it carefully and look for the kinds of attributes I’ve attributed to subjective experience… behaviors, anatomical structures, formal structures, etc… and compare it to my accumulated knowledge to make a decision.
Isn’t that how you answer such questions as well?
If not, then I’ll ask you the same question: what, in light of whatever non-Dennett-type approach you prefer, can we identify as conscious or not?
Ok, well given that responses to pain/pleasure can equally be explained by more direct evolutionary reasons, I’m not sure that the inference from action to experience is very useful. Why would you ever connect these things with expereince rather than other, more directly measurable things?
But the point is definitely not that I have a magic bullet or easy solution: it’s that I think there’s a real and urgent question—are they conscious—which I don’t see how information about responses etc. can answer. Compare to the cases of containment, or heat, or life—all the urgent questions are already resolved before those issues are even raised.
As I say, the best way I know of to answer “is it conscious?” about X is to compare X to other systems about which I have confidence-levels about its consciousness and look for commonalities and distinctions.
If there are alternative approaches that you think give us more reliable answers, I’d love to hear about them.
I have no reliable answers! And I have low meta-confidence levels (in that it seems clear to me that people and most other creatures are conscious but I have no confidence in why I think this)
If the Dennett position still sees this as a complete bafflement but thinks it will be resolved with the so-called ‘soft’ problem, I have less of an issue than I thought I did. Though I’d still regard the view that the issue will become clear one of hope rather than evidence.
I’m splitting up my response to this into several pieces because it got long. Some other stuff:
I expect so, sure. For example, I report having experiences; one explanation of that (though hardly the only possible one) starts with my actually having experiences and progresses forward in a causal fashion.
Sure, there are many causal explanations of many phenomena, including but not limited to how bats use echolocation, that don’t posit subjective experience as part of their causal chain. For example, humas do all kinds of things without the subjective experience of doing them.
Certainly.
In the examples you give, yes, it is being ignored. So? Lots of things are being ignored in those examples… mass, electrical conductivity, street address, level of fluency in Russian, etc. If these things aren’t necessary to explain the examples, there’s nothing wrong with ignoring these things.
On the other hand, if we look at an example for which experience ought to be part of the causal chain (for example, as I note above, reporting having those experiences), subjective experience is not ignored. X happens, as a consequence of X a subjective experience Y arises, as a consequence of Y a report Z arises, and so forth. (Of course, for some reports we do have explanations that don’t presume Y… e.g., confabulation, automatic writing, etc. But that needn’t be true for all reports. Indeed, it would be surprising if it were.)
“But we don’t know what Xes give rise to the Y of subjective experience, so we don’t fully understand subjective experience!” Well, yes, that’s true. We don’t fully understand fluency in Russian, either. But we don’t go around as a consequence positing some mysterious essence of Russian fluency that resists neurobiological explanation… though two centuries ago, we might have done so. Nor should we. Neither should we posit some mysterious essence of subjective experience.
“But subjective experience is different! I can imagine what a mechanical explanation of Russian fluency would be like, but I can’t imagine what a mechanical explanation of subjective experience would be like.” Sure, I understand that. Two centuries ago, the notion of a mechanical explanation of Russian fluency would raise similar incredulity… how could a machine speak Russian? I’m not sure how I could go about answering such incredulity convincingly, but I don’t thereby conclude that machines can’t speak Russian. My incredulity may be resistant to my reason, but it doesn’t therefore compel or override my reason.
I have a lot of sympathy for this. The most plausible position of reductive materialism is simply that at some future scientific point this will become clear. But this is inevitably a statement of faith, rather than acknoweldging current acheivement. It’s very hard to compare current apparent mysteries to solved mysteries—I do get that. Having said that, I can’t even see what the steps on the way to explaining consciousness would be, and claiming there is not such thing seems not to be an option (unlike ‘life’, ‘free will’ etc.) whereas most other cases you rely on saying that you can’t see how the full extent could be achieved: a machine might speak crap Russian in some circumstances etc.
Also, if a machine can speak Russian, you can check that. I don’t know how we’d check a machine was conscious.
BTW, when I said ‘it seems subjective experience is just being ignored’, I meant ignored in your and Dennett’s arguments, not in specific explanations. I have nothing against analysing things in ways that ignore consciousness, if they work.
I don’t know what the mechanical explanation would look like, either. But I’m sufficiently aware of how ignorant my counterparts two centuries ago would have been of what a mechanical explanation for speaking Russian would look like that I don’t place too much significance on my ignorance.
I agree that testing whether a system is conscious or not is a tricky problem. (This doesn’t just apply to artificial systems.)
Indeed: though artificial systems are more intuitively difficult as we don’t have as clear an intuitive expectation.
You can take an outside view and say ‘this will dissolve like the other mysteries’. I just genuinely find this implausible, if only because you can take steps towards the other mysteries (speaking bad Russian occasionally) and because you have a clear empirical standard (russians). Whereas for consciousness I don’t have any standard for identifying another’s consciousness: I do it only by analogy with myself and by the implausibility of me having an apparently causal element that others who act similarly to me lack.
I agree that the “consciousness-detector” problem is a hard problem. I just can’t think of a better answer than the generalizing-from-commonalities strategy I discussed previously, so that’s the approach I go with. It seems capable of making progress for now.
And I understand that you find it implausible. That said, I suspect that if we solve the “soft” problem of consciousness well enough that a typical human is inclined to treat an artificial system as though it were conscious, it will start to seem more plausible.
Perhaps it will be plausible and incorrect, and we will happily go along treating computers as conscious when they are no such thing. Perhaps we’re currently going along treating dogs and monkeys and 90% of humans as conscious when they are no such thing.
Perhaps not.
Either way, plausibility (or the absence of it) doesn’t really tell us much.
Yes. This is what worries me: I can see more advances making everyone sure that computers are conscious, but my suspicion is that this will not be logical. Take the same processor and I suspect the chances of it being seen as conscious will rise sharply if it’s put in a moving machine, rise sharply again for a humanoid, again for face/voice and again for physically indistinguishable.
The problem with generalising from commonalities is that I have precisely one direct example of consciousness. Although having said that, I don’t find epiphenomenal accounts convincing, so it’s reasonable for me to think that as my statements about qualia seem to follow causally from experiencing said qualia, that other people don’t have a totally separate framework for their statements about qualia. I wouldn’t be that confident though, and it gets harder with artificial consciousness.
Sure. By the same token, if you take me, remove my ability to communicate, and encase me in an opaque cylinder, nobody will recognize me as a being with subjective experience. Or, for that matter, as a being with the ability to construct English sentences.
We are bounded intellects reasoning under uncertainty in a noisy environment. We will get stuff wrong. Sometimes it will be important stuff.
I agree. And, as I said initially, I apply the same reasoning not only to the statements I make in English, but to all manner of behaviors that “seem to rise from my qualia,” as you put it… all of it is evidence in favor of other organisms also having subjective experience, even organisms that don’t speak English.
How confident are you that I possess subjective experience?
Would that confidence rise significantly if we met in person and you verified that I have a typical human body?
Agreed.
Consciousness does seem different in that we can have a better and better understanding of all the various functional elements but that we’re 1) left with a sort of argument from analogy for others having qualia 2) even if we can resolve(1), I can’t see how we can start to know whether my green is your red etc. etc.
I can’t think of many comparable cases: certainly I don’t think containership is comparable. You and I could end up looking at the AI in the moment before it destroys/idealises/both the world and say ‘gosh, I wonder if it’s conscious’. This is nothing like the casuistic ‘but what about this container gives it its containerness’. I think we’re on the same point here, though?
I’m intuitively very confident you’re conscious: and yes, seeing you were human would help (in that one of the easiest ways I can imagine you weren’t conscious is that you’re actually a computer designed to post about things on less wrong. This would also explain why you like Dennett—I’ve always suspected he’s a qualia-less robot too! ;-)
Yes, I agree that we’re much more confused about subjective experience than we are about containership.
We’re also more confused about subjective experience than we are about natural language, about solving math problems, about several other aspects of cognition. We’re not _un_confused about those things, but we’re less confused than we used to be. I expect us to grow still less confused over time.
I disagree about the lack of comparable cases. I agree about containers; that’s just an intuition pump. But the issues that concern you here arise for any theoretical construct for which we have only indirect evidence. The history of science is full of such things. Electrons. Black holes. Many worlds. Fibromyalgia. Phlogiston. Etc.
What makes subjective experience different is not that we lack the ability to perceive it directly; that’s pretty common. What makes it different is that we can perceive it directly in one case, as opposed to the other stuff where we perceive it directly in zero cases.
Of course, it’s also different from many of them in that it matters to our moral reasoning in many cases. I can’t think of a moral decision that depends on whether phlogiston exists, but I can easily think of a moral decision that depends on whether cows have subjective experiences. OTOH, it still isn’t unique; some people make moral decisions that depend on the actuality of theoretical constructs like many worlds and PTSD.
Fair enough. As an intuition pump, for me at least, it’s unhelpful: it gave the impression that you thought that consciousness was merely a label being mistaken for a thing (like ‘life’ as something beyond its parts).
Only having indirect evidence isn’t the problem. For a black hole, I care about the observable functional parts. I wouldn’t be being sucked towards it and being crushed while going ‘but is it really a black hole?’ A black hole is like a container here: what matter are the functional bits that make it up. For consciousness, I care if a robot can reason and can display conscious-type behaviour, but I also care if it can experience and feel.
Many worlds could be comparable if there is evidence that means that there are ‘many worlds’ but people are vague about if these worlds actually exist. And you’re right, this is also a potentially morally relevant point.
Insofar as people infer from the fact of subjective experience that there is some essence of subjective experience that is, as you say, “beyond its parts” (and their patterns of interaction), I do in fact think they are mistaking a label for a thing.
I dunno about essences. The point is that you can observe lots of interactions of neurons and behaviours and be left with an argument from analogy to say “they must be conscious because I am and they are really similar, and the idea that my consciousness is divorced from what I do is just wacky”.
You can observe all the externally observable, measurable things that a black hole or container can do, and then if someone argues about essences you wonder if they’re actually referring to anything: it’s a purely semantic debate. But you can observe all the things a fish, or tree, or advanced computer can do, predict it for all useful purposes, and still not know if it’s conscious. This is bothersome. But it’s not to do with essences, necessarily.
Insofar as people don’t infer something else beyond the parts of a (for example) my body and their pattern of interactions that account for (for example) my subjective experience, I don’t think they are mistaking a label for a thing.
Well, until we know how to identify if something/someone is conscious, it’s all a bit of a mystery: I couldn’t rule out consciousness being some additional thing. I have an inclination to do so because it seems unparsimonious, but that’s it.
I’m splitting up my response to this into several pieces because it got long.
The key bit, IMHO:
And I would agree with you.
“No,” replies A, “you miss the point completely. I don’t ask whether a container can contain things; clearly it can, I observe it doing so. I ask how it contains things. What is the explanation for its demonstrated ability to contain things? Containership is not just a function,” A insists, “though I understand you want to treat it as one. No, containership is a fundamental essence. You can’t simply ignore the hard question of “is X a container?” in favor of thinking about simpler, merely functional questions like “can X contain Y?”. And, while we’re at it,” A coninues, “what makes you think that an artificial container, such as we build all the time, is actually containing anything rather than merely emulating containership? Sure, perhaps we can’t tell the difference, but that doesn’t mean there isn’t a difference.”
I take it you don’t find A’s argument convincing, and neither do I, but it’s not clear to me what either of us could say to A that A would find at all compelling.
Maybe we couldn’t, but A is simply asserting that containership is a concept beyond its parts, whereas I’m appealing directly to experience: the relevance of this is that whether something has experience matters. Ultimately for any case, if others just express bewilderment in your concepts and apparently don’t get what you’re talking about, you can’t prove it’s an issue. But at any rate, most people seem to have subjective experience.
Being conscious isn’t a label I apply to certain conscious-type systems that I deem ‘valuable’ or ‘true’ in some way. Rather, I want to know what systems should be associated with the clearly relevant and important category of ‘conscious’
My thoughts about how I go about associating systems with the expectation of subjective experience are elsewhere and I have nothing new to add to it here.
As regards you and A… I realize that you are appealing directly to experience, whereas A is merely appealing to containment, and I accept that it’s obvious to you that experience is importantly different from containment in a way that makes your position importantly non-analogous to A’s.
I have no response to A that I expect A to find compelling… they simply don’t believe that containership is fully explained by the permeability and topology of containers. And, you know, maybe they’re right… maybe some day someone will come up with a superior explanation of containerhood that depends on some previously unsuspected property of containers and we’ll all be amazed at the realization that containers aren’t what we thought they were. I don’t find it likely, though.
I also have no response to you that I expect you to find compelling. And maybe someday someone will come up with a superior explanation of consciousness that depends on some previously unsuspected property of conscious systems, and I’ll be amazed at the realization that such systems aren’t what I thought they were, and that you were right all along.
Are you saying you don’t experience qualia and find them a bit surprising (in a way you don’t for containerness)? I find it really hard to not see arguments of this kind as a little disingenous: is the issue genuinely not difficult for some people, or is this a rhetorical stance intended to provoke better arguments, or awareness of the weakness of current arguments?
I have subjective experiences. If that’s the same thing as experiencing qualia, then I experience qualia.
I’m not quite sure what you mean by “surprising” here… no, it does not surprise me that I have subjective experiences, I’ve become rather accustomed to it over the years. I frequently find the idea that my subjective experiences are a function of the formal processes my neurobiology implements a challenging idea… is that what you’re asking?
Then again, I frequently find the idea that my memories of my dead father are a function of the formal processes my neurobiology implements a challenging idea as well. What, on your view, am I entitled to infer from that?
Yes, I meant surprising in light of other discoveries/beliefs.
On memory: is it the conscious experience that’s challenging (in which case it’s just a sub-set of the same issue) or do you find the functional aspects of memory challenging? Even though I know almost nothing about how memory works, I can see plausible models and how it could work, unlike consciousness.
Isn’t our objection to A’s position that it doesn’t pay rent in anticipated experience? If one thinks there is a “hard problem of consciousness” such that different answers would cause one to behave differently, then one must take up the burden of identifying what the difference would look like, even if we can’t create a measuring device to find it just now.
If A means that we cannot determine the difference in principle, then there’s nothing we should do differently. If A means that a measuring device does not currently exist, he needs to identify the range of possible outputs of the device.
This may be a situation where that’s a messy question. After all, qualia are experience. I keep expecting experiences, and I keep having experiences. Do experiences have to be publicly verifiable?
If two theories both lead me to anticipate the same experience, the fact that I have that experience isn’t grounds for choosing among them.
So, sure, the fact that I keep having experiences is grounds for preferring a theory of subjective-experience-explaining-but-otherwise-mysterious qualia over a theory that predicts no subjective experience at all, but not necessarily grounds for preferring it to a theory of subjective-experience-explaining-neural-activity.
They don’t necessarily once you start talking about uploads, or the afterlife for that matter.
different answers to the HP would undoubtedly change our behaviour, because they would indicate different classes of entity have feelings impacting morality. Indeed, it is pretty hard to think of anything more impactive.
The measuring device for conscious experience is consciousness, which is the whole problem.
Sure. But in this sense, false believed answers to the HP are no different from true believed answers.… that is, they would both potentially change our behavior the way you describe.
I suspect that’s not what TimS meant.
That is the case for most any belief you hold (unless you mean “in the exact same way”, not as “change behavior”). You may believe there’s a burglar in your house, and that will impact your actions, whether it be false or true. Say you believe that it’s more likely there is a burglar, you are correct in acting upon that belief even if it turns out to incorrect. It’s not AIXI’s fault if it believes in the wrong thing for the right reasons.
In that sense, you can choose an answer for example based on complexity considerations. In the burglar example, the answer you choose (based on data such as crime rate, cat population etc.) can potentially be further experimentally “verified” (the probability increased) as true or false, but even before such verification, your belief can still be strong enough to act upon.
After all, you do act upon your belief that “I am not living in a simulation which will eventually judge and reward me only for the amount of cheerios I’ve eaten”. It also doesn’t lead to different expected experiences at the present time, yet you also choose to act thus as if it were true. Prior based on complexity considerations alone, yet strong enough to act upon. Same when thinking about whether the sun has qualia (“hot hot hot hot hot”).
(Bit of a hybrid fusion answer also meant to refer to our neighboring discussion branch.)
Cheerio!
Yes, I agree with all of this.
Well, in the case of “do landslides have qualia”, Occam’s Razor could be used to assign probabilities just the same as we assign probabilities in the “cheerio simulation” example. So we’ve got methodology, we’ve got impact, enough to adopt a stance on the “psychic unity of the cosmos”, no?
I’m having trouble following you, to be honest.
My best guess is that you’re suggesting that, with respect to systems that do not manifest subjective experience in any way we recognize or understand, Occam’s Razor provides grounds to be more confident that they have subjective experience than that they don’t.
If that’s what you mean, I don’t see why that should be.
If that’s not what you mean, can you rephrase the question?
I think it’s conceivable if not likely that Occam’s Razor would favor or disfavor qualia as a property of more systems than just those that seem to show or communicate them in terms we’re used to. I’m not sure which, but it is a question worth pondering, with an impact on how we view the world, and accessible through established methodology, to a degree.
I’m not advocating assigning a high probability to “landslides have raw experience”, I’m advocating that it’s an important question, the probability of which can be argued. I’m an advocate of the question, not the answer, so to speak. And as such opposed to “I really can’t see why anyone should care one way or the other”.
Ah, I see.
So, I stand by my assertion that in the absence of evidence one way or the other, I really can’t see why anyone should care.
But I agree that to the extent that Occam’s Razor type reasoning provides evidence, that’s a reason to care.
And if it provided strong evidence one way or another (which I don’t think it does, and I’m not sure you do either) that would provide a strong reason to care.
I have evidence in the form of by personal experience of qualia. Granted, I have no way of showing you that evidence, but that doesn’t mean I don’t have it.
Agreed that the ability to share evidence with others is not a necessary condition of having evidence. And to the extent that I consider you a reliable evaluator of (and reporter of) evidence, your report is evidence, and to that extent I have a reason to care.
The point has been made that we should care because qualia have moral implications.
Moral implications of a proposition in the absence of evidence one way or another for that proposition are insufficient to justify caring.
If I actually care about the experiences of minds capable of experiences, I do best to look for evidence for the presence or absence of such experiences.
Failing such evidence, I do best to concentrate my attention elsewhere.
It’s possible to have both a strong reason to care, and weak evidence, ie due the moral hazard dependent on some doubtful proposition. People often adopt precautionary principles in such scenarios.
I don’t think that’s the situation here though. That sounds like a description of this situation: (imagine) we have weak evidence that 1) snakes are sapient, and we grant that 2) sapience is morally significant. Therefore (perhaps) we should avoid wonton harm to snakes.
Part of why this argument might make sense is that (1) and (2) are independent. Our confidence in (2) is not contingent on the small probability that (1) is true: whether or not snakes are sapient, we’re all agreed (lets say) that sapience is morally significant.
On the other hand, the situation with qualia is one where we have weak evidence (suppose) that A) qualia are real, and we grant that B) qualia are morally significant.
The difference here is that (B) is false if (A) is false. So the fact that we have weak evidence for (A) means that we can have no stronger (and likely, we must have yet weaker) evidence for (B).
Does the situation change significantly if “the situation with qualia” is instead framed as A) snakes have qualia and B) qualia are morally significant?
Yes, if the implication of (A) is that we’re agreed on the reality of qualia but are now wondering whether or not snakes have them. No, if (A) is just a specific case of the general question ‘are qualia real?’. My point was probably put in a confusing way: all I mean to say was that Juno seemed to be arguing as if it were possible to be very confident about the moral significance of qualia while being only marginally confident about their reality.
(nods) Makes sense.
What? Are you saying we have weak evidence for even in ourselves?
What I think of the case for qualia is beside the point, I was just commenting on your ‘moral hazard’ argument. There you said that even if we assume that we have only weak evidence for the reality of qualia, we should take the possibility seriously, since we can be confident that qualia are morally significant. I was just pointing out that this argument is made problematic by the fact that our confidence in the moral significance of qualia can be no stronger than our confidence in their reality, and therefore by assumption must be weak.
But of course it can. i can be much more confident in
(P → Q)
than I am in P. For instance, I can be highly confident that if i won the lottery, I could buy a yacht.
I am guessing that Juno_Watt means that strong evidence for our own perception of qualia makes them real enough to seriously consider their moral significance, whether or not they are “objectively real”.
Yes, they often do.
On your view, is there a threshold of doubtfulness of a proposition below which it is justifiable to not devote resources to avoiding the potential moral hazard of that proposition being true, regardless of the magnitude of that moral hazard?
i don’t think ir’s likely my house will catch fire, but i take out fire insurance. OTOH, if i don’t set a lower bound I will be susceptible to pascal’s muggings.
He may have meant something like “Qualiaphobia implies we would have no expereinces at all”. However, that all depends on what you mean by experience. I don’t think the Expected Experience criterion is useful here (or anywhere else)