There is little prospect of an outcome that realizes even the value of being interesting, unless the first superintelligences undergo detailed inheritance from human values
No doubt a Martian Yudkowsy would make much the same argument—but they can’t both be right. I think that neither of them are right—and that the conclusion is groundless.
Complexity theory shows what amazing things can arise from remarkably simple rules. Values are evidently like that—since even “finding prime numbers” fills the galaxy with an amazing, nanotech-capable spacefaring civilization—and if you claim that a nanotech-capable spacefaring civilization is not “interesting” you severely need recalibrating.
I think Martian Yudkowsky is a dangerous intuition pump. We’re invited to imagine a creature just like Eliezer except green and with antennae; we naturally imagine him having values as similar to us as, say, a Star Trek alien. From there we observe the similarity of values we just pushed in, and conclude that values like “interesting” are likely to be shared across very alien creatures. Real Martian Yudkowsky is much more alien than that, and is much more likely to say
There is little prospect of an outcome that realizes even the value of being flarn, unless the first superintelligences undergo detailed inheritance from Martian values.
Imagine, an intelligence that didn’t have the universal emotion of badweather!
Of course, extraterrestrial sentients may possess physiological states corresponding to limbic-like emotions that have no direct analog in human experience. Alien species, having evolved under a different set of environmental constraints than we, also could have a different but equally adaptive emotional repertoire. For example, assume that human observers land on another and discover an intelligent animal with an acute sense of absolute humidity and absolute air pressure. For this creature, there may exist an emotional state responding to an unfavorable change in the weather. Physiologically, the emotion could be mediated by the ET equivalent of the human limbic system; it might arise following the secretion of certain strength-enhancing and libido-arousing hormones into the alien’s bloodstream in response to the perceived change in weather. Immediately our creature begins to engage in a variety of learned and socially-approved behaviors, including furious burrowing and building, smearing tree sap over its pelt, several different territorial defense ceremonies, and vigorous polygamous copulations with nearby females, apparently (to humans) for no reason at all. Would our astronauts interpret this as madness? Or love? Lust? Fear? Anger? None of these is correct, of course the alien is feeling badweather.
I suggest you guys taboo interesting, because I strongly suspect you’re using it with slightly different meanings. (And BTW, as a Martian Yudkowsky I imagine something with values at least as alien as Babyeaters’ or Superhappys’.)
It’s another discussion, really, but it sounds as though you are denying the idea of “interestingness” as a universal instrumental value—whereas I would emphasize that “interestingness” is really just our name for whether something sustains our interest or not—and ‘interest’ is a pretty basic functional property of any agent with mobile sensors. There’ll be other similarities in the area too—such as novelty-seeking. So shared common ground is only to be expected.
Anyway, I am not too wedded to Martian Yudkowsky. The problematical idea is that you could have a nanotech-capable spacefaring civilization that is not “interesting”. If such a thing isn’t “interesting” then—WTF?
So: do you really think that humans wouldn’t find a martian civilization interesting? Surely there would be many humans who would be incredibly interested.
I find Jupiter interesting. I think a paperclip maximizer (choosing a different intuition pump for the same point) could be more interesting than Jupiter, but it would generate an astronomically tiny fraction of the total potential for interestingness in this universe.
Life isn’t much of an “interestingness” maximiser. Expecting to produce more than a tiny fraction of the total potential for interestingness in this universe seems as though it would be rather unreasonable.
I agree that a paperclip maximiser would be more boring than an ordinary entropy-maximising civilization—though I don’t know by how much—probably not by a huge amount—the basic problems it faces are much the same—the paperclip maximiser just has fewer atoms to work with.
since even “finding prime numbers” fills the galaxy with an amazing, nanotech-capable spacefaring civilization
The goal “finding prime numbers” fills the galaxy with an amazing, nonotech-capable spacefaring network of computronium which finds prime numbers, not a civilization, and not interesting.
Maybe we should taboo the term interesting? My immediate reaction was that that sounded really interesting. This suggests that the term may not be a good one.
Fair enough. By “not interesting”, I meant it is not the sort of future that I want to achieve. Which is a somewhat ideosyncratic usage, but I think inline with the context.
Not just computronium—also sensors and actuators—a lot like any other cybernetic system. There would be mining, spacecraft caft, refuse collection, recycling, nanotechnology, nuclear power and advanced machine intelligence with planning, risk assessment, and so forth. You might not be interested—but lots of folk would be amazed and fascinated.
If using another creature’s values is effective at producing something “interesting”, then ‘detailed inheritance from human values’ is clearly not needed to produce this effect.
There is little prospect of an outcome that realizes even the value of being interesting, unless the first superintelligences undergo detailed inheritance from human values
and Mars Yudkowsky (MY) argues:
There is little prospect of an outcome that realizes even the value of being interesting, unless the first superintelligences undergo detailed inheritance from martian values
and that one of these things has to be incorrect? But if martian and human values are similar, then they can both be right, and if martian and human values are not similar, then they refer to different things by the word “interesting”.
In any case, I read EY’s statement as one of probability-of-working-in-the-actual-world-as-it-is, not a deep philosophical point—“this is the way that would be most likely to be successful given what we know”. In which case, we don’t have access to martian values and therefore invoking detailed inheritance from them would be unlikely to work. MY would presumably be in an analogous situation.
But if martian and human values are similar, then they can both be right
I was assuming that ‘detailed inheritance from human values’ doesn’t refer to the same thing as “detailed inheritance from martian values”.
if martian and human values are not similar, then they refer to different things by the word “interesting”.
Maybe—but humans not finding martians interesting seems contrived to me. Humans have a long history of being interested in martians—with feeble evidence of their existence.
In any case, I read EY’s statement as one of probability-of-working-in-the-actual-world-as-it-is, not a deep philosophical point—“this is the way that would be most likely to be successful given what we know”. In which case, we don’t have access to martian values and therefore invoking detailed inheritance from them would be unlikely to work
Right—so, substitute in “dolphins”, “whales”, or another advanced intelligence that actually exists.
Do you actually disagree with my original conclusion? Or is this just nit-picking?
I actually disagree that tiling the universe with prime number calculators would result in an interesting universe from my perspective (dead). I think it’s nonobvious that dolphin-CEV-AI-paradise would be human-interesting. I think it’s nonobvious that martian-CEV-AI-paradise would be human-interesting, given that these hypothetical martians diverge from humans to a significant extent.
I actually disagree that tiling the universe with prime number calculators would result in an interesting universe from my perspective (dead).
I think it’s violating the implied premises of the thought experiment to presume that the “interestingness evaluator” is dead. There’s no terribly-compelling reason to assume that—it doesn’t follow from the existence of a prime number maximizer that all humans are dead.
I may have been a little flip there.
My understanding of the thought experiment is—something extrapolates some values and maximizes them, probably using up most of the universe, probably becoming the most significant factor in the species’ future and that of all sentients, and the question is whether the result is “interesting” to us here and now, without specifying the precise way to evaluate that term. From that perspective, I’d say a vast uniform prime-number calculator, whether or not it wipes out all (other?) life, is not “interesting”, in that it’s somewhat conceptually interesting as a story but a rather dull thing to spend most of a universe on.
Today’s ecosystems maximise entropy. Maximising primeness is different, but surely not greatly more interesting—since entropy is widely regarded as being tedious and boring.
Intriguing! But even granting that, there’s a big difference between extrapolating the values of a screwed-up offshoot of an entropy-optimizing process and extrapolating the value of “maximize entropy”. Or do you suspect that a FOOMing AI would be much less powerful and more prone to interesting errors than Eliezer believes?
Truly maximizing entropy would involve burning everything you can burn, tearing the matter of solar systems apart, accelerating stars towards nova, trying to accelerate the evaporation of black holes and prevent their formation, and other things of this sort. It’d look like a dark spot in the sky that’d get bigger at approximately the speed of light.
Fires are crude entropy maximisers. Living systems destroy energy dradients at all scales, resulting in more comprehensive devastation than mere flames can muster.
Of course, maximisation is often subject to constraints. Your complaint is rather like saying that water doesn’t “truly minimise” its altitude—since otherwise it would end up at the planet’s core. That usage is simply not what the terms “maximise” and “minimise” normally refer to.
It says:
No doubt a Martian Yudkowsy would make much the same argument—but they can’t both be right. I think that neither of them are right—and that the conclusion is groundless.
Complexity theory shows what amazing things can arise from remarkably simple rules. Values are evidently like that—since even “finding prime numbers” fills the galaxy with an amazing, nanotech-capable spacefaring civilization—and if you claim that a nanotech-capable spacefaring civilization is not “interesting” you severely need recalibrating.
To end with, a quote from E.Y.:
I think Martian Yudkowsky is a dangerous intuition pump. We’re invited to imagine a creature just like Eliezer except green and with antennae; we naturally imagine him having values as similar to us as, say, a Star Trek alien. From there we observe the similarity of values we just pushed in, and conclude that values like “interesting” are likely to be shared across very alien creatures. Real Martian Yudkowsky is much more alien than that, and is much more likely to say
Imagine, an intelligence that didn’t have the universal emotion of badweather!
I suggest you guys taboo interesting, because I strongly suspect you’re using it with slightly different meanings. (And BTW, as a Martian Yudkowsky I imagine something with values at least as alien as Babyeaters’ or Superhappys’.)
It’s another discussion, really, but it sounds as though you are denying the idea of “interestingness” as a universal instrumental value—whereas I would emphasize that “interestingness” is really just our name for whether something sustains our interest or not—and ‘interest’ is a pretty basic functional property of any agent with mobile sensors. There’ll be other similarities in the area too—such as novelty-seeking. So shared common ground is only to be expected.
Anyway, I am not too wedded to Martian Yudkowsky. The problematical idea is that you could have a nanotech-capable spacefaring civilization that is not “interesting”. If such a thing isn’t “interesting” then—WTF?
Yes, I am; I think that the human value of interestingness is much, much more specific than the search space optimization you’re pointing at.
[This reply was to an earlier version of timtyler’s comment]
So: do you really think that humans wouldn’t find a martian civilization interesting? Surely there would be many humans who would be incredibly interested.
I find Jupiter interesting. I think a paperclip maximizer (choosing a different intuition pump for the same point) could be more interesting than Jupiter, but it would generate an astronomically tiny fraction of the total potential for interestingness in this universe.
Life isn’t much of an “interestingness” maximiser. Expecting to produce more than a tiny fraction of the total potential for interestingness in this universe seems as though it would be rather unreasonable.
I agree that a paperclip maximiser would be more boring than an ordinary entropy-maximising civilization—though I don’t know by how much—probably not by a huge amount—the basic problems it faces are much the same—the paperclip maximiser just has fewer atoms to work with.
The goal “finding prime numbers” fills the galaxy with an amazing, nonotech-capable spacefaring network of computronium which finds prime numbers, not a civilization, and not interesting.
Maybe we should taboo the term interesting? My immediate reaction was that that sounded really interesting. This suggests that the term may not be a good one.
Fair enough. By “not interesting”, I meant it is not the sort of future that I want to achieve. Which is a somewhat ideosyncratic usage, but I think inline with the context.
What if we added a module that sat around and was really interested in everything going on?
Not just computronium—also sensors and actuators—a lot like any other cybernetic system. There would be mining, spacecraft caft, refuse collection, recycling, nanotechnology, nuclear power and advanced machine intelligence with planning, risk assessment, and so forth. You might not be interested—but lots of folk would be amazed and fascinated.
Why?
If using another creature’s values is effective at producing something “interesting”, then ‘detailed inheritance from human values’ is clearly not needed to produce this effect.
So you’re saying Earth Yudkowsky (EY) argues:
and Mars Yudkowsky (MY) argues:
and that one of these things has to be incorrect? But if martian and human values are similar, then they can both be right, and if martian and human values are not similar, then they refer to different things by the word “interesting”.
In any case, I read EY’s statement as one of probability-of-working-in-the-actual-world-as-it-is, not a deep philosophical point—“this is the way that would be most likely to be successful given what we know”. In which case, we don’t have access to martian values and therefore invoking detailed inheritance from them would be unlikely to work. MY would presumably be in an analogous situation.
I was assuming that ‘detailed inheritance from human values’ doesn’t refer to the same thing as “detailed inheritance from martian values”.
Maybe—but humans not finding martians interesting seems contrived to me. Humans have a long history of being interested in martians—with feeble evidence of their existence.
Right—so, substitute in “dolphins”, “whales”, or another advanced intelligence that actually exists.
Do you actually disagree with my original conclusion? Or is this just nit-picking?
I actually disagree that tiling the universe with prime number calculators would result in an interesting universe from my perspective (dead). I think it’s nonobvious that dolphin-CEV-AI-paradise would be human-interesting. I think it’s nonobvious that martian-CEV-AI-paradise would be human-interesting, given that these hypothetical martians diverge from humans to a significant extent.
I think it’s violating the implied premises of the thought experiment to presume that the “interestingness evaluator” is dead. There’s no terribly-compelling reason to assume that—it doesn’t follow from the existence of a prime number maximizer that all humans are dead.
I may have been a little flip there. My understanding of the thought experiment is—something extrapolates some values and maximizes them, probably using up most of the universe, probably becoming the most significant factor in the species’ future and that of all sentients, and the question is whether the result is “interesting” to us here and now, without specifying the precise way to evaluate that term. From that perspective, I’d say a vast uniform prime-number calculator, whether or not it wipes out all (other?) life, is not “interesting”, in that it’s somewhat conceptually interesting as a story but a rather dull thing to spend most of a universe on.
Today’s ecosystems maximise entropy. Maximising primeness is different, but surely not greatly more interesting—since entropy is widely regarded as being tedious and boring.
Intriguing! But even granting that, there’s a big difference between extrapolating the values of a screwed-up offshoot of an entropy-optimizing process and extrapolating the value of “maximize entropy”. Or do you suspect that a FOOMing AI would be much less powerful and more prone to interesting errors than Eliezer believes?
Truly maximizing entropy would involve burning everything you can burn, tearing the matter of solar systems apart, accelerating stars towards nova, trying to accelerate the evaporation of black holes and prevent their formation, and other things of this sort. It’d look like a dark spot in the sky that’d get bigger at approximately the speed of light.
Fires are crude entropy maximisers. Living systems destroy energy dradients at all scales, resulting in more comprehensive devastation than mere flames can muster.
Of course, maximisation is often subject to constraints. Your complaint is rather like saying that water doesn’t “truly minimise” its altitude—since otherwise it would end up at the planet’s core. That usage is simply not what the terms “maximise” and “minimise” normally refer to.
Yeah! Compelling, but not “interesting”. Likewise, I expect that actually maximizing the fitness of a species would be similarly “boring”.