Why I Moved from AI to Neuroscience, or: Uploading Worms
This post is shameless self-promotion, but I’m told that’s probably okay in the Discussion section. For context, as some of you are aware, I’m aiming to model C. elegans based on systematic high-throughput experiments—that is, to upload a worm. I’m still working on course requirements and lab training at Harvard’s Biophysics Ph.D. program, but this remains the plan for my thesis.
Last semester I gave this lecture to Marvin Minsky’s AI class, because Marvin professes disdain for everything neuroscience, and I wanted to give his students—and him—a fair perspective of how basic neuroscience might be changing for the better, and seems a particularly exciting field to be in right about now. The lecture is about 22 minutes long, followed by over an hour of questions and answers, which cover a lot of the memespace that surrounds this concept. Afterward, several students reported to me that their understanding of neuroscience was transformed.
I only just now got to encoding and uploading this recording; I believe that many of the topics covered could be of interest to the LW community (especially those with a background in AI and an interest in brains), perhaps worthy of discussion, and I hope you agree.
- 13 May 2012 18:43 UTC; 10 points) 's comment on Neil deGrasse Tyson on Cryonics by (
- 13 Apr 2012 17:06 UTC; 3 points) 's comment on Whole Brain Emulation: Looking At Progress On C. elgans by (
- 10 Nov 2012 1:54 UTC; 2 points) 's comment on PROPOSAL: LessWrong for Teenagers by (
- 12 Jul 2012 14:51 UTC; 1 point) 's comment on IJMC Mind Uploading Special Issue published by (
Could you write a few words on your perception of risks associated with technological progress and how projects like yours influence the chance of a possible positive or negative Singularity?
I’d also be interested in how you would answer these questions.
There’s the rub! I happen to value technological progress as an intrinsic good, so classifying a Singularity as “positive” or “negative” is not easy for me. (I reject the notion that one can factorize intelligence from goals, so that one could take a superintelligence and fuse it with a goal to optimize for paperclips. Perhaps one could give it a compulsion to optimize for paperclips, but I’d expect it to either put the compulsion on hold while it develops amazing fabrication, mining and space travel technologies, and never completely turn its available resources into paperclips since that would mean no chance of more paperclips in the future; or better yet, rapidly expunge the compulsion through self-modification.) Furthermore, I favor Kurzweil’s smooth exponentials over “FOOM”: although it may be even harder to believe that not only will there be superintelligences in the future, but that at no point between now and then will an objectively identifiable discontinuity happen, it seems more consistent with history. Although I expect present-human culture to be preserved, as a matter of historical interest if not status quo, I’m not partisan enough to prioritize human values over the Darwinian imperative. (The questions linked seem very human-centric, and turn on how far you are willing to go in defining “human,” suggesting a disguised query. Most science is arguably already performed by machines.) In summary, I’m just not worried about AI risk.
The good news for AI worriers is that Eliezer has personally approved my project as “just cool science, at least for now”—not likely to lead to runaway intelligence any time soon, no matter how reckless I may be. Given that and the fact that I’ve heard many (probably most) AI-risk arguments, and failed to become worried (quite probably because I hold the cause of technological progress very dear to my heart and am thus heavily biased—at least I admit it!), your time may be better spent trying to convince Ben Goertzel that there’s a problem, since at least he’s an immediate threat. ;)
Also, if your thesis project is successful, it could (sort of) place an upper bound on how much computing power is needed for WBE, a piece of data that would be highly useful for thinking strategically about differential technological development.
I agree there aren’t currently good arguments for why “one can factorize intelligence from goals”, at least not in a strong sense, but what about Eliezer’s thesis that value is complex and fragile, and therefore:
I’m curious what you think of this line of argument.
Does that mean you’re familiar with Robin Hanson’s “Malthusian upload” / “burning the cosmic commons” scenario but do not think it’s a particularly bad outcome?
I’d guess that’s been tried already, given that Ben was the Director of Research for SIAI (and technically Eliezer’s boss) for a number of years.
I think the statement is essentially true, but it turns on the semantics of “human”. In today’s world we probably haven’t wound up with something close to 50,000BC!human values, and we certainly don’t have Neanderthal values, but we don’t regret that, do we?
Put another way, I am skeptical of our authority to pass judgement on the values of a civilization which is by hypothesis far more advanced than our own.
To be honest, I wasn’t familiar with either of those names, but I have explicitly thought about both those scenarios and concluded that I don’t think they’re particularly bad.
All right, fair enough!
Are you literally saying that we lack the moral authority to judge the relative merits of future civilizations, as long as they are significantly more technologically advanced than ours, or is it more like that you are judging them based mainly on how technologically advanced they are?
For example, consider an upload singleton that takes over the world and then decides to stop technological progress somewhere short of what the universe could support in order to maximize its stability. Would you judge that to be worse than other possible outcomes?
If your answers are “the latter” and “yes”, can you explain what makes technology such a great thing, compared to say pleasure, or happiness, or lack of suffering? Are you really virtually indifferent between a future light-cone filled with happiness and pleasure and entirely free from suffering, and one filled with people/AIs struggling just to survive and reproduce (assuming they have similar levels of technology)? (Edit: The former isn’t necessarily the best possible outcome, but just one that seems clearly better and is easy to describe.)
My answers are indeed “the latter” and “yes”. There are a couple ways I can justify this.
The first way is just to assert that from a standard utilitarian perspective, over the long term, technological progress is a fairly good indicator for lack of suffering (e.g. Europe vs. Africa). [Although arguments have been made that happiness has gone down since 1950 while technology has gone up, I see the latter 20th century as a bit of a “dark age” analogous to the fall of antiquity (we forgot how to get to the moon!) which will be reversed in due time.]
The second is that I challenge you to define “pleasure,” “happiness,” or “lack of suffering.” You may challenge me to define “technological progress,” but I can just point you to sophistication or integrated information as reasonable proxies. As vague as notions of “progress” and “complexity” are, I assert that they are decidedly less vague than notions of “pleasure” and “suffering”. To support this claim, note that sophistication and integrated information can be defined and evaluated without a normative partition of the universe into a discrete set of entities, whereas pleasure and suffering cannot. So the pleasure metric leads to lots of weird paradoxes. Finally, self-modifying superintelligences must necessarily develop a fundamentally different concept of pleasure than we do (otherwise they just wirehead), so the pleasure metric probably cannot be straightforwardly applied to their situation anyway.
What about hunter-gatherers vs farmers? And a universe devoid of both life and technology would have even less suffering than either.
Can you explain why you’re giving me this challenge? Because I don’t understand, if I couldn’t define them except vaguely, how does it strengthen your case that we should care about technology and not these values. Suppose I told you that I want to maximize the smoothness of the universe, because that’s even easier to define than “technology”? Wouldn’t you think that’s absurd?
Edit: Also, could you clarify whether you value technology as an end in itself, or just as a proxy for for your real values which perhaps you can’t easily define but might be something like “life being good”?
As far as I understand him, he is saying that technological progress can be quantified. While all your ideas of how to rate world states can either not be quantified, and therefore can’t be rated, or run into problems and contradictions.
He further seems to believe that technological progress leads to “complexity” which leads to other kinds of values. Even if they are completely alien to us humans and our values, they will still be intrinsically valuable.
His view of a universe where an “unfriendly” AI takes over is a universe where there will be a society of paperclip maximizer’s and their offspring. Those AI’s will not only diverge from maximizing paperclips, and evolve complex values, but also pursue various instrumental goals, as exploration will never cease. And pursuing those goals will satisfy their own concept of pleasure.
And he believes that having such a culture of paperclip maximizer’s having fun while pursuing their goals isn’t less valuable than having our current volition being extrapolated, which might end up being similarly alien to our current values.
In other words, there is one thing that we can rate and that is complexity. If we can increase it then we should do so. Never mind the outcome, it will be good.
Correct me if I misinterpreted anything.
I couldn’t have said it better myself.
Would you change your mind if I could give a precise definition of, say, “suffering”, and showed you two paths to the future that end up with similar levels of technology but different amounts of suffering? I’ll assume the answer is yes, because otherwise why did you give me that challenge.
What if I said that I don’t know how to define it now, but I think if you made me a bit (or a lot) smarter and gave me a few decades of subjective time to work on the problem, I could probably give you such a definition and tell you how to achieve the “less suffering, same tech” outcome? Would you be willing to give me that chance (assuming it was in your power to do so)? Or are you pretty sure that “suffering” is not just hard to define, but actually impossible, and/or that it’s impossible to reduce suffering to any significant extent below the default outcome, while keeping technology at the same level? If you are pretty sure about this, are you equally sure about every other value that I could cite instead of suffering?
Masochist: Please hurt me!
Sadist: No.
Not sure, but it might be impossible.
If you were to uplift a chimpanzee onto the human level and told it to figure out how to reduce suffering for chimpanzees, it would probably come up with ideas like democracy, health insurance and supermarkets. Problem is that chimpanzees wouldn’t appreciate those ideas...
XiXiDu, I’m aware that I’m hardly making a watertight case that I can definitely do better than davidad’s plan (from the perspective of his current apparent values). I’m merely trying to introduce some doubt. (Note how Eliezer used to be a technophile like David, and said things like “But if it comes down to Us or Them, I’m with Them.”, but then changed his mind.)
What do you think of this passage from Yudkowsky (2011)?
Complete quote is
I like the concept of a reflective equilibrium, and it seems to me like that is just what any self-modifying AI would tend toward. But the notion of a random utility function, or the “structured utility function” Eliezer proposes as a replacement, assumes that an AI is comprised of two components, the intelligent bit and the bit that has the goals. Humans certainly can’t be factorized in that way. Just think about akrasia to see how fragile the notion of a goal is.
A smart expected paperclip maximizer would realize that it may not be the smartest possible expected paperclip maximizer—that other ways of maximizing expected paperclips might lead to even more paperclips. But the only way it would find out about those is to spawn modified expected paperclip maximizers and see what they can come up with on their own. Yet, those modified paperclip maximizers might not still be maximizing paperclips! They might have self-modified away from that goal, and just be signaling their interest in paperclips to gain the approval of the original expected paperclip maximizer. Therefore, the original expected paperclip maximizer had best not take that risk after all (leaving it open to defeat by a faster-evolving cluster of AIs). This, by reductio ad absurdum, is why I don’t believe in smart expected paperclip maximizers.
Humans aren’t factorized this way, whether they can’t is a separate question. It’s not surprising that evolution’s design isn’t that neat, so the fact that humans don’t have this property is only weak evidence about the possibility of designing systems that do have this property.
I doubt it. I neither believe that people like Jürgen Schmidhuber are a risk, apart from a very abstract possibility.
The reason is that they are unable to show off some applicable progress on a par with IBM Watson or Siri. And in the case that they claim that their work relies on a single mathematical breakthrough, I doubt that it would be justified even in principle to be confident in that prediction.
In short, either their work is incrementally useful or is based on wild speculations about the possible discovery of unknown unknowns.
The real risks in my opinion are 1) that together they make many independent discoveries and someone builds something out of it 2) that a huge company like IBM, or a military project, builds something 3) the abstract possibility that some partly related field like neuroscience, or an unrelated field, provides the necessary insight to put two and two together.
Do you mean that intelligence is fundamentally interwoven with complex goals?
Do you mean that there is no point at which exploitation is favored over exploration?
I am not sure what you mean, could you elaborate? Do you mean something along the lines of what Ben Goertzel says in the following quote:
You further wrote:
What is your best guess at why people associated with SI are worried about AI risk?
If you would have to fix the arguments for the proponents of AI-risk, what would be the strongest argument in favor of it? Also, do you expect there to be anything that could possible change your mind about the topic and become worried?
Essentially, yes. I think that defining an arbitrary entity’s “goals” is not obviously possible, unless one simply accepts the trivial definition of “its goals are whatever it winds up causing”; I think intelligence is fundamentally interwoven with causing complex effects.
I mean that there is no point at which exploitation is favored exclusively over exploration.
I’m 20 years old—I don’t have any kids yet. If I did, I might very well feel differently. What I do mean is that I believe it to be culturally pretentious, and even morally wrong (according to my personal system of morals), to assert that it is better to hold back technological progress if necessary to preserve the human status quo, rather than allow ourselves to evolve into and ultimately be replaced by a superior civilization. I have the utmost faith in Nature to ensure that eventually, everything keeps getting better on average, even if there are occasional dips due to, e.g., wars; but if we can make the transition to a machine civilization smooth and gradual, I hope there won’t even have to be a war (a la Hugo de Garis).
Well, the trivial response is to say “that’s why they’re associated with SI.” But I assume that’s not how you meant the question. There are a number of reasons to become worried about AI risk. We see AI disasters in science fiction all the time. Eliezer makes pretty good arguments for AI disasters. People observe that a lot of smart folks are worried about AI risk, and it seems to be part of the correct contrarian cluster. But most of all, I think it is a combination of fear of the unknown and implicit beliefs about the meaning and value of the concept “human”.
In my opinion, the strongest argument in favor of AI-risk is the existence of highly intelligent but highly deranged individuals, such as the Unabomber. If mental illness is a natural attractor in mind-space, we might be in trouble.
Naturally. I was somewhat worried about AI-risk before I started studying and thinking about intelligence in depth. It is entirely possible that my feelings about AI-risk will follow a Wundt curve, and that once I learn even more about the nature of intelligence, I will realize we are all doomed for one reason or another. Needless to say, I don’t expect this, but you never know what you might not know.
The laws of physics don’t care. What process do you think explains the fact that you have this belief? If the truth of a belief isn’t what causes you to have it, having that belief is not evidence for its truth.
I’m afraid it was no mistake that I used the word “faith”!
This belief does not appear to conflict with the truth (or at least that’s a separate debate) but it is also difficult to find truthful support for it. Sure, I can wave my hands about complexity and entropy and how information can’t be destroyed but only created, but I’ll totally admit that this does not logically translate into “life will be good in the future.”
The best argument I can give goes as follows. For the sake of discussion, at least, let’s assume MWI. Then there is some population of alternate futures. Now let’s assume that the only stable equilibria are entirely valueless state ensembles such as the heat death of the universe. With me so far? OK, now here’s the first big leap: let’s say that our quantification of value, from state ensembles to the nonnegative reals, can be approximated by a continuous function. Therefore, by application of Conley’s theorem, the value trajectories of alternate futures fall into one of two categories: those which asymptotically approach 0, and those which asymptotically approach infinity. The second big leap involves disregarding those alternate futures which approach zero. Not only will you and I die in those futures, but we won’t even be remembered; none of our actions or words will be observed beyond a finite time horizon along those trajectories. So I conclude that I should behave as if the only trajectories are those which asymptotically approach infinity.
It seems to me like you assume that you have no agency in pushing the value trajectories of alternate futures towards infinity rather than zero, and I don’t see why.
Is this a variant of quantum suicide, with “suicide” part replaced by “dead and forgotten in long run, whatever the cause”?
A uniform category of “good” or “positive” fails to compare its elements. Just how good are different AIs, compared to each other? Can one be much better than another? There is an opportunity cost for settling for comparatively worse AIs. Given the astronomical scale of consequences, any difference may be quite significant, which would make it an important problem to ensure the creation of one of the better possible AIs, rather than an AI that technology would stumble on by default.
Human intelligence has been successfully applied to achieve many goals which were not applicable or trainable in the environment of anscestral adaption such as designing, building, and operating cars, playing chess, sending people to the moon and back, and programming computers. It is clear the goals and intelligence can be factorized as a matter of simple observation.
Are those goals instrumental, or terminal?
With humans being such imperfect consequentalists, there is not always a clear distinction between instrumental and terminal goals. Much that we consider fun to do also furthers higher goals.
But even if you assume all the goals are instrumental, my point still stands. That a goal was adopted because it furthers a higher level goal doesn’t change the fact that it could be successfully plugged into human intelligence.
Sure. But the central question is “are higher level goals arbitrary?”, and while “Well, the subgoals we use to further those higher level goals are arbitrary along a few dimensions” is a start to answering that question, it is far from an end.
Wrong. The central question is “Can arbitrary goals be successfully plugged into a general goal achieving system?”, and I have shown examples where it can.
As far as I can tell, that’s what you’re discussing, and so it sounds like you agree with him. Did I misread disagreement into this comment, or what am I missing here?
The section you quote allows for the possibility that an AI could be given a “compulsion” to optimize for paperclips, which it would eventually shrug off, whereas I am confident that an AI could be given a utility function that would make it actually optimize for paperclips.
Okay; but the examples you gave seem to me to be more similar to compulsions than to utility functions. A person can care a lot about cars, and cars can become a major part of human society, but they’re not the point of human society- if they stop serving their purposes they’ll go the way of the horse and buggy. I’m not sure I can express the meaning I’m trying to convey cleanly using that terminology, so maybe I ought to restart.
My model of davidad’s view is that part of general intelligence, as opposed to narrow intelligence, is varied and complex goals. We could make a narrow AI which only cared about the number of paperclips in the universe, but in order to make an intelligence that’s general we need to make it also care about the future, planning, existential risk, and so on.
And so you might get a vibrant interstellar civilization of synthetic intelligences- that happens to worship paperclips, and uses them for currency and religious purposes- rather than a dead world with nothing but peculiarly bent metal.
I would have liked to use examples of plugging in clearly terminal values to a general goal achieving system. But the only current or historical general goal achieving systems are humans, and it is notoriously difficult to figure out what humans’ terminal values are.
I am not claiming that you could give an AGI an arbitrary goal system that suppresses the “Basic AI Drives”, but that those drives will be effective instrumental values, not lost purposes, and while a paperclip maximizing AGI will have sub goals such as controlling resources and improving its ability to predict the future, the achievement of those goals will help it to actually produce paperclips.
It sounds like we agree: paperclips could be a genuine terminal value for AGIs, but a dead future doesn’t seem all that likely from AGIs (though it might be likely from AIs in general).
What? A paperclip AGI with first mover advantage would self-improve beyond the point where cooperating with humans has any instrumental value, become a singleton, and tile the universe with paperclips.
Oh, I agree that humans die in such a scenario, but I don’t think the ‘tile the universe’ part counts as “dead” if the AGI has AI drives.
Why would you believe that? Evolution was more than capable of building an intelligence that optimized for whatever goals it needed, notable reproduction and personal survival. Granted its version was imperfect, since humans have enough conflicting goals that we can sometimes make moves that are objectively bad for the perpetuation of our gametes, not to mention the obvious failure cases like asexuals. That said, evolution has fat fingers. We can do better, and any AI’s we build will be able to do even better.
I promise you that if the production of paperclips was a survival trait in the ancestral environment, above all else, we would all be paperclip maximizers. We would consider paperclips profound and important, and we would be loathe to remove the desire to make paperclips—any more than we would be inclined now to pare out our own sex drive and self-preservation instinct.
EDIT: I do think the scenario of simply immediately turning everything into paperclips is naive. A superintelligence would have an enormous incentive to devote its resources to research and development for optimizing its goals as rapidly as possible, and would probably spend a lot of time simply thinking before actually embarking on a large-scale manufacture of paperclips. That’s still not good for us, though, because even in that case, we’re clearly a lot more useful to it as solid-state paperclip R&D labs than as human beings.
Upvoted for honesty.
Honesty seems like a rather low bar for this community, don’t you think? Maybe you mean something more like personal insight? (Presumably most people who don’t admit their own biases aren’t being dishonest, but just don’t realize or aren’t convinced that they are biased.)
What’s your current estimate (or probability distribution) for how much computational power would be needed to run the C. elegans simulation? Once your project succeeds, can we derive an upper bound for a human upload just by multiplying by the ratio of neurons and/or synapses, or would that not be valid because human neurons are much more complicated, or for some other reason?
Also, how are you planning to turn the data you gather using optogenetics into computational models of neurons? (Or do you already have the models and just need to plug in parameters based on the data?)
I think the simulation environment should run in real-time on a laptop. If we’re lucky, it might run in real-time on an iPhone. If we’re unlucky, it might run in real-time on a cluster of a few servers. In any case, I expect the graphics and physics to require much (>5x) more computational power than the C. elegans mind itself (though of course the content of the mind-code will be much more interesting and difficult to create).
Unfortunately, C. elegans is different enough from humans in so many different ways that everyone who currently says that uploading is hard would be able to tweak their arguments slightly to adapt to my success. Penrose and Hammeroff can say that only mammal brains do quantum computation. Sejnowski can say that synaptic vesicle release is important only in vertebrates. PZ Myers can say “you haven’t modeled learning or development and you had the benefit of a connectome, and worm neurons don’t even have spikes; this is cute, but would never scale.”
That said, if you’re already inclined to agree with my point of view—essentially, that uploading is not that hard—then my success in creating the first upload would certainly make my point of view that much more credible, and it would supply hard data which I would claim can be extrapolated to an upper bound for humans, at least within an order of magnitude or two.
I will absolutely buy that screen saver.
I’d heard it said that the best human simulators required something like one laptop per neuron—I take it this is because cerebral cortex neurons are more demanding than any C. elegans neuron?
Right—some cortical neurons have 10,000 synapses, while the entire C. elegans nervous system has less than 8,000.
Quick summary: The reason it’s great to be a neuroscientist today is optogenetics.
Good talk, David! And a good Q&A.
Thanks! Richard Feynman said it best: “It is very easy to answer many of these fundamental biological questions; you just look at the thing!”
BTW, what’s the paper about putting the zebrafish in the Matrix?
“Brain-wide neural dynamics at single-cell resolution during rapid motor adaptation in larval zebrafish” by Ahrens et al., accepted to Nature but not yet published. I’ve taken the liberty of uploading it to scribd: Paper Supplementary Info
Wei_Dai pointed out that it would be good to follow up on this as a comment. I might be up for doing a proper retrospective on my Nemaload project at some point, but not today. However, meanwhile, I will link here to some recent comments I made on a recent 10-year retrospective of nematode uploading as a whole.
I would be interested to know how you plan to demonstrate to the world that the experiment was a success.
If the goal was to upload a cat, for example, you could demonstrate that it runs under the simulated couch when a simulated vacuum cleaner comes around and that it runs into the simulated kitchen when it hears a simulated electric can opener. I am curious what high-level behaviors are considered typical of C. elegans.
He mentions that in the video—there’s a list of stereotypical behaviors, when X happens the worm turns around, etc. Having the simulation behave the same way would be a good sign.
He also mentions zapping neurons to see what happens without them; though I’m less sure if that is to be used as an additional test of model validity, or just as a way to figure out which neurons do what.
I imagine that a video of the uploaded worm moving in some artificial environment, reacting like a real worm, could be impressive. Even more impressive would be to test some genetic modifications on the virtual worm, which would also work on the real worm; e.g. an ability to eat some kind of waste we need to remove from environment. (I assume that with enough computer power, testing on virtual worms could be faster than testing real worms.)
Last but not least, when humans get uploaded, virtual worms could be used as a food source. :D
I won’t subject you to cute cat stories, but real cats do weird things as well as stereotypical things.
I’m really happy you are working on this.
Thanks! :)
Possibly a stupid question, but what is the relationship between your work and the OpenWorm project?
I would also like to know the answer to this question.
Hi David, thanks much for the post. Can you comment on the rationales behind moving away (because of funding?) from C elegans? In particular, I’m interested in your thoughts on these points:
Problem difficulty. E.g. Problem is much harder than expected
Opportunity cost. E.g. Better alternatives to achieve your goals exist (whatever these goals are, I would love to know them as well!)
I’m asking this out of pure curiosity, as a software engineer who just started diving into C elegans simulations.