OpenWorm and differential technological development
According to Robin Hanson’s arguments in this blog post, we want to promote research in to cell modeling technology (ideally at the expense of research in to faster computer hardware). That would mean funding this kickstarter, which is ending in 11 hours (it may still succeed; there are a few tricks for pushing borderline kickstarters through). I already pledged $250; I’m not sure if I should pledge significantly more on the strength of one Hanson blog post. Thoughts from anyone? (I also encourage other folks to pledge! Maybe we can name neurons after characters in HPMOR or something. EDIT: Or maybe funding OpenWorm is a bad idea; see this link.)
I’m also curious on what people think about the efficiency of trying to pursue differential technological development directly this way vs funding MIRI/FHI. I haven’t read the entire conversation referenced here, but this bit from the blog post sounds correct-ish to me:
People doing philosophical work to try to reduce existential risk are largely wasting their time. Tyler doesn’t think it’s a serious effort, though it may be good publicity for something that will pay off later. A serious effort looks more like the parts of the US government that trained people to infiltrate the post-collapse Soviet Union and then locate and neutralize nuclear weapons. There was also a serious effort by the people who set up hotlines between leaders to be used to quickly communicate about nuclear attacks (e.g., to help quickly convince a leader in country A that a fishy object on their radar isn’t an incoming nuclear attack).
Edit: For some reason I forgot about this previous discussion on this topic, which makes the case for funding OpenWorm look less clear-cut.
- 19 May 2014 4:17 UTC; 2 points) 's comment on Open Thread, May 12 − 18, 2014 by (
Thanks for pledging and encouraging others to pledge! Full disclosure: I’m the coordinator for the project. I’ve been having a look through the discussions on your references and I’d offer the following thoughts:
I think Hanson’s three part break down (Computing power, brain scanning, cell modeling) is useful and I agree that cell modeling is an important research investment that has not had enough focus, either academically or industrially. Better cell models is one of the technological advances that OpenWorm helps to address due to its approach to model a complete organism with such few cells.
I would add that none of these discussions seem to pick up on the additional benefits of cell modeling outside of the context of brain emulation, which include advances in complexity science in general, increased potential for tissue regeneration and repair, and better diagnostics and therapies for diseases. Remember, all living things have cells, so advanced cell modeling could give us a debugger and an editor for biology unlike anything we’ve ever seen.
In terms of funding open science via crowd funding as a differential technological development strategy, I would also point out that the results are held in a public commons (GitHub in our case) and this transparency and open access may be an important factor. Work like this is likely going to be done at some point, but if it isn’t publicly funded then it is likely to be privately funded and also privately held, and may add to asymmetrical control over these technologies. Personally, I prefer power to be distributed as a bulwark against tyranny. The more of these technological advances are out in the open, I think, the less likely the power of them will be concentrated in the hands of the few and used improperly.
Hi Stephen, thanks for chiming in! Did you take a look at this that I linked to at the bottom?
I did—could you summarize what parts of this you take away to discourage you specifically? There’s quite a few things you could be resonating with there.
Sure. So to give you some context, WBE is an acronym for Whole Brain Emulation. You can read this report by the Future of Humanity Institute at Oxford where they detail the technological development that’d be required to achieve this. If it did end up getting achieved, the consequences for society could be enormous. An emulated brain could do the same work as a software developer, researcher, etc., but the emulated brain could be run at a much faster subjective rate (say, doing a week’s worth of subjective thinking in the space of 5 minutes) and for less money (given continued decreases in the cost of computer hardware). Robin Hanson, the GMU economics professor who wrote the “Bad Emulation Advance” blog post, is very interested in ems… here is a presentation where he fleshes out the possibilities of an emulation-filled future in some detail. It’s not a terrible future, but it’s not a terrifically bright one either.
One future that probably would be pretty terrible is if we had an extremely intelligent artificial intelligence that was built on some technologies inspired by the human brain but was not a high-resolution exact copy of any living human (“neuromorphic AI”), and was not carefully constructed to work towards achieving human values. You can read this for a short summary of why this would likely be terrible, or this for a longer, more fleshed-out argument (with entertaining background info).
So this is why we want to tread carefully. It’s suspected that neuromorphic AI is harder to construct in a robust, provably safe way, and given the dangers of haphazardly constructed superintelligences, it seems like we’d rather see more mathematically pure AI methods advance, if any at all. That’s what this link is in reference to.
I know that was a pretty high density of crazy ideas in a pretty short period of time, and there are some leaps of reasoning that I left out… let me know if you’ve got any thoughts or questions. To a futurist like me, things like better diagnostics and therapies for diseases, while interesting and exciting, are small potatoes in the grand scheme of things. When there’s the possibility of society being completely transformed by a technology, that’s when I start to pay attention. (Hey, it’s happened plenty of times before. Imagine explaining the modern world to a Cro-Magnon.) So the main way I’m seeing your project is in terms of its incremental advances along various technological dimensions that could lead to societal transformation. Hopefully that makes some sense.
A human is very massively sub-mankind level intelligent, and some rough approximation running at sub-realtime speeds with several orders of magnitude higher daily sustenance cost even more so.
Granted, it’s a great plot device once you give it superpowers, and so there have been many high profile movies concerning such scenarios, and you see worlds destroyed by AI on the big screen. And your internal probability evaluator—evolved before CGI—uses the frequency of scenarios you seen with your own eyes.
No disagreement there.
Reversed stupidity isn’t intelligence.
Then why you’re discussing sudden super-intelligence? The faster cell simulation technologies advance, the weaker is the hardware they’ll run on.
Direct stupidity is not intelligence either. The a-priori likelihood of an arbitrary made up prediction being anywhere near correct is pretty damn low. If I tell you that 7 932 384 626 is some lottery number from yesterday, you may believe it, but once you see that it’s the digits of pi fairly close to the start, your credence should drop. A lot.
If hardware growth strictly followed Moore’s Law and CPUs (or GPUs, etc.) were completely general-purpose, this would be true. But, if cell simulation became a dominant application for computing hardware, one could imagine instruction set extensions or even entire architecture changes designed around it. Obviously, it would also take some time for software to take advantage of hardware change.
Well, first it has to become dominant enough (for which it’d need to be common enough, for which it needs to be useful enough—used for what?), then the hardware specialization is not easy either, and on top of that specialized hardware locks the designs in (prevents easy modification and optimization). Especially if we’re speaking of specializing beyond how GPUs are specialized for parallel floating point computations.
I’m afraid you’re going to have to explain yourself better if you want me to respond… I confess I don’t see clearly how what you’re saying pertains to our argument.
The point is, cell simulation won’t yield this stupid AI movie plot threat that you guys are concerned about. This is because it doesn’t result in sudden superintelligence, but a very gradual transition.
And in so much as there’s some residual possibility that it could, this possibility is lessened by working on it earlier.
I’m puzzled why you focus on the AI movie plot threat when discussing any AI related technology, but my suspicion is that it’s because it is a movie plot threat.
edit: as for the “robust provably safe AI”, as a subset of “safe” an AI must be able to look at—say—an electronic circuit diagram (or an even lower level representation) and tell if said circuit diagram implements a tortured sentient being. You’d need neurobiology to merely define what’s bad. The problem is that “robust provably safe” is nebulous enough that you can’t link it to anything concrete.
You seem awfully confident. I agree that you’re likely right but I think it’s hard to know for sure and most people who speak on this issue are too confident (including you and both EY/RH in their AI foom debate).
Just to clarify: so you mostly agree with the “Bad Emulation Advance” blog post?
It’s not clear to me that a gradual transition completely defeats the argument against neuromorphic AI. If neuromorphic AI is less predictable (to put things poetically, “harder to wield”) than AI constructed so that it provably satisfies certain properties, then you can imagine humanity wielding a bigger and bigger weapon that’s hard to control. How long do you think the world would last if everyone had a pistol that fired tactical nuclear weapons? What if the pistol has a one in six chance of firing in a random direction?
Want to point to a specific case where I did that?
That’s an interesting point. I think it probably makes sense to think of “robust provably safe” as being a continuous parameter. You’ve got your module that determines what’s ethical and what isn’t and you’ve got your module that makes predictions and you’ve got your module that generates plans. The probability of your AI being “safe” is the product of the probabilities of all your modules being “safe”. If a neuromorphic AI self-modifies in a less predictable way that seems like a lose, keeping the ethics module constant.
There’s a false equivalence, similar to what’d happen if I were predicting “the lottery will not roll 12345134” and someone else predicting “the lottery will roll 12345134″. Predicting some sudden change in a growth curve along with the cause of such change, that’s a guess into a large space of possibilities; if such guess is equally unsupported with it’s negation, it’s extremely unlikely and negation is much more likely.
That strikes me as a rather silly way to look at it. The future generations of biological humans are not predictable or controllable either.
The point is that you need bottom-up understanding of, for example, suffering, to be able to even begin working at an “ethics module” which recognizes suffering as bad. (We get away without conscious understanding of such only because we can feel it ourselves and thus implicitly embody a definition of such). On the road to that, you obviously have cell simulation and other neurobiology.
The broader picture is that with zero clue as to the technical process of actually building the “ethics module”, when you look at, say, openworm, and it doesn’t seem like it helps build an ethical module, that’s not representative in any way as to whenever it would or would not help, but only representative of it being a concrete and specific advance and the “ethics module” being too far off and nebulous.
This sounds to me like an argument over priors; I’ll tap out at this point.
Well, do you trust humans with humanity’s future? I’m not sure I do.
Maybe.
If you just make stuff up, the argument will be about priors. Observe: there’s a teapot in the asteroid belt.
Well yeah and I could trivially “defeat” any argument of yours by declaring my prior for it to be very low. My priors for the future are broadly distributed because the world we are in would seem very weird to a hunter-gatherer, so I think it’s likely that the world of 6,000 years from now will seem very weird to us. Heck, World War II would probably sound pretty fantastic if you described it to Columbus.
I’ll let you have the last word :)
Priors can’t go arbitrarily high before the sum over incompatible propositions becomes greater than 1.
If we were to copy your brain a trillion times over and ask it to give your “broadly distributed” priors for various mutually incompatible and very specific propositions, the result should sum to 1 (or less than 1 if its non exhaustive), which means that most propositions should receive very, very low priors. I strongly suspect that it wouldn’t be even remotely the case—you’ll be given a proposition, then you can’t be sure it’s wrong “because the world of future would look strange”, and so you give it some prior heavily biased towards 0.5 , and then over all the propositions, the summ will be very huge .
When you’re making very specific stuff up about what the world of 6000 years from now will look like, it’s necessarily quite unlikely that you’re right and quite likely that you’re wrong, precisely because that future would seem strange. That the future is unpredictable works against specific visions of the future.
Are there alternatives..?
PM’s precise point is also raised in the Sequences.
Yes, and John’s point is that just because it’s possible to arrive at that conclusion by a wrong method doesn’t actually mean it’s wrong.
Re: Tyler’s comment on “philosophical work” on x-risk reduction being largely a waste of time.
I’m not sure what he means by “philosophical work” here, but if he means “broad strategic work of the type sometimes done at FHI and MIRI,” well, the whole point of that work is to help answer questions exactly like the one you’re conflicted about here: whether OpenWorm is good or bad differential technological development. It’s precisely because of such work at FHI and MIRI that we have the concept of “differential technological development” in the first place, and that we have a collection of arguments for and against different kinds of differential technological development, even if the answers aren’t yet clear.
Before one country invades another, or cuts or supplies $1B in funding for some project, it would be nice to know whether doing so would be good or bad. That’s why Teller et al. studied the question of whether an atom bomb could ignite the atmosphere, and it’s why FHI and MIRI are doing much of the research we do.
I agree with Luke. It’s funny that Tyler, a pundit, says that pundits are useless for reducing existential risk.
Funding concrete projects is often relatively easy. People can see it and get excited about it. Asking to fund higher-level research is harder.
Concrete work is what governments do. OpenWorm is not going to compete with Obama’s $3 billion BRAIN Initiative. Pundits make these issues political, raise awareness, and thereby lead to huge amounts of funding down the road.
As for the question of whether to favor WBE: I’d be nervous about it. It could also accelerate non-WBE AI through spillover technology, enhancing interest in general in these topics, etc. I don’t have a clear opinion, but the fact that the question is so hard suggests to me that this isn’t the most cost-effective place to push. There are many other donation targets where the benefits clearly outweigh the risks.
Even leaving the definitions of “good” and “bad” aside, can you know for a sufficiently long time horizon?
Five-year consequences we can get a reasonable forecast for. Hundred-year consequences, I submit no one has the slightest clue about.
I like Brian Tomasik’s case for trying to increase reflectiveness in the general population. I would expect that increasing reflectiveness in the general population would, if anything, be self-reinforcing; I’d be a bit surprised to be a “reflectiveness backlash” where people would decide they wanted to start being very unreflective as a result of reflectiveness being promoted too strongly. So increasing reflectiveness would seem to me to be an intervention that is pretty likely to steer humanity in a good direction, or at the very least, make it so that wherever we are in 10 years, we’ll have a better distribution of outcomes in front of us and factors to lean on.
My guess is there are other interventions that also fall in to this category, e.g. improving the quality of political discourse and generally increasing peoples’ rationality. Basically things that would prepare society to better deal with a broad range of tough situations that we might face 100 years out.
I don’t know about that. First, I’m automatically suspicious of arguments which go “General population should be more like me!” and, truth be told, intellectuals tend to be rather fond of such arguments.
Second, reflectiveness is like narcissism, it’s just that instead of focusing on your body you focus on your mind instead. I am not convinced it falls into the “the more the better” category.
Third, the suggested ways of going about it are all very handwavy and wishy-washy.
This is tautology—“we will improve things by improving things”. Try tabooing words like “better” or “improve”. What specific, concrete, practical changes would you make to the political discourse? Why do you think these changes will turn out to have positive consequences in a hundred years?
Well that would be my prior.
That sounds like an argument from analogy to me. You’re not describing any causal pathway by which reflectiveness makes the world a worst case. You’re saying “reflectiveness looks vaguely like this other thing [which is actually totally different], and people to seem to think that thing is bad, therefore reflectiveness is bad”.
I’d like to hear your answers to those questions before I answer them, if you don’t mind.
You link leads to a (rather obvious) observation that smarter-on-the-average populations do better, economically, than dumber-on-the-average populations. Though, as the PRC example shows, sociopolitical structures do matter.
However being reflective is not at all the same thing as being smart. I’m viewing statements “People should be more reflective” by someone who’s clearly more reflective than the average as epistemically suspect. I don’t see how your link addresses this issue.
That’s not difficult. For example, an increase in reflectiveness is generally accompanied by a decrease in decisiveness. Analysis paralysis is a common problem for highly reflective people. For another example, reflectiveness leads your focus inside, to yourself, and it’s not hard to come up with situations where you should be thinking about the outside world more and about yourself less. Navel gazing is a highly reflective but rarely productive activity.
This looks like an ad hominem argument to me.
Thanks for making concrete arguments. Analysis paralysis is a problem I would like to see more of on a societal level :) But maybe things will get tricky if the least reflective people end up moving first and making society’s decisions. So maybe what we want to aim for is increasing the reflectiveness of the least reflective people who hold power.
I think it makes sense to disentangle the self-focus that you mention and treat it as an orthogonal vector. I’m in favor of people reflecting about important stuff but not unimportant stuff… I hope that clarifies my goals. Insofar as there’s a tradeoff where getting people to reflect about important stuff also means they will waste time reflecting about unimportant stuff, I’m not sure how best to manage that tradeoff.
I’m not finding this conversation especially productive, so I’ll let you have the last word if you want it.
If it’s done, it’s done.
The latest edition of The Economist (May 24th, 2014) includes a story on OpenWorm. Some excerpts: