Fake Fake Utility Functions
Followup to: Most of my posts over the last month...
Every now and then, you run across someone who has discovered the One Great Moral Principle, of which all other values are a mere derivative consequence.
I run across more of these people than you do. Only in my case, it’s people who know the amazingly simple utility function that is all you need to program into an artificial superintelligence and then everything will turn out fine...
It’s incredible how one little issue can require so much prerequisite material. My original schedule called for “Fake Utility Functions” to follow “Fake Justification” on Oct 31.
Talk about your planning fallacy. I’ve been planning to post on this topic in “just a few days” for the past month. A fun little demonstration of underestimated inferential distances.
You see, before I wrote this post, it occurred to me that if I wanted to properly explain the problem of fake utility functions, it would be helpful to illustrate a mistake about what a simple optimization criterion implied. The strongest real-world example I knew was the Tragedy of Group Selectionism. At first I thought I’d mention it in passing, within “Fake Utility Functions”, but I decided the Tragedy of Group Selectionism was a long enough story that it needed its own blog post...
So I started to write “The Tragedy of Group Selectionism”. A few hours later, I noticed that I hadn’t said anything about group selectionism yet. I’d been too busy introducing basic evolutionary concepts. Select all the introductory stuff, cut, Compose New Post, paste, title… “An Alien God”. Then keep writing until the “Alien God” post gets too long, and start taking separate subjects out into their own posts: “The Wonder of Evolution”, “Evolutions Are Stupid”, and at this point it became clear that, since I was planning to say a few words on evolution anyway, that was the time. Besides, a basic familiarity with evolution would help to shake people loose of their human assumptions when it came to visualizing nonhuman optimization processes.
So, finally I posted “The Tragedy of Group Selectionism”. Now I was ready to write “Fake Utility Functions”, right? The post that was supposed to come immediately afterward? So I thought, but each time I tried to write the post, I ended up recursing on a prerequisite post instead. Such as “Fake Selfishness”, “Fake Morality”, and “Fake Optimization Criteria”.
When I got to “Fake Optimization Criteria”, I really thought I could do “Fake Utility Functions” the next day. But then it occurred to me that I’d never explained why a simple utility function wouldn’t be enough. We are a thousand shards of desire, as I said in “Thou Art Godshatter”. Only that first required discussing “Evolutionary Psychology”, which required explaining that human minds are “Adaptation-Executers, not Fitness-Maximizers”, plus the difference between “Protein Reinforcement and DNA Consequentialism”.
Furthermore, I’d never really explained the difference between “Terminal Values and Instrumental Values”, without which I could hardly talk about utility functions.
Surely now I was ready? Yet I thought about conversations I’d had over the years, and how people seem to think a simple instruction like “Get my mother out of that burning building!” contains all the motivations that shape a human plan to rescue her, so I thought that first I’d do “The Hidden Complexity of Wishes”. But, really, the hidden complexity of planning, and all the special cases needed to patch the genie’s wish, was part of the general problem of recording outputs without absorbing the process that generates the outputs—as I explained in “Artificial Addition” and “Truly Part Of You”. You don’t want to keep the local goal description and discard the nonlocal utility function: “Leaky Generalizations” and “Lost Purposes”.
Plus it occurred to me that evolution itself made an interesting genie, so before all that, came “Conjuring An Evolution To Serve You”.
One kind of lost purpose is artificial pleasure, and “happiness” is one of the Fake Utility Functions I run into more often: “Not for the Sake of Happiness (Alone)”. Similarly, it was worth taking the time to establish that fitness is not always your friend (“Evolving to Extinction”) and that not everything in the universe is subject to significant selection pressures (“No Evolutions for Corporations or Nanodevices”), to avoid the Fake Utility Function of “genetic fitness”.
Right after “Lost Purposes” seemed like a good time to point out the deep link between keeping track of your original goal and keeping track of your original question: “Purpose and Pragmatism”.
Into the home stretch! No, wait, this would be a good time to discuss “Affective Death Spirals”, since that’s one of the main things that goes wrong when someone discovers The One True Valuable Thingy—they keep finding nicer and nicer things to say about it. Well, you can’t discuss affective death spirals unless you first discuss “The Affect Heuristic”, but I’d been meaning to do that for a while anyway. “Evaluability” illustrates the affect heuristic and leads to an important point about “Unbounded Scales and Futurism”. The second key to affective death spirals is “The Halo Effect”, which we can see illustrated in “Superhero Bias” and “Mere Messiahs”. Then it’s on to affective death spirals and how to “Resist the Happy Death Spiral” and “Uncritical Supercriticality”.
A bonus irony is that “Fake Utility Functions” isn’t a grand climax. It’s just one of many Less Wrong posts relevant to my AI work, with plenty more scheduled. This particular post just turned out to require just a little more prerequisite material which—I thought on each occasion—I would have to write anyway, sooner or later.
And that’s why blogging is difficult, and why it is necessary, at least for me. I would have been doomed, yea, utterly doomed, if I’d tried to write all this as one publication rather than as a series of blog posts. One month is nothing for this much material.
But now, it’s done! Now, after only slightly more than an extra month of prerequisite material, I can do the blog post originally scheduled for November 1st!
Except...
Now that I think about it...
This post is pretty long already, right?
So I’ll do the real “Fake Utility Functions” tomorrow.
- GAZP vs. GLUT by 7 Apr 2008 1:51 UTC; 77 points) (
- Fake Utility Functions by 6 Dec 2007 16:55 UTC; 69 points) (
- The Psychological Unity of Humankind by 24 Jun 2008 7:12 UTC; 61 points) (
- Not for the Sake of Pleasure Alone by 11 Jun 2011 23:21 UTC; 50 points) (
- ...And Say No More Of It by 9 Feb 2009 0:15 UTC; 43 points) (
- What I Think, If Not Why by 11 Dec 2008 17:41 UTC; 41 points) (
- Not for the Sake of Selfishness Alone by 2 Jul 2011 17:37 UTC; 34 points) (
- Heading Toward Morality by 20 Jun 2008 8:08 UTC; 27 points) (
- Setting Up Metaethics by 28 Jul 2008 2:25 UTC; 27 points) (
- Help us Optimize the Contents of the Sequences eBook by 19 Sep 2013 4:31 UTC; 18 points) (
- I think I’ve found the source of what’s been bugging me about “Friendly AI” by 10 Jun 2012 14:06 UTC; 15 points) (
- The Outside Critics of Effective Altruism by 5 Jan 2015 18:37 UTC; 12 points) (EA Forum;
- 27 Jun 2011 17:44 UTC; 10 points) 's comment on asking an AI to make itself friendly by (
- 15 Jan 2016 4:41 UTC; 7 points) 's comment on Open Thread, January 11-17, 2016 by (
- 9 Feb 2013 18:56 UTC; 6 points) 's comment on How to offend a rationalist (who hasn’t thought about it yet): a life lesson by (
- 11 Mar 2011 5:31 UTC; 5 points) 's comment on Why not just write failsafe rules into the superintelligent machine? by (
- 4 Oct 2013 18:53 UTC; 4 points) 's comment on How to Become a 1000 Year Old Vampire by (
- 6 Jan 2015 17:32 UTC; 2 points) 's comment on The Outside Critics of Effective Altruism by (EA Forum;
- 31 Dec 2011 6:35 UTC; 2 points) 's comment on Stupid Questions Open Thread by (
- 15 Jun 2010 11:16 UTC; 2 points) 's comment on Open Thread June 2010, Part 3 by (
- 24 Jul 2011 4:23 UTC; 0 points) 's comment on Rewriting the sequences? by (
Brilliant! I like how this post is an annotated bibliography to many of your previous posts. This is really handy for this website. The one gripe I have with this website is how inconvenient it is to read old posts. It would be great if one could display articles by an author in one place.
Alternatively, it would be a great idea if you wrote something like this for ALL your previous posts, and for all future posts once a month or so.
This post made me think that I’m reading your writings in a very inefficient manner if I do not follow the order they’re written in, yet I cannot find a simple way to do so (without opening a few dozen tabs in firefox and gritting my teeth through the inevitable subsequent memory leak!). Am I missing something?
All articles (posts)
I wonder if we’ll see any discussion of utilitarianism as the One Great Moral Principle?
I’ve often wondered why you didn’t make this material into a book or something, because you clearly could have given your expansive knowledge. The fact that you are doing it in this mostly non-profit format impresses me, although it is no surprise given the honesty and knowledge in your writing, and your stated desire to “save the world.”
Anyway, thanks.
This post put a big smile on my face. Thanks Eliezer.
“I’ve often wondered why you didn’t make this material into a book or something”
Here’s the answer: Why I’m blooking
“This post made me think that I’m reading your writings in a very inefficient manner if I do not follow the order they’re written in, yet I cannot find a simple way to do so”
Well one simple way to do it, is to subscribe to the RSS feed and then in the reader, search “Eliezer Yudkowsky”, for example, and there you go. Every post in OB by Eliezer in the right order.
hmm… well, it was an interesting month but I’m not sure I didn’t enjoy this blog more when it was a blog just about, well, overcoming bias. Eliezer, perhaps you should have your own blog...
BTW I saw you on the TV the other night, talking about the singularity.
It’s always odd, isn’t it, seeing a person for the first time whom you have previously known only in writing. As always, you didn’t look/sound at all like I expected.
Which led me to an interesting line of thought: why does reading someone’s writing produce any expectation at all of what the person looks/sounds like? It’s a bias. Is there a halo effect: we expect intelligent writers to look tall and beautiful? Or is there a einstein-bias: we expect intelligent writers to be wild-looking with messy hair, talking quickly?
Anyway, there you were on the screen. Looking different. “Quick!”, I yelled to my wife, who was in the kitchen, “quick, come here! It’s Eliezer Yudkowsky!”
hmmm...perhaps I need to get out more.