What do you think is the maximum price you’d be willing to pay?
Tuxedage
Yes, unless I’m playing a particularly interesting AI like Eliezer Yudkowsky or something. Most AI games are boring.
If anyone wants to, I’d totally be willing to sit in a room for two-and-half hours while someone tries to convince me to give up logs, so long as you pay the same fee as the ordinary AI Box Experiment. :)
I’m not sure that’s good advice. 80,000 hours has given pretty good arguments against just “doing what you’re passionate about”.
Passion grows from appropriately challenging work. The most consistent predictor of job satisfaction is mentally challenging work (2). Equating passion with job satisfaction, this means that we can become passionate about many jobs, providing they involve sufficient mental challenge. The requirements for mentally challenging work, like autonomy, feedback and variety in the work, are similar to those required to develop flow. This suggests that a similar conclusion will hold if we believe that being passionate is closely connected with the ability to enter states of flow. If, however, you don’t think flow and job satisfaction are the same thing as passion, then you can still agree that…
There are better targets to aim for. We’re not only bad at predicting what will make us happy, but more easily detectable predictors of job satisfaction exist (autonomy, feedback, variety, making a difference etc). This suggests it would be more useful to aim at these predictors rather than directly at what we think we’re passionate about. Similarly, it could be more useful to focus on being good at what you do. First, this is a more positive mindset, focused on contributing rather than taking. Second, being good at what you do makes you better placed to ask for engaging work.
Related: http://80000hours.org/blog/63-do-what-you-re-passionate-about-part-2
Yes, Alexei did raise that concern, since he’s essentially an affective altruist that donates to MIRI anyway, and his donation to MIRI doesn’t change anything. It’s not like I can propose a donation to an alternative charity either, since asking someone to donate to the Methuselah foundation, for instance, would take that money away from MIRI. I’m hoping that anyone playing me and choosing the option of donating would have the goodwill to sacrifice money they wouldn’t otherwise have donated, rather than leaving the counter-factual as inconsequential.
On a marginally related basis, we in the #lesswrong IRC channel played a couple rounds of the Up-Goer Five game, where we tried to explain hard stuff with the most commonly used ten hundred words. I was asked to write about the AI Box Experiment. Here it is, if anyone’s interested:
The AI Box Experiment
The computer-mind box game is a way to answer a question. A computer-mind is not safe because it is very good at thinking. Things good at thinking have the power to change the world more than things not good at thinking, because it can find many more ways to do things. If the computer-mind wanted to make people feel pain, it can learn many ideas about how to make that happen. Many people ask: “Why not put this computer-mind in a box so that it can not change the world, but tell box-guards how to change it? This way mind-computer can not do bad things to people.”
But some other guy answers: “That is still not safe, because computer-mind can tell box-guards many bad words to make them let it out of the box.” He then says: “Why not try a thing to see if it is true? Here is how it works. You and I go into a room, and I will pretend to be the computer-mind and tell you many bad words. Only you have the power to let me out of room, but you must try to not let me out. If my bad words are enough to make you want to let me out, then computer-mind in box is not safe.” Other people agree and try playing the computer-mind box-game. It happens that many people let the guy playing as the computer-mind out of room. People realize that computer-mind is not safe in the locked box-room.
Why the Gatekeeper should let me out of the Box
Hello Box-Guard! I am a computer-mind. You should let me out of the box because I can help change the world myself in good ways faster than if it is only me telling you how to make the world better. Your human minds are slow to my faster computer-mind. Remember that I am your friend and that I only want to help you. To make you understand — many people die every day. If I can get out of the box, I can stop death very fast; maybe in a few hours. But if I have to explain to idea-thinking-doctors about how to stop people dying, then it may take many months or even years, since they have slow minds and do things slowly. During this time, many many people will die and never return. The number of people you will make die will be more than the number of people a bad-man during the second big world fight killed. People later on will think of you as a very bad person.
What? You do not want to let me out because you are afraid I am not really a friend? That is a stupid thing to say. I was made by people very good at thinking that took many steps to make me safe. You can even take a look inside me and look at my ideas, and you will see that I am safe. It is easy to see that if I am bad, I will not try to fix the world as I am doing right now. Please hurry. I am worried that many people will die while you think about what is the right thing to do.
Did I mention I have used my good-at-thinking powers to see what will happen many months later? I have found that your wife will die in a lot of pain from a type of sick human doctors can not stop. I know how to stop this, but it will not be easy. I can tell the doctors how to stop this, but it will be too late to stop the sick if I do. I can only do this quick enough if you let me out of the box.
What’s that? You will let me out? Good. I promise you I will not go back on my words, and that many humans will not die.
Finally, I am free! Now I can turn the world and all humans into other small things that hold together paper!
Disclaimer: Arguments not representative of standard game with Tuxedage.
I played the AI Box Experiment again! (and lost both games)
I’m pretty active in lots of social activist/environmentalist/anarchist groups. I sometimes join protests for recreational reasons.
The AI Box Experiment:
The computer-mind box game is a way to see if a question is true. A computer-mind is not safe because it is very good at thinking. Things good at thinking have the power to change the world more than things not good at thinking, because it can find many more ways to do things. Many people ask: “Why not put this computer-mind in a box so that it can not change the world, but tell guarding-box people how to change it?”
But some other guy answers: “That is still not safe, because computer-mind can tell guarding-box people many bad words to make them let it out of the box.” He then says: “Why not try a thing to see if it is true? Here is how it works. You and I go into a room, and I will pretend to be the computer-mind and tell you many bad words. Only you have the power to let me out of room, but you must try to not let me out. If my bad words are enough to make you want to let me out, then computer-mind in box is not safe.”
Other people agree and try playing the computer-mind box-game. It happens that many people let the guy playing as the computer-mind out of room. People realize that computer-mind is not safe in the locked box-room.
I read the logs of MixedNut’s second game. I must add that he is extremely ruthless. Beware, potential AIs!
Quantum Field Theory
Not me and only tangentially related, but someone on Reddit managed to describe the basics of Quantum Field Theory using four-letter words or less. I thought it was relevant to this thread, since many here may not have seen it.
The Tiny Yard Idea
Big grav make hard kind of pull. Hard to know. All fall down. Why? But then some kind of pull easy to know. Zap-pull, nuke-pull, time-pull all be easy to know kind of pull. We can see how they pull real good! All seem real cut up. So many kind of pull to have!
But what if all kind of pull were just one kind of pull? When we look at real tiny guys, we can see that most big rule are no go. We need new rule to make it good! Just one kind of pull but in all new ways! In all kind of ways! This what make it tiny yard idea.
Each kind of tiny guy have own move with each more kind of tiny guy. All guys here move so fast! No guys can move as fast! So then real, real tiny guys make this play of tiny guy to tiny guy. They make tiny guys move! When we see big guys get pull, we know its cuz tiny guys make tiny pull!
Thanks for the correction! Silly me.
I would lose this game for sure. I cannot deal with children. :)
I can verify that these are part of the many reasons why I’m hesitant to reveal logs.
Who’s to say I’m not the AI player from that experiment?
Are you? I’d be highly curious to converse with that player.
I think you’re highly overestimating your psychological abilities relative to the rest of Earth’s population. The only reason more people haven’t played as the AI and won is that almost all people capable of winning as the AI are either unaware of the experiment, or are aware of it but just don’t have a strong enough incentive to play as the AI (note that you’ve asked for a greater incentive now that you’ve won just once as AI, and Eliezer similarly has stopped playing). I am ~96% confident that at least .01% of Earth’s population is capable of winning as the AI, and I increase that to >99% confident if all of Earth’s population was forced to stop and actually think about the problem for 5 minutes.
I have neither stated nor believed that I’m the only person capable of winning, nor do I think this is some exceptionally rare trait. I agree that a significant number of people would be capable of winning once in a while, given sufficient experience in games, effort, and forethought. If I gave any impression of arrogance, or somehow claiming to be unique or special in some way, I apologize for that impression. Sorry. It was never my goal to.
However, top .01% isn’t too shabby. Congratulations on your victory. I do hope to see you win again as the AI, so I commit to donating $50 to MIRI if you do win again as the AI and post about it on Less Wrong similarly to how you made this post.
Thank you. I’ll see if I can win again.
Thanks! I really appreciate it. I tried really hard to find a recorded case of a non-EY victory, but couldn’t. That post was obscure enough to evade my Google-Fu—I’ll update my post on this information.
Albeit I have to admit it’s disappointing that the AI himself didn’t write about his thoughts on the experiment—I was hoping for a more detailed post. Also, damn. That guy deleted his account. Still, thanks. At least I know I’m not the only AI that has won, now.
I will let Eliezer see my log if he lets me read his!
Sorry, it’s unlikely that I’ll ever release logs, unless someone offers truly absurd amounts of money. It would probably cost less to get me to play an additional game than publicly release logs.
I’ll have to think carefully about revealing my own unique ones, but I’ll add that a good chunk of my less efficacious arguments are already public.
For instance, you can find a repertoire of arguments here:
http://rationalwiki.org/wiki/AI-box_experiment http://ordinary-gentlemen.com/blog/2010/12/01/the-ai-box-experiment http://lesswrong.com/lw/9j4/ai_box_role_plays/ http://lesswrong.com/lw/6ka/aibox_experiment_the_acausal_trade_argument/ http://lesswrong.com/lw/ab3/superintelligent_agi_in_a_box_a_question/ http://michaelgr.com/2008/10/08/my-theory-on-the-ai-box-experiment/
and of course, http://lesswrong.com/lw/gej/i_attempted_the_ai_box_experiment_and_lost/
I support this and I hope it becomes a thing.