CFAR Handbook: Introduction
The Center for Applied Rationality is a Bay Area non-profit that, among other things, ran lots of workshops to offer people tools and techniques for solving problems and improving their thinking. Those workshops were accompanied by a reference handbook, which has been available as a PDF since 2020.
The handbook hasn’t been substantially updated since it was written in 2016, but it remains a fairly straightforward primer for a lot of core rationality content. The LW team, working with the handbook’s author Duncan Sabien, have decided to republish it as a lightly-edited sequence, so that each section can be linked on its own.
In the workshop context, the handbook was a supplement to lectures, activities, and conversations taking place between participants and staff. Care was taken to emphasize the fact that each tool or technique or perspective was only as good as it was effectively applied to one’s problems, plans, and goals. The workshop was intentionally structured to cause participants to actually try things (including iterating on or developing their own versions of what they were being shown), rather than simply passively absorb content. Keep this in mind as you read—mere knowledge of how to exercise does not confer the benefits of exercise!
Discussion is strongly encouraged, and disagreement and debate are explicitly welcomed. Many LWers (including the staff of CFAR itself) have been tinkering with these concepts for years, and will have developed new perspectives on them, or interesting objections to them, or thoughts about how they work or break in practice. What follows is a historical artifact—the rough state-of-the-art at the time the handbook was written, circa 2017. That’s an excellent jumping-off point, especially for newcomers, but there’s been a lot of scattered progress since then, and we hope some of it will make its way into the comments.
- All images from the WaitButWhy sequence on AI by 8 Apr 2023 7:36 UTC; 73 points) (
- AI Safety is Dropping the Ball on Clown Attacks by 22 Oct 2023 20:09 UTC; 73 points) (
- Critique of some recent philosophy of LLMs’ minds by 20 Jan 2023 12:53 UTC; 52 points) (
- Which LessWrong/Alignment topics would you like to be tutored in? [Poll] by 19 Sep 2024 1:35 UTC; 43 points) (
- Upgrading the AI Safety Community by 16 Dec 2023 15:34 UTC; 42 points) (
- How can I get help becoming a better rationalist? by 13 Jul 2023 13:41 UTC; 40 points) (
- [Linkpost] George Mack’s Razors by 27 Nov 2023 17:53 UTC; 38 points) (
- 19 Sep 2024 2:16 UTC; 25 points) 's comment on Which LessWrong/Alignment topics would you like to be tutored in? [Poll] by (
- Kick-off meeting: Doing Things Better—A course in the art of applied rationality by 11 Jan 2023 13:58 UTC; 11 points) (EA Forum;
- 7 Jan 2024 3:05 UTC; 10 points) 's comment on The Sequences on YouTube by (
- 14 May 2023 15:12 UTC; 9 points) 's comment on Dark Forest Theories by (
- 6 Aug 2023 19:29 UTC; 7 points) 's comment on How can we improve Infohazard Governance in EA Biosecurity? by (EA Forum;
- What AI Posts Do You Want Distilled? by 25 Aug 2023 9:01 UTC; 7 points) (
- What’s the best time-efficient alternative to the Sequences? by 16 Dec 2022 20:17 UTC; 7 points) (
- 24 Jul 2022 2:27 UTC; 4 points) 's comment on Curating “The Epistemic Sequences” (list v.0.1) by (
- 4 Nov 2023 23:48 UTC; 3 points) 's comment on We are already in a persuasion-transformed world and must take precautions by (
- 24 Jul 2022 3:52 UTC; 2 points) 's comment on Curating “The Epistemic Sequences” (list v.0.1) by (
- 17 Mar 2024 20:37 UTC; 1 point) 's comment on I’ve read the Effective Altruism Handbook. Here’s what I learned. by (EA Forum;
- 5 Aug 2023 18:53 UTC; -1 points) 's comment on video games > IQ tests by (
- AI Safety is Dropping the Ball on Clown Attacks by 21 Oct 2023 23:15 UTC; -17 points) (EA Forum;
Thanks. This is the kind of content I originally came to LW for a decade ago, but seems to have become less popular
Couple questions about this sequence.
Is there any plan to write down and post more of the surrounding content like activities/lectures/etc.?
How does CFAR feel about “off-brand”/”knockoff” versions of these workshops being run at meetups? If OK with it, how should those be announced/disclaimed to make it clear that they’re not affiliated with CFAR?
I’m interested in this as an organizer, and based on conversations at the meetup organizers’ retreat this weekend, I think a number of other organizers would be interested as well.
There aren’t currently plans to write up e.g. descriptions of the classes and activities, but there are lots of people who have been to CFAR workshops who can offer their anecdotes, and you may be able to reach out to CFAR directly for descriptions of what a workshop is like.
(Also, there are going to be workshops in Europe this fall that you could attend if you want.)
As for spreading off-brand versions of the content: CFAR is enthusiastically pro the idea! Their main request is just that you clearly headline:
That CFAR originated the content you’re attempting to convey (e.g. credit them for terms like “TAPs”)
That you are teaching your version of CFAR’s TAPs (or whatever); that this is “what I, Maia, got out of attempting to learn the CFAR technique called TAPs.”
As long as you’re crediting the creators and not claiming to speak with authority about the thing you’re teaching, CFAR is (very) happy to have other people spreading the content.
Nitpick (about a line that I otherwise quite liked):
Should “convey” be “confer” here? “Convey” implies that the thing changing is typically information (i.e. knowledge), whereas “confer” implies that someone has been granted possession of something (i.e. health).
Seems right. =)
>What follows is a historical artifact—the rough state-of-the-art at the time the handbook was written, circa 2017.
The PDF hosted on CFAR’s website is dated January 2021. Is that version of the handbook more up-to-date than this one?
This was written in mid-2022, so it’s probably at least somewhat more up-to-date (but also might be slightly tweaked in the direction of ‘stuff Duncan endorses’. Duncan no longer works at CFAR and if CFAR makes further updates to the handbook I could imagine them diverging)
Here’s a simple test which can be used to evaluate the qualifications of all individuals and groups claiming to be qualified to teach rational thinking.
How much have they written or otherwise contributed on the subject of nuclear weapons?
As example, as a thought experiment imagine that I walk around all day with a loaded gun in my mouth, but I typically don’t find the gun interesting enough to discuss. In such a case, would you consider me an authority on rational thinking? In this example, the gun in one person’s mouth represents the massive hydrogen bombs in all of our mouths.
Almost all intellectual elites will fail this test. Once this is seen one’s relationship with intellectual elites can change substantially.
Note: despite the different username, I’m the author of the handbook and a former CFAR staff member.
I disagree with this take as specifically outlined, even though I do think there’s a kernel of truth to it.
Mainly, I disagree with it because it presupposes that obviously the important thing to talk about is nuclear weapons!
I suspect that Phil is unaware that the vast majority of both CFAR staff and prolific LWers have indeed 100% passed the real version of his test, which is writing and contributing to the subject of existential risk, especially that from artificial intelligence.
Phil may disagree with the claim that nuclear weapons are something like third on the list, rather than the top item, but that doesn’t mean he’s right. And CFAR staff certainly clear the bar of “spending a lot of time focusing on what seems to them to be the actually most salient threat.”
I agree that if somebody seems to be willfully ignoring a salient threat, they have gaps in their rationality that should give you pause.
Hi again Duncan,
Can AI destroy modern civilization in the next 30 minutes? Can a single human being unilaterally decide to make that happen, right now, today?
I feel that nuclear weapons are a very useful tool for analysis because unlike emerging technologies like AI, genetic engineering etc they are very easily understood by almost the entire population. So if we’re not talking about nukes, which we overwhelmingly are not across the culture at every level of society, it’s not because we don’t understand. It’s because we are in a deep denial similar to how we relate to our own personal mortality. To debunk my own posts, puncturing such deep denial with mere logic is not very promising, but one does what one knows how to do.
Except that is not the test I proposed. That’s ignoring the most pressing threat to engage a threat that’s more fun to talk about. That said, any discussion of X risk must be applauded, and I do so applaud.
The challenge I’ve presented is not to the EA community in particular, who seem far ahead of other intellectual elites on the subject of X risk generally. I’m really challenging the entire leadership of our society. I tend to focus the challenge mostly at intellectual elites of all types, because that’s who I have the highest expectations of. You know, it’s probably pointless to challenge politicians and the media on such subjects.
Doubt it, but it might depend on how much of an overhang we have. My timelines aren’t that short, but if there were an overhang and we were just a few breakthroughs away from recursive self-improvement, would the world look any different than it does now?
Oh, good point. Pilots have intentionally crashed planes full of passengers. Kids have shot up schools, not expecting to come out alive. Murder-suicide is a thing humans have been known to do. There have been a number of well-documented close calls in the Cold War. As nuclear powers proliferate, MAD becomes more complicated.
It’s still about #3 on my catastrophic risk list depending on how you count things. But the number of humans who could plausibly do this remains relatively small. How many human beings could plausibly bioengineer a pandemic? I think the number is greater, and increasing as biotech advances. Time is not the only factor in risk calculations.
And likely neither of these results in human extinction, but the pandemic scares me more. No, nuclear war wouldn’t do it. That would require salted bombs, which have been theorized, but never deployed. Can’t happen in the next 30 minutes. Fallout become survivable (if unhealthy) in a few days. Nobody is really interested in bombing New Zealand. They’re too far away from everybody else to matter. Nuclear winter risk has been greatly exaggerated, and humans are more omnivorous than you’d think, especially with even simple technology helping to process food sources. Not to say that a nuclear war wouldn’t be catastrophic, but there would be survivors. A lot of them.
A communicable disease that’s too deadly (like SARS-1) tends to burn itself out before spreading much, but an engineered (or natural!) pandemic could plausibly thread the needle and become something at least as bad as smallpox. A highly contagious disease that doesn’t kill outright but causes brain damage or sterility might be similarly devastating to civilization, without being so self-limiting. Even New Zealand might not be safe. A nuclear war ends. A pandemic festers. Outcomes could be worse, and it’s more likely to happen, and becoming more likely to happen. It’s #2 for me.
And #1 is an intelligence explosion. This is not just a catastrophic risk, but an existential one. An unaligned AI destroys all value, by default. It’s not going to have a conscience unless we put one in. Nobody knows how to do that. And short of a collapse of civilization, an AI takeover seems inevitable in short order. We either figure out how to build one that’s aligned before that happens, and it solves all the other solvable risks, or everybody dies.