I haven’t had time to reread this sequence in depth, but I wanted to at least touch on how I’d evaluate it. It seems to be aiming to be both a good introductory sequence, while being a “complete and compelling case I can for why the development of AGI might pose an existential threat”.
The question is who is this sequence for, what is it’s goal, and how does it compare to other writing targeting similar demographics.
Some writing that comes to mind to compare/contrast it with includes:
Scott Alexander’s Superintelligence FAQ. This is the post I’ve found most helpful for convincing people (including myself), that yes, AI is just actually a big deal and an extinction risk. It’s 8000 words. It’s written fairly entertainingly. What I find particularly compelling here are a bunch of factual statements about recent AI advances that I hadn’t known about at the time.
Tim Urban’s Road To Superintelligence series. This is even more optimized for entertainingness. I recall it being a bit more handwavy and making some claims that were either objectionable, or at least felt more objectionable. It’s 22,000 words.
Alex Flint’s AI Risk for Epistemic Minimalists. This goes in a pretty different direction – not entertaining, and not really comprehensive either . It came to mind because it’s doing a sort-of-similar thing of “remove as many prerequisites or assumptions as possible”. (I’m not actually sure it’s that helpful, the specific assumptions it’s avoiding making don’t feel like issues I expect to come up for most people, and then it doesn’t make a very strong claim about what to do)
(I recall Scott Alexander once trying to run a pseudo-study where he had people read a randomized intro post on AI alignment, I think including his own Superintelligence FAQ and Tim Urban’s posts among others, and see how it changed people’s minds. I vaguely recall it didn’t find that big a difference between them. I’d be curious how this compared)
At a glance, AGI Safety From First Principles seems to be more complete than Alex Flint’s piece, and more serious/a-bit-academic than Scott or Tim’s writing. I assume it’s aiming for a somewhat skeptical researcher, and is meant to not only convince them the problem exists, but give them some technical hooks of how to start thinking about it. I’m curious how well it actually succeeds at that.
I haven’t had time to reread this sequence in depth, but I wanted to at least touch on how I’d evaluate it. It seems to be aiming to be both a good introductory sequence, while being a “complete and compelling case I can for why the development of AGI might pose an existential threat”.
The question is who is this sequence for, what is it’s goal, and how does it compare to other writing targeting similar demographics.
Some writing that comes to mind to compare/contrast it with includes:
Scott Alexander’s Superintelligence FAQ. This is the post I’ve found most helpful for convincing people (including myself), that yes, AI is just actually a big deal and an extinction risk. It’s 8000 words. It’s written fairly entertainingly. What I find particularly compelling here are a bunch of factual statements about recent AI advances that I hadn’t known about at the time.
Tim Urban’s Road To Superintelligence series. This is even more optimized for entertainingness. I recall it being a bit more handwavy and making some claims that were either objectionable, or at least felt more objectionable. It’s 22,000 words.
Alex Flint’s AI Risk for Epistemic Minimalists. This goes in a pretty different direction – not entertaining, and not really comprehensive either . It came to mind because it’s doing a sort-of-similar thing of “remove as many prerequisites or assumptions as possible”. (I’m not actually sure it’s that helpful, the specific assumptions it’s avoiding making don’t feel like issues I expect to come up for most people, and then it doesn’t make a very strong claim about what to do)
(I recall Scott Alexander once trying to run a pseudo-study where he had people read a randomized intro post on AI alignment, I think including his own Superintelligence FAQ and Tim Urban’s posts among others, and see how it changed people’s minds. I vaguely recall it didn’t find that big a difference between them. I’d be curious how this compared)
At a glance, AGI Safety From First Principles seems to be more complete than Alex Flint’s piece, and more serious/a-bit-academic than Scott or Tim’s writing. I assume it’s aiming for a somewhat skeptical researcher, and is meant to not only convince them the problem exists, but give them some technical hooks of how to start thinking about it. I’m curious how well it actually succeeds at that.