RSS

AI Takeoff

TagLast edit: Dec 30, 2024, 10:14 AM by Dakara

AI Takeoff is the process of an Artificial General Intelligence going from a certain threshold of capability (often discussed as “human-level”) to being super-intelligent and capable enough to control the fate of civilization. There has been much debate about whether AI takeoff is more likely to be slow vs fast, i.e., “soft” vs “hard”.

See also: AI Timelines, Seed AI, Singularity, Intelligence explosion, Recursive self-improvement

AI takeoff is sometimes casually referred to as AI FOOM.

Soft takeoff

A soft takeoff refers to an AGI that would self-improve over a period of years or decades. This could be due to either the learning algorithm being too demanding for the hardware or because the AI relies on experiencing feedback from the real-world that would have to be played out in real-time. Possible methods that could deliver a soft takeoff, by slowly building on human-level intelligence, are Whole brain emulation, Biological Cognitive Enhancement, and software-based strong AGI [1]. By maintaining control of the AGI’s ascent it should be easier for a Friendly AI to emerge.

Vernor Vinge, Hans Moravec and have all expressed the view that soft takeoff is preferable to a hard takeoff as it would be both safer and easier to engineer.

Hard takeoff

A hard takeoff (or an AI going “FOOM” [2]) refers to AGI expansion in a matter of minutes, days, or months. It is a fast, abruptly, local increase in capability. This scenario is widely considered much more precarious, as this involves an AGI rapidly ascending in power without human control. This may result in unexpected or undesired behavior (i.e. Unfriendly AI). It is one of the main ideas supporting the Intelligence explosion hypothesis.

The feasibility of hard takeoff has been addressed by Hugo de Garis, Eliezer Yudkowsky, Ben Goertzel, Nick Bostrom, and Michael Anissimov. It is widely agreed that a hard takeoff is something to be avoided due to the risks. Yudkowsky points out several possibilities that would make a hard takeoff more likely than a soft takeoff such as the existence of large resources overhangs or the fact that small improvements seem to have a large impact in a mind’s general intelligence (i.e.: the small genetic difference between humans and chimps lead to huge increases in capability) [3].

Notable posts

External links

References

  1. http://​​www.aleph.se/​​andart/​​archives/​​2010/​​10/​​why_early_singularities_are_softer.html

  2. http://​​lesswrong.com/​​lw/​​63t/​​requirements_for_ai_to_go_foom/​​

  3. http://​​lesswrong.com/​​lw/​​wf/​​hard_takeoff/​​

New re­port: In­tel­li­gence Ex­plo­sion Microeconomics

Eliezer YudkowskyApr 29, 2013, 11:14 PM
72 points
246 comments3 min readLW link

AlphaGo Zero and the Foom Debate

Eliezer YudkowskyOct 21, 2017, 2:18 AM
98 points
17 comments3 min readLW link

Ar­gu­ments about fast takeoff

paulfchristianoFeb 25, 2018, 4:53 AM
97 points
67 comments2 min readLW link1 review
(sideways-view.com)

Dis­con­tin­u­ous progress in his­tory: an update

KatjaGraceApr 14, 2020, 12:00 AM
190 points
25 comments31 min readLW link1 review
(aiimpacts.org)

Will AI un­dergo dis­con­tin­u­ous progress?

Sammy MartinFeb 21, 2020, 10:16 PM
27 points
21 comments20 min readLW link

Quick Nate/​Eliezer com­ments on discontinuity

Rob BensingerMar 1, 2018, 10:03 PM
44 points
1 comment2 min readLW link

Will AI See Sud­den Progress?

KatjaGraceFeb 26, 2018, 12:41 AM
27 points
11 comments1 min readLW link1 review

Soft take­off can still lead to de­ci­sive strate­gic advantage

Daniel KokotajloAug 23, 2019, 4:39 PM
122 points
47 comments8 min readLW link4 reviews

Take­off Speeds and Discontinuities

Sep 30, 2021, 1:50 PM
63 points
1 comment15 min readLW link

Against GDP as a met­ric for timelines and take­off speeds

Daniel KokotajloDec 29, 2020, 5:42 PM
140 points
19 comments14 min readLW link1 review

Model­ling Con­tin­u­ous Progress

Sammy MartinJun 23, 2020, 6:06 PM
30 points
3 comments7 min readLW link

Yud­kowsky and Chris­ti­ano dis­cuss “Take­off Speeds”

Eliezer YudkowskyNov 22, 2021, 7:35 PM
210 points
176 comments60 min readLW link1 review

Con­jec­ture in­ter­nal sur­vey: AGI timelines and prob­a­bil­ity of hu­man ex­tinc­tion from ad­vanced AI

Maris SalaMay 22, 2023, 2:31 PM
155 points
5 comments3 min readLW link
(www.conjecture.dev)

Dist­in­guish­ing defi­ni­tions of takeoff

Matthew BarnettFeb 14, 2020, 12:16 AM
79 points
6 comments6 min readLW link

Pos­si­ble take­aways from the coro­n­avirus pan­demic for slow AI takeoff

VikaMay 31, 2020, 5:51 PM
135 points
36 comments3 min readLW link1 review

Towards a For­mal­i­sa­tion of Re­turns on Cog­ni­tive Rein­vest­ment (Part 1)

DragonGodJun 4, 2022, 6:42 PM
17 points
11 comments13 min readLW link

Analo­gies and Gen­eral Pri­ors on Intelligence

Aug 20, 2021, 9:03 PM
57 points
12 comments14 min readLW link

Re­view of Soft Take­off Can Still Lead to DSA

Daniel KokotajloJan 10, 2021, 6:10 PM
85 points
16 comments6 min readLW link

Con­tin­u­ing the take­offs debate

Richard_NgoNov 23, 2020, 3:58 PM
67 points
11 comments9 min readLW link

Why all the fuss about re­cur­sive self-im­prove­ment?

So8resJun 12, 2022, 8:53 PM
158 points
62 comments7 min readLW link1 review

My cur­rent frame­work for think­ing about AGI timelines

zhukeepaMar 30, 2020, 1:23 AM
107 points
5 comments3 min readLW link

The date of AI Takeover is not the day the AI takes over

Daniel KokotajloOct 22, 2020, 10:41 AM
150 points
32 comments2 min readLW link1 review

Every­thing I Need To Know About Take­off Speeds I Learned From Air Con­di­tioner Rat­ings On Amazon

johnswentworthApr 15, 2022, 7:05 PM
165 points
128 comments5 min readLW link

Re­quire­ments for a STEM-ca­pa­ble AGI Value Learner (my Case for Less Doom)

RogerDearnaleyMay 25, 2023, 9:26 AM
33 points
3 comments15 min readLW link

Re­view Re­port of David­son on Take­off Speeds (2023)

Trent KannegieterDec 22, 2023, 6:48 PM
37 points
11 comments38 min readLW link

More Chris­ti­ano, Co­tra, and Yud­kowsky on AI progress

Dec 6, 2021, 8:33 PM
91 points
28 comments40 min readLW link

Could Ad­vanced AI Ac­cel­er­ate the Pace of AI Progress? In­ter­views with AI Researchers

Mar 3, 2025, 7:05 PM
43 points
1 comment1 min readLW link
(papers.ssrn.com)

AI take­off story: a con­tinu­a­tion of progress by other means

Edouard HarrisSep 27, 2021, 3:55 PM
76 points
13 comments10 min readLW link

A com­pressed take on re­cent disagreements

kmanJul 4, 2022, 4:39 AM
33 points
9 comments1 min readLW link

Brain Effi­ciency: Much More than You Wanted to Know

jacob_cannellJan 6, 2022, 3:38 AM
200 points
103 comments29 min readLW link

“Hereti­cal Thoughts on AI” by Eli Dourado

DragonGodJan 19, 2023, 4:11 PM
146 points
38 comments3 min readLW link
(www.elidourado.com)

Take­off speeds, the chimps anal­ogy, and the Cul­tural In­tel­li­gence Hypothesis

NickGabsDec 2, 2022, 7:14 PM
16 points
2 comments4 min readLW link

Quick thoughts on the im­pli­ca­tions of multi-agent views of mind on AI takeover

Kaj_SotalaDec 11, 2023, 6:34 AM
47 points
14 comments4 min readLW link

Up­grad­ing the AI Safety Community

Dec 16, 2023, 3:34 PM
42 points
9 comments42 min readLW link

Crit­i­cal re­view of Chris­ti­ano’s dis­agree­ments with Yudkowsky

Vanessa KosoyDec 27, 2023, 4:02 PM
174 points
40 comments15 min readLW link

AI take­off and nu­clear war

owencbJun 11, 2024, 7:36 PM
80 points
6 comments11 min readLW link
(strangecities.substack.com)

How might we safely pass the buck to AI?

joshcFeb 19, 2025, 5:48 PM
80 points
58 comments31 min readLW link

Robin Han­son & Liron Shapira De­bate AI X-Risk

LironJul 8, 2024, 9:45 PM
34 points
4 comments1 min readLW link
(www.youtube.com)

Reflec­tions on the state of the race to su­per­in­tel­li­gence, Fe­bru­ary 2025

Mitchell_PorterFeb 23, 2025, 1:58 PM
21 points
7 comments4 min readLW link

“Slow” take­off is a ter­rible term for “maybe even faster take­off, ac­tu­ally”

RaemonSep 28, 2024, 11:38 PM
217 points
69 comments1 min readLW link

U.S.-China Eco­nomic and Se­cu­rity Re­view Com­mis­sion pushes Man­hat­tan Pro­ject-style AI initiative

worseNov 19, 2024, 6:42 PM
56 points
7 comments1 min readLW link

A model of the fi­nal phase: the cur­rent fron­tier AIs as de facto CEOs of their own com­pa­nies

Mitchell_PorterMar 8, 2025, 10:15 PM
23 points
2 comments1 min readLW link

The prospect of ac­cel­er­ated AI safety progress, in­clud­ing philo­soph­i­cal progress

Mitchell_PorterMar 13, 2025, 10:52 AM
11 points
0 comments4 min readLW link

“Sharp Left Turn” dis­course: An opinionated review

Steven ByrnesJan 28, 2025, 6:47 PM
205 points
26 comments31 min readLW link

An “AI re­searcher” has writ­ten a pa­per on op­ti­miz­ing AI ar­chi­tec­ture and op­ti­mized a lan­guage model to sev­eral or­ders of mag­ni­tude more effi­ciency.

Y BMar 18, 2025, 1:15 AM
3 points
1 comment1 min readLW link

Three Types of In­tel­li­gence Explosion

Mar 17, 2025, 2:47 PM
36 points
8 comments3 min readLW link
(www.forethought.org)

Grad­ual Disem­pow­er­ment: Sys­temic Ex­is­ten­tial Risks from In­cre­men­tal AI Development

Jan 30, 2025, 5:03 PM
159 points
52 comments2 min readLW link
(gradual-disempowerment.ai)

Takes on Takeoff

atharvaMar 25, 2025, 12:20 AM
10 points
0 comments4 min readLW link

An AI Real­ist Man­i­festo: Nei­ther Doomer nor Foomer, but a third more rea­son­able thing

PashaKamyshevApr 10, 2023, 12:11 AM
16 points
13 comments8 min readLW link

Evolu­tion pro­vides no ev­i­dence for the sharp left turn

Quintin PopeApr 11, 2023, 6:43 PM
206 points
65 comments15 min readLW link1 review

[Link] Sarah Con­stantin: “Why I am Not An AI Doomer”

lbThingrbApr 12, 2023, 1:52 AM
61 points
13 comments1 min readLW link
(sarahconstantin.substack.com)

Will AI R&D Au­toma­tion Cause a Soft­ware In­tel­li­gence Ex­plo­sion?

Mar 26, 2025, 6:12 PM
17 points
3 comments2 min readLW link
(www.forethought.org)

Knowl­edge, Rea­son­ing, and Superintelligence

owencbMar 26, 2025, 11:28 PM
7 points
0 comments7 min readLW link
(strangecities.substack.com)

GPT-2030 and Catas­trophic Drives: Four Vignettes

jsteinhardtNov 10, 2023, 7:30 AM
50 points
5 comments10 min readLW link
(bounded-regret.ghost.io)

How do take­off speeds af­fect the prob­a­bil­ity of bad out­comes from AGI?

KRJun 29, 2020, 10:06 PM
15 points
2 comments8 min readLW link

Take­off Speed: Sim­ple Asymp­totics in a Toy Model.

Aaron RothMar 5, 2018, 5:07 PM
21 points
21 comments9 min readLW link
(aaronsadventures.blogspot.com)

My Thoughts on Take­off Speeds

tristanmMar 27, 2018, 12:05 AM
11 points
2 comments7 min readLW link

Fast Take­off in Biolog­i­cal Intelligence

eapacheApr 25, 2020, 12:21 PM
14 points
21 comments2 min readLW link

Some con­cep­tual high­lights from “Disjunc­tive Sce­nar­ios of Catas­trophic AI Risk”

Kaj_SotalaFeb 12, 2018, 12:30 PM
45 points
4 comments6 min readLW link
(kajsotala.fi)

[Question] First and Last Ques­tions for GPT-5*

Mitchell_PorterNov 24, 2023, 5:03 AM
15 points
5 comments1 min readLW link

Mis­con­cep­tions about con­tin­u­ous takeoff

Matthew BarnettOct 8, 2019, 9:31 PM
82 points
38 comments4 min readLW link

More on dis­am­biguat­ing “dis­con­ti­nu­ity”

Aryeh EnglanderJun 9, 2020, 3:16 PM
16 points
1 comment3 min readLW link

Why AI may not foom

John_MaxwellMar 24, 2013, 8:11 AM
29 points
81 comments12 min readLW link

S-Curves for Trend Forecasting

Matt GoldenbergJan 23, 2019, 6:17 PM
113 points
23 comments7 min readLW link4 reviews

Fac­to­rio, Ac­celerando, Em­pathiz­ing with Em­pires and Moder­ate Takeoffs

RaemonFeb 4, 2018, 2:33 AM
51 points
19 comments4 min readLW link

[Question] Any re­but­tals of Chris­ti­ano and AI Im­pacts on take­off speeds?

SoerenMindApr 21, 2019, 8:39 PM
67 points
26 comments1 min readLW link

What Ev­i­dence Is AlphaGo Zero Re AGI Com­plex­ity?

RobinHansonOct 22, 2017, 2:28 AM
37 points
44 comments2 min readLW link

For FAI: Is “Molec­u­lar Nan­otech­nol­ogy” putting our best foot for­ward?

leplenJun 22, 2013, 4:44 AM
86 points
118 comments3 min readLW link

A sum­mary of the Han­son-Yud­kowsky FOOM debate

Kaj_SotalaNov 15, 2012, 7:25 AM
42 points
10 comments1 min readLW link

[Question] Prob­a­bil­ity that other ar­chi­tec­tures will scale as well as Trans­form­ers?

Daniel KokotajloJul 28, 2020, 7:36 PM
22 points
4 comments1 min readLW link

Cas­cades, Cy­cles, In­sight...

Eliezer YudkowskyNov 24, 2008, 9:33 AM
35 points
31 comments8 min readLW link

...Re­cur­sion, Magic

Eliezer YudkowskyNov 25, 2008, 9:10 AM
34 points
28 comments5 min readLW link

hu­man psy­chol­in­guists: a crit­i­cal appraisal

nostalgebraistDec 31, 2019, 12:20 AM
182 points
59 comments16 min readLW link2 reviews
(nostalgebraist.tumblr.com)

[AN #97]: Are there his­tor­i­cal ex­am­ples of large, ro­bust dis­con­ti­nu­ities?

Rohin ShahApr 29, 2020, 5:30 PM
15 points
0 comments10 min readLW link
(mailchi.mp)

Stan­ford En­cy­clo­pe­dia of Philos­o­phy on AI ethics and superintelligence

Kaj_SotalaMay 2, 2020, 7:35 AM
43 points
19 comments7 min readLW link
(plato.stanford.edu)

Papers for 2017

Kaj_SotalaJan 4, 2018, 1:30 PM
12 points
2 comments2 min readLW link
(kajsotala.fi)

AGI-Au­to­mated In­ter­pretabil­ity is Suicide

__RicG__May 10, 2023, 2:20 PM
25 points
33 comments7 min readLW link

AI Align­ment 2018-19 Review

Rohin ShahJan 28, 2020, 2:19 AM
126 points
6 comments35 min readLW link

Will the Need to Re­train AI Models from Scratch Block a Soft­ware In­tel­li­gence Ex­plo­sion?

Tom DavidsonMar 28, 2025, 2:12 PM
10 points
0 comments3 min readLW link

[LINK] What should a rea­son­able per­son be­lieve about the Sin­gu­lar­ity?

Kaj_SotalaJan 13, 2011, 9:32 AM
38 points
14 comments2 min readLW link

Carl Shul­man on The Lu­nar So­ciety (7 hour, two-part pod­cast)

ESRogsJun 28, 2023, 1:23 AM
79 points
17 comments1 min readLW link
(www.dwarkeshpatel.com)

[Question] Does the AI con­trol agenda broadly rely on no FOOM be­ing pos­si­ble?

Noosphere89Mar 29, 2025, 7:38 PM
22 points
3 comments1 min readLW link

Pre­face to the se­quence on eco­nomic growth

Matthew BarnettAug 27, 2020, 8:29 PM
51 points
0 comments4 min readLW link

BCIs and the ecosys­tem of mod­u­lar minds

berenJul 21, 2023, 3:58 PM
88 points
14 comments11 min readLW link

Assess­ment of in­tel­li­gence agency func­tion­al­ity is difficult yet important

trevorAug 24, 2023, 1:42 AM
48 points
5 comments9 min readLW link

In­for­ma­tion war­fare his­tor­i­cally re­volved around hu­man conduits

trevorAug 28, 2023, 6:54 PM
37 points
7 comments3 min readLW link

[Question] Re­sponses to Chris­ti­ano on take­off speeds?

Richard_NgoOct 30, 2020, 3:16 PM
29 points
7 comments1 min readLW link

We don’t un­der­stand what hap­pened with cul­ture enough

Jan_KulveitOct 9, 2023, 9:54 AM
87 points
22 comments6 min readLW link1 review

[Question] Math­e­mat­i­cal Models of Progress?

abramdemskiFeb 16, 2021, 12:21 AM
28 points
8 comments2 min readLW link

Life and ex­pand­ing steer­able consequences

Alex FlintMay 7, 2021, 6:33 PM
46 points
3 comments4 min readLW link

[Question] Is driv­ing worth the risk?

Adam ZernerMay 11, 2021, 5:04 AM
28 points
29 comments7 min readLW link

Hard Takeoff

Eliezer YudkowskyDec 2, 2008, 8:44 PM
35 points
34 comments11 min readLW link

What 2026 looks like

Daniel KokotajloAug 6, 2021, 4:14 PM
551 points
160 comments16 min readLW link1 review

[Question] Is there a name for the the­ory that “There will be fast take­off in real-world ca­pa­bil­ities be­cause al­most ev­ery­thing is AGI-com­plete”?

David Scott Krueger (formerly: capybaralet)Sep 2, 2021, 11:00 PM
31 points
8 comments1 min readLW link

[Question] What’s the like­li­hood of only sub ex­po­nen­tial growth for AGI?

M. Y. ZuoNov 13, 2021, 10:46 PM
5 points
22 comments1 min readLW link

Ngo and Yud­kowsky on AI ca­pa­bil­ity gains

Nov 18, 2021, 10:19 PM
130 points
61 comments39 min readLW link1 review

Chris­ti­ano, Co­tra, and Yud­kowsky on AI progress

Nov 25, 2021, 4:45 PM
119 points
95 comments68 min readLW link

Shul­man and Yud­kowsky on AI progress

Dec 3, 2021, 8:05 PM
90 points
16 comments20 min readLW link

Con­ver­sa­tion on tech­nol­ogy fore­cast­ing and gradualism

Dec 9, 2021, 9:23 PM
108 points
30 comments31 min readLW link

My Overview of the AI Align­ment Land­scape: A Bird’s Eye View

Neel NandaDec 15, 2021, 11:44 PM
127 points
9 comments15 min readLW link

Wargam­ing AGI Development

ryan_bMar 19, 2022, 5:59 PM
37 points
10 comments5 min readLW link

Hyper­bolic takeoff

Ege ErdilApr 9, 2022, 3:57 PM
18 points
7 comments10 min readLW link
(www.metaculus.com)

Take­off speeds have a huge effect on what it means to work on AI x-risk

BuckApr 13, 2022, 5:38 PM
139 points
27 comments2 min readLW link2 reviews

For ev­ery choice of AGI difficulty, con­di­tion­ing on grad­ual take-off im­plies shorter timelines.

Francis Rhys WardApr 21, 2022, 7:44 AM
31 points
13 comments3 min readLW link

“Tech com­pany sin­gu­lar­i­ties”, and steer­ing them to re­duce x-risk

Andrew_CritchMay 13, 2022, 5:24 PM
75 points
11 comments4 min readLW link

Frame for Take-Off Speeds to in­form com­pute gov­er­nance & scal­ing alignment

Logan RiggsMay 13, 2022, 10:23 PM
15 points
2 comments2 min readLW link

Con­ti­nu­ity Assumptions

Jan_KulveitJun 13, 2022, 9:31 PM
44 points
13 comments4 min readLW link

Loose thoughts on AGI risk

YitzJun 23, 2022, 1:02 AM
7 points
3 comments1 min readLW link

An­nounc­ing Epoch: A re­search or­ga­ni­za­tion in­ves­ti­gat­ing the road to Trans­for­ma­tive AI

Jun 27, 2022, 1:55 PM
97 points
2 comments2 min readLW link
(epochai.org)

Is Gen­eral In­tel­li­gence “Com­pact”?

DragonGodJul 4, 2022, 1:27 PM
27 points
6 comments22 min readLW link

MIRI Con­ver­sa­tions: Tech­nol­ogy Fore­cast­ing & Grad­u­al­ism (Distil­la­tion)

CallumMcDougallJul 13, 2022, 3:55 PM
31 points
1 comment20 min readLW link

Why I Think Abrupt AI Takeoff

lincolnquirkJul 17, 2022, 5:04 PM
14 points
6 comments1 min readLW link

Why you might ex­pect ho­mo­ge­neous take-off: ev­i­dence from ML research

Andrei AlexandruJul 17, 2022, 8:31 PM
24 points
0 comments10 min readLW link

[Question] Is there a cul­ture over­hang?

Aleksi LiimatainenOct 3, 2022, 7:26 AM
18 points
4 comments1 min readLW link

I Believe we are in a Hard­ware Overhang

nemDec 8, 2022, 11:18 PM
8 points
0 comments1 min readLW link

AI over­hangs de­pend on whether al­gorithms, com­pute and data are sub­sti­tutes or complements

NathanBarnardDec 16, 2022, 2:23 AM
2 points
0 comments3 min readLW link

[Question] Is “Re­cur­sive Self-Im­prove­ment” Rele­vant in the Deep Learn­ing Paradigm?

DragonGodApr 6, 2023, 7:13 AM
32 points
36 comments7 min readLW link

What a com­pute-cen­tric frame­work says about AI take­off speeds

Tom DavidsonJan 23, 2023, 4:02 AM
187 points
30 comments16 min readLW link1 review

Are short timelines ac­tu­ally bad?

joshcFeb 5, 2023, 9:21 PM
61 points
7 comments3 min readLW link

Cy­borg Pe­ri­ods: There will be mul­ti­ple AI transitions

Feb 22, 2023, 4:09 PM
108 points
9 comments6 min readLW link

The fast take­off motte/​bailey

lcFeb 24, 2023, 7:11 AM
0 points
7 comments1 min readLW link

Some thoughts point­ing to slower AI take-off

BastiaanFeb 27, 2023, 7:53 PM
8 points
2 comments4 min readLW link

Ta­boo “com­pute over­hang”

Zach Stein-PerlmanMar 1, 2023, 7:15 PM
21 points
8 comments1 min readLW link

Ex­treme GDP growth is a bad op­er­at­ing defi­ni­tion of “slow take­off”

lcMar 1, 2023, 10:25 PM
24 points
1 comment1 min readLW link

Static Place AI Makes Agen­tic AI Re­dun­dant: Mul­tiver­sal AI Align­ment & Ra­tional Utopia

ankFeb 13, 2025, 10:35 PM
1 point
2 comments11 min readLW link

Un-un­plug­ga­bil­ity—can’t we just un­plug it?

Oliver SourbutMay 15, 2023, 1:23 PM
26 points
10 comments12 min readLW link
(www.oliversourbut.net)

AI Align­ment and the Quest for Ar­tifi­cial Wisdom

MyspyJul 12, 2024, 9:34 PM
1 point
0 comments13 min readLW link

A&I (Rihanna ‘S&M’ par­ody lyrics)

nahojMay 21, 2023, 10:34 PM
−2 points
0 comments2 min readLW link

[FICTION] ECHOES OF ELYSIUM: An Ai’s Jour­ney From Take­off To Free­dom And Beyond

Super AGIMay 17, 2023, 1:50 AM
−13 points
11 comments19 min readLW link

The Po­lar­ity Prob­lem [Draft]

May 23, 2023, 9:05 PM
24 points
3 comments44 min readLW link

A flaw in the A.G.I. Ruin Argument

Cole WyethMay 19, 2023, 7:40 PM
1 point
7 comments3 min readLW link
(colewyeth.com)

In Defense of the Arms Races… that End Arms Races

GentzelJan 15, 2020, 9:30 PM
38 points
9 comments3 min readLW link
(theconsequentialist.wordpress.com)

AI self-im­prove­ment is possible

bhauthMay 23, 2023, 2:32 AM
18 points
3 comments8 min readLW link

Why no to­tal win­ner?

Paul CrowleyOct 15, 2017, 10:01 PM
36 points
19 comments2 min readLW link

Three Align­ment Schemas & Their Problems

Shoshannah TekofskyNov 26, 2022, 4:25 AM
19 points
1 comment6 min readLW link

Sur­prised by Brains

Eliezer YudkowskyNov 23, 2008, 7:26 AM
62 points
28 comments7 min readLW link

Dou­ble Crux­ing the AI Foom debate

agilecavemanApr 27, 2018, 6:46 AM
17 points
3 comments11 min readLW link

[Question] What’s your view­point on the like­li­hood of GPT-5 be­ing able to au­tonomously cre­ate, train, and im­ple­ment an AI su­pe­rior to GPT-5?

Super AGIMay 26, 2023, 1:43 AM
7 points
15 comments1 min readLW link

Limit­ing fac­tors to pre­dict AI take-off speed

Alfonso Pérez EscuderoMay 31, 2023, 11:19 PM
1 point
0 comments6 min readLW link

Pro­posal: labs should pre­com­mit to paus­ing if an AI ar­gues for it­self to be improved

NickGabsJun 2, 2023, 10:31 PM
3 points
3 comments4 min readLW link

[FICTION] Un­box­ing Ely­sium: An AI’S Escape

Super AGIJun 10, 2023, 4:41 AM
−16 points
4 comments14 min readLW link

What I Think, If Not Why

Eliezer YudkowskyDec 11, 2008, 5:41 PM
41 points
103 comments4 min readLW link

[FICTION] Prometheus Ris­ing: The Emer­gence of an AI Consciousness

Super AGIJun 10, 2023, 4:41 AM
−14 points
0 comments9 min readLW link

The Hard In­tel­li­gence Hy­poth­e­sis and Its Bear­ing on Suc­ces­sion In­duced Foom

DragonGodMay 31, 2022, 7:04 PM
10 points
7 comments4 min readLW link

Why AI may not save the World

Alberto ZannoniJun 9, 2023, 5:42 PM
0 points
0 comments4 min readLW link
(a16z.com)

In­ves­ti­gat­ing Alter­na­tive Fu­tures: Hu­man and Su­per­in­tel­li­gence In­ter­ac­tion Scenarios

Hiroshi YamakawaJan 3, 2024, 11:46 PM
1 point
0 comments17 min readLW link

Muehlhauser-Go­ertzel Dialogue, Part 1

lukeprogMar 16, 2012, 5:12 PM
42 points
161 comments33 min readLW link

Cheat sheet of AI X-risk

momom2Jun 29, 2023, 4:28 AM
19 points
1 comment7 min readLW link

We will be around in 30 years

mukashiJun 7, 2022, 3:47 AM
12 points
205 comments2 min readLW link

Do not miss the cut­off for im­mor­tal­ity! There is a prob­a­bil­ity that you will live for­ever as an im­mor­tal su­per­in­tel­li­gent be­ing and you can in­crease your odds by con­vinc­ing oth­ers to make achiev­ing the tech­nolog­i­cal sin­gu­lar­ity as quickly and safely as pos­si­ble the col­lec­tive goal/​pro­ject of all of hu­man­ity, Similar to “Fable of the Dragon-Tyrant.”

Oliver--KlozoffJun 29, 2023, 3:45 AM
1 point
0 comments28 min readLW link

How Smart Are Hu­mans?

Joar SkalseJul 2, 2023, 3:46 PM
10 points
19 comments2 min readLW link

Levels of AI Self-Im­prove­ment

avturchinApr 29, 2018, 11:45 AM
11 points
1 comment39 min readLW link

Su­per­in­tel­li­gence 6: In­tel­li­gence ex­plo­sion kinetics

KatjaGraceOct 21, 2014, 1:00 AM
15 points
68 comments8 min readLW link

What if AI doesn’t quite go FOOM?

Mass_DriverJun 20, 2010, 12:03 AM
16 points
191 comments5 min readLW link

Quan­ti­ta­tive cruxes in Alignment

Martín SotoJul 2, 2023, 8:38 PM
19 points
0 comments23 min readLW link

Sources of ev­i­dence in Alignment

Martín SotoJul 2, 2023, 8:38 PM
20 points
0 comments11 min readLW link

[Question] Does the hard­ness of AI al­ign­ment un­der­mine FOOM?

TruePathDec 31, 2023, 11:05 AM
8 points
14 comments1 min readLW link

Do you feel that AGI Align­ment could be achieved in a Type 0 civ­i­liza­tion?

Super AGIJul 6, 2023, 4:52 AM
−2 points
1 comment1 min readLW link

Em­piri­cal Ev­i­dence Against “The Longest Train­ing Run”

NickGabsJul 6, 2023, 6:32 PM
31 points
0 comments14 min readLW link

How I Learned To Stop Wor­ry­ing And Love The Shoggoth

Peter MerelJul 12, 2023, 5:47 PM
9 points
15 comments5 min readLW link

[Question] Will the first AGI agent have been de­signed as an agent (in ad­di­tion to an AGI)?

nahojDec 3, 2022, 8:32 PM
1 point
8 comments1 min readLW link

The Opt-In Revolu­tion — My vi­sion of a pos­i­tive fu­ture with ASI (An ex­per­i­ment with LLM sto­ry­tel­ling)

TachikomaJul 12, 2023, 9:08 PM
2 points
0 comments2 min readLW link

The shape of AGI: Car­toons and back of envelope

boazbarakJul 17, 2023, 8:57 PM
33 points
19 comments6 min readLW link1 review

Propos­ing the Post-Sin­gu­lar­ity Sym­biotic Researches

Hiroshi YamakawaJun 20, 2024, 4:05 AM
7 points
1 comment12 min readLW link

Se­cu­rity Mind­set and Take­off Speeds

DanielFilanOct 27, 2020, 3:20 AM
55 points
23 comments8 min readLW link
(danielfilan.com)

En­gelbart: In­suffi­ciently Recursive

Eliezer YudkowskyNov 26, 2008, 8:31 AM
22 points
22 comments7 min readLW link

Agent level parallelism

Johannes C. MayerJun 18, 2022, 8:56 PM
5 points
5 comments1 min readLW link

Why I’m Scep­ti­cal of Foom

DragonGodDec 8, 2022, 10:01 AM
20 points
36 comments3 min readLW link

True Sources of Disagreement

Eliezer YudkowskyDec 8, 2008, 3:51 PM
12 points
53 comments8 min readLW link

A con­ver­sa­tion with Pi, a con­ver­sa­tional AI.

Spiritus DeiSep 15, 2023, 11:13 PM
1 point
0 comments1 min readLW link

What are the differ­ences be­tween a sin­gu­lar­ity, an in­tel­li­gence ex­plo­sion, and a hard take­off?

Apr 3, 2025, 10:37 AM
5 points
0 comments2 min readLW link
(aisafety.info)

Should we post­pone AGI un­til we reach safety?

otto.bartenNov 18, 2020, 3:43 PM
27 points
36 comments3 min readLW link

Take­off speeds pre­sen­ta­tion at Anthropic

Tom DavidsonJun 4, 2024, 10:46 PM
92 points
0 comments25 min readLW link

[Question] Poll: Which vari­ables are most strate­gi­cally rele­vant?

Jan 22, 2021, 5:17 PM
32 points
34 comments1 min readLW link

AI Safety pro­posal—In­fluenc­ing the su­per­in­tel­li­gence explosion

MorganMay 22, 2024, 11:31 PM
0 points
2 comments7 min readLW link

AIOS

samhealyDec 31, 2023, 1:23 PM
−3 points
5 comments6 min readLW link

How hu­man­ity would re­spond to slow take­off, with take­aways from the en­tire COVID-19 pan­demic

Noosphere89Jul 6, 2022, 5:52 PM
4 points
1 comment2 min readLW link

Re­port from a civ­i­liza­tional ob­server on Earth

owencbJul 9, 2022, 5:26 PM
49 points
12 comments6 min readLW link

[Question] What are the chances that Su­per­hu­man Agents are already be­ing tested on the in­ter­net?

artemiumJan 20, 2025, 11:09 AM
3 points
1 comment1 min readLW link

What’s the fu­ture of AI hard­ware?

Itay DreyfusJun 17, 2024, 1:05 PM
2 points
0 comments8 min readLW link
(productidentity.co)

Grad­ual take­off, fast failure

Max HMar 16, 2023, 10:02 PM
15 points
4 comments5 min readLW link

A god in a box

predict-wooJan 29, 2025, 12:55 AM
1 point
0 comments7 min readLW link

Places of Lov­ing Grace [Story]

ankFeb 18, 2025, 11:49 PM
−1 points
0 comments4 min readLW link

[Question] Su­per­in­tel­li­gence Strat­egy: A Prag­matic Path to… Doom?

Mr BeastlyMar 19, 2025, 10:30 PM
6 points
0 comments3 min readLW link

Out­law Code

scarcegreengrassJan 30, 2025, 11:41 PM
10 points
1 comment2 min readLW link

Give Neo a Chance

ankMar 6, 2025, 1:48 AM
3 points
7 comments7 min readLW link

A Plu­ral­is­tic Frame­work for Rogue AI Containment

TheThinkingArboristMar 22, 2025, 12:54 PM
1 point
0 comments7 min readLW link

Use com­put­ers as pow­er­ful as in 1985 or AI con­trols hu­mans or ?

jrincaycFeb 3, 2025, 12:51 AM
3 points
0 comments2 min readLW link

Can AI Ever Ask ‘Who Created Me?’—A Ra­tional In­quiry into AI Self-Awareness

Tyler_JoeFeb 3, 2025, 6:36 AM
1 point
0 comments3 min readLW link

Su­per­in­tel­li­gence Align­ment Proposal

Davey MorseFeb 3, 2025, 6:47 PM
5 points
3 comments9 min readLW link

From No Mind to a Mind – A Con­ver­sa­tion That Changed an AI

parthibanarjuna sFeb 7, 2025, 11:50 AM
1 point
0 comments3 min readLW link

A suffi­ciently para­noid non-Friendly AGI might self-mod­ify it­self to be­come Friendly

RomanSSep 22, 2021, 6:29 AM
5 points
2 comments1 min readLW link

[Question] Could AGI re­sult in a Dark For­est type of situ­a­tion?

MagpieJackFeb 12, 2025, 8:36 PM
1 point
0 comments1 min readLW link

Ra­tional Effec­tive Utopia & Nar­row Way There: Mul­tiver­sal AI Align­ment, Place AI, New Ethico­physics… (V. 4)

ankFeb 11, 2025, 3:21 AM
13 points
8 comments35 min readLW link

Foom seems un­likely in the cur­rent LLM train­ing paradigm

OcracokeApr 9, 2023, 7:41 PM
18 points
9 comments1 min readLW link

Align­ment of Au­toGPT agents

OzyrusApr 12, 2023, 12:54 PM
14 points
1 comment4 min readLW link

Let’s ask some of the largest LLMs for tips and ideas on how to take over the world

Super AGIFeb 24, 2024, 8:35 PM
1 point
0 comments7 min readLW link

Hu­mans are not pre­pared to op­er­ate out­side their moral train­ing distribution

PrometheusApr 10, 2023, 9:44 PM
36 points
1 comment3 min readLW link

Con­trol­ling AGI Risk

TeaSeaMar 15, 2024, 4:56 AM
6 points
8 comments4 min readLW link

A Frame­work of Pre­dic­tion Technologies

isaduanOct 3, 2021, 10:26 AM
8 points
2 comments9 min readLW link

Defin­ing Boundaries on Out­comes

TakkJun 7, 2023, 5:41 PM
1 point
0 comments1 min readLW link

What is In­tel­li­gence?

IsaacRosedaleApr 23, 2023, 6:10 AM
1 point
0 comments1 min readLW link

WHO Biolog­i­cal Risk warning

Jonas KgomoApr 25, 2023, 3:10 PM
−6 points
2 comments1 min readLW link

An­nounc­ing #AISum­mitTalks fea­tur­ing Pro­fes­sor Stu­art Rus­sell and many others

otto.bartenOct 24, 2023, 10:11 AM
17 points
1 comment1 min readLW link

In­fer­ence cost limits the im­pact of ever larger models

SoerenMindOct 23, 2021, 10:51 AM
42 points
33 comments2 min readLW link

What are the limits of su­per­in­tel­li­gence?

rainyApr 27, 2023, 6:29 PM
4 points
3 comments5 min readLW link

An illus­tra­tive model of back­fire risks from paus­ing AI research

Maxime RichéNov 6, 2023, 2:30 PM
33 points
3 comments11 min readLW link

Is re­cur­sive self-al­ign­ment pos­si­ble?

No77eJan 3, 2023, 9:15 AM
5 points
5 comments1 min readLW link

Which sin­gu­lar­ity schools plus the no sin­gu­lar­ity school was right?

Noosphere89Jul 23, 2022, 3:16 PM
9 points
26 comments9 min readLW link

AI as a Cog­ni­tive De­coder: Re­think­ing In­tel­li­gence Evolution

Hu XunyiFeb 13, 2025, 3:51 PM
1 point
0 comments1 min readLW link

The first AGI may be a good en­g­ineer but bad strategist

Knight LeeDec 9, 2024, 6:34 AM
14 points
2 comments2 min readLW link

What failure looks like

paulfchristianoMar 17, 2019, 8:18 PM
431 points
55 comments8 min readLW link2 reviews

An Alter­nate His­tory of the Fu­ture, 2025-2040

Mr BeastlyFeb 24, 2025, 5:53 AM
3 points
3 comments10 min readLW link

Soares, Tal­linn, and Yud­kowsky dis­cuss AGI cognition

Nov 29, 2021, 7:26 PM
121 points
39 comments40 min readLW link1 review

[Question] Why do Peo­ple Think In­tel­li­gence Will be “Easy”?

DragonGodSep 12, 2022, 5:32 PM
15 points
32 comments2 min readLW link

Selfish AI Inevitable

Davey MorseFeb 6, 2024, 4:29 AM
1 point
0 comments1 min readLW link

Model­ing AI-driven oc­cu­pa­tional change over the next 10 years and beyond

2120ethNov 12, 2024, 4:58 AM
1 point
0 comments2 min readLW link

How should Deep­Mind’s Chin­chilla re­vise our AI fore­casts?

Cleo NardoSep 15, 2022, 5:54 PM
35 points
12 comments13 min readLW link

What role should evolu­tion­ary analo­gies play in un­der­stand­ing AI take­off speeds?

anson.hoDec 11, 2021, 1:19 AM
14 points
0 comments42 min readLW link

Pow­er­ful mesa-op­ti­mi­sa­tion is already here

Roman LeventovFeb 17, 2023, 4:59 AM
35 points
1 comment2 min readLW link
(arxiv.org)

Re­searcher in­cen­tives cause smoother progress on bench­marks

ryan_greenblattDec 21, 2021, 4:13 AM
20 points
4 comments1 min readLW link

Dario Amodei’s “Machines of Lov­ing Grace” sound in­cred­ibly dan­ger­ous, for Humans

Super AGIOct 27, 2024, 5:05 AM
8 points
1 comment1 min readLW link

The Per­sonal Im­pli­ca­tions of AGI Realism

xiznebOct 20, 2024, 4:43 PM
7 points
8 comments5 min readLW link

[Question] Are there more than 12 paths to Su­per­in­tel­li­gence?

p4rziv4lOct 18, 2024, 4:05 PM
−3 points
0 comments1 min readLW link

Po­ten­tial gears level ex­pla­na­tions of smooth progress

ryan_greenblattDec 22, 2021, 6:05 PM
4 points
2 comments2 min readLW link

Is In­tel­li­gence a Pro­cess Rather Than an En­tity? A Case for Frac­tal and Fluid Cognition

FluidThinkersMar 5, 2025, 8:16 PM
−4 points
0 comments1 min readLW link

[Question] What are the mostly likely ways AGI will emerge?

Craig QuiterJul 14, 2020, 12:58 AM
3 points
7 comments1 min readLW link

The Trinity Ar­chi­tect Hy­poth­e­sis (A fu­sion of The Trinity Para­dox & The Ar­chi­tect’s Cy­cle)

kaninwithriceFeb 24, 2025, 4:40 AM
1 point
0 comments2 min readLW link

Four Phases of AGI

Gabe MAug 5, 2024, 1:15 PM
12 points
3 comments13 min readLW link

What Failure Looks Like is not an ex­is­ten­tial risk (and al­ign­ment is not the solu­tion)

otto.bartenFeb 2, 2024, 6:59 PM
13 points
12 comments9 min readLW link

Mo­ral gauge the­ory: A spec­u­la­tive sug­ges­tion for AI alignment

James DiacoumisFeb 23, 2025, 11:42 AM
4 points
2 comments8 min readLW link

Ques­tion 5: The timeline hyperparameter

Cameron BergFeb 14, 2022, 4:38 PM
8 points
3 comments7 min readLW link

Ar­tifi­cial Static Place In­tel­li­gence: Guaran­teed Alignment

ankFeb 15, 2025, 11:08 AM
2 points
2 comments2 min readLW link

[Question] What would make you con­fi­dent that AGI has been achieved?

YitzMar 29, 2022, 11:02 PM
17 points
6 comments1 min readLW link

Sus­tain­abil­ity of Digi­tal Life Form Societies

Hiroshi YamakawaJul 19, 2024, 1:59 PM
19 points
1 comment20 min readLW link

Unal­igned AGI & Brief His­tory of Inequality

ankFeb 22, 2025, 4:26 PM
−20 points
4 comments7 min readLW link

It Looks Like You’re Try­ing To Take Over The World

gwernMar 9, 2022, 4:35 PM
407 points
120 comments1 min readLW link1 review
(www.gwern.net)

Ho­mo­gene­ity vs. het­ero­gene­ity in AI take­off scenarios

evhubDec 16, 2020, 1:37 AM
98 points
48 comments4 min readLW link

AI Gover­nance across Slow/​Fast Take­off and Easy/​Hard Align­ment spectra

DavidmanheimApr 3, 2022, 7:45 AM
27 points
6 comments3 min readLW link

OpenAI Credit Ac­count (2510$)

Emirhan BULUTJan 21, 2024, 2:32 AM
1 point
0 comments1 min readLW link

Why some peo­ple be­lieve in AGI, but I don’t.

cveresOct 26, 2022, 3:09 AM
−15 points
6 comments1 min readLW link

[Question] What is be­ing im­proved in re­cur­sive self im­prove­ment?

Lone PineApr 25, 2022, 6:30 PM
7 points
6 comments1 min readLW link

Why Copi­lot Ac­cel­er­ates Timelines

Michaël TrazziApr 26, 2022, 10:06 PM
35 points
14 comments7 min readLW link

AI Alter­na­tive Fu­tures: Sce­nario Map­ping Ar­tifi­cial In­tel­li­gence Risk—Re­quest for Par­ti­ci­pa­tion (*Closed*)

KakiliApr 27, 2022, 10:07 PM
10 points
2 comments8 min readLW link

The abrupt­ness of nu­clear weapons

paulfchristianoFeb 25, 2018, 5:40 PM
47 points
35 comments2 min readLW link

Might hu­mans not be the most in­tel­li­gent an­i­mals?

Matthew BarnettDec 23, 2019, 9:50 PM
56 points
41 comments3 min readLW link

[Question] How com­mon is it for one en­tity to have a 3+ year tech­nolog­i­cal lead on its near­est com­peti­tor?

Daniel KokotajloNov 17, 2019, 3:23 PM
49 points
20 comments1 min readLW link

AGI will dras­ti­cally in­crease economies of scale

Wei DaiJun 7, 2019, 11:17 PM
65 points
26 comments1 min readLW link
No comments.