RSS

AI Risk Con­crete Stories

TagLast edit: Dec 30, 2024, 10:11 AM by Dakara

AI Risk Concrete Stories are narratives that illustrate potential catastrophic scenarios involving advanced AI systems, often used to make abstract risks more tangible and relatable. These stories typically describe specific ways AI systems might cause harm to humanity.

See also Threat Models

Catas­trophic Risks from AI #6: Dis­cus­sion and FAQ

Jun 27, 2023, 11:23 PM
24 points
1 comment13 min readLW link
(arxiv.org)

“If we go ex­tinct due to mis­al­igned AI, at least na­ture will con­tinue, right? … right?”

plexMay 18, 2024, 2:09 PM
47 points
23 comments2 min readLW link
(aisafety.info)

Will GPT-5 be able to self-im­prove?

Nathan Helm-BurgerApr 29, 2023, 5:34 PM
18 points
22 comments3 min readLW link

It Looks Like You’re Try­ing To Take Over The World

gwernMar 9, 2022, 4:35 PM
407 points
120 comments1 min readLW link1 review
(www.gwern.net)

What Mul­tipo­lar Failure Looks Like, and Ro­bust Agent-Ag­nos­tic Pro­cesses (RAAPs)

Andrew_CritchMar 31, 2021, 11:50 PM
282 points
65 comments22 min readLW link1 review

Another plau­si­ble sce­nario of AI risk: AI builds mil­i­tary in­fras­truc­ture while col­lab­o­rat­ing with hu­mans, defects later.

avturchinJun 10, 2022, 5:24 PM
10 points
2 comments1 min readLW link

Clar­ify­ing “What failure looks like”

Sam ClarkeSep 20, 2020, 8:40 PM
97 points
14 comments17 min readLW link

A plau­si­ble story about AI risk.

DeLesley HutchinsJun 10, 2022, 2:08 AM
16 points
2 comments4 min readLW link

LifeKeeper Diaries: Ex­plor­ing Misal­igned AI Through In­ter­ac­tive Fiction

Nov 9, 2024, 8:58 PM
15 points
5 comments2 min readLW link

Slow mo­tion videos as AI risk in­tu­ition pumps

Andrew_CritchJun 14, 2022, 7:31 PM
241 points
41 comments2 min readLW link1 review

The Sim­plest Good

Jesse HooglandFeb 2, 2025, 7:51 PM
72 points
6 comments5 min readLW link

Rishi to out­line his vi­sion for Bri­tain to take the world lead in polic­ing AI threats when he meets Joe Biden

Mati_RoyJun 6, 2023, 4:47 AM
25 points
1 comment1 min readLW link
(www.dailymail.co.uk)

Catas­trophic Risks from AI #1: Introduction

Jun 22, 2023, 5:09 PM
40 points
1 comment5 min readLW link
(arxiv.org)

Catas­trophic Risks from AI #2: Mal­i­cious Use

Jun 22, 2023, 5:10 PM
38 points
1 comment17 min readLW link
(arxiv.org)

Catas­trophic Risks from AI #3: AI Race

Jun 23, 2023, 7:21 PM
18 points
9 comments29 min readLW link
(arxiv.org)

The next decades might be wild

Marius HobbhahnDec 15, 2022, 4:10 PM
175 points
42 comments41 min readLW link1 review

Catas­trophic Risks from AI #4: Or­ga­ni­za­tional Risks

Jun 26, 2023, 7:36 PM
23 points
0 comments21 min readLW link
(arxiv.org)

Catas­trophic Risks from AI #5: Rogue AIs

Jun 27, 2023, 10:06 PM
15 points
0 comments22 min readLW link
(arxiv.org)

“Hu­man­ity vs. AGI” Will Never Look Like “Hu­man­ity vs. AGI” to Humanity

Thane RuthenisDec 16, 2023, 8:08 PM
190 points
34 comments5 min readLW link

What suc­cess looks like

Jun 28, 2022, 2:38 PM
19 points
4 comments1 min readLW link
(forum.effectivealtruism.org)

Re­spond­ing to ‘Beyond Hyper­an­thro­po­mor­phism’

ukc10014Sep 14, 2022, 8:37 PM
9 points
0 comments16 min readLW link

AI Safety Endgame Stories

Ivan VendrovSep 28, 2022, 4:58 PM
31 points
11 comments11 min readLW link

A Story of AI Risk: In­struc­tGPT-N

peterbarnettMay 26, 2022, 11:22 PM
24 points
0 comments8 min readLW link

A bridge to Dath Ilan? Im­proved gov­er­nance on the crit­i­cal path to AI al­ign­ment.

Jackson WagnerMay 18, 2022, 3:51 PM
24 points
0 comments12 min readLW link

Hu­man level AI can plau­si­bly take over the world

anithiteMar 1, 2023, 11:27 PM
27 points
12 comments2 min readLW link

Grad­ual take­off, fast failure

Max HMar 16, 2023, 10:02 PM
15 points
4 comments5 min readLW link

Brain­storm­ing: Slow Takeoff

David PiepgrassJan 23, 2024, 6:58 AM
3 points
0 comments51 min readLW link

Places of Lov­ing Grace [Story]

ankFeb 18, 2025, 11:49 PM
−1 points
0 comments4 min readLW link

Mus­ings on Sce­nario Fore­cast­ing and AI

Alvin ÅnestrandMar 6, 2025, 12:28 PM
10 points
0 comments11 min readLW link
(forecastingaifutures.substack.com)

A god in a box

predict-wooJan 29, 2025, 12:55 AM
1 point
0 comments7 min readLW link

Are we the Wolves now? Hu­man Eu­gen­ics un­der AI Control

BritJan 30, 2025, 8:31 AM
−2 points
1 comment2 min readLW link

Out­law Code

scarcegreengrassJan 30, 2025, 11:41 PM
10 points
1 comment2 min readLW link

How AI Takeover Might Hap­pen in 2 Years

joshcFeb 7, 2025, 5:10 PM
394 points
131 comments29 min readLW link
(x.com)

Scale Was All We Needed, At First

Gabe MFeb 14, 2024, 1:49 AM
295 points
34 comments8 min readLW link
(aiacumen.substack.com)

The Peril of the Great Leaks (writ­ten with ChatGPT)

bvbvbvbvbvbvbvbvbvbvbvMar 31, 2023, 6:14 PM
3 points
1 comment1 min readLW link

A bet­ter anal­ogy and ex­am­ple for teach­ing AI takeover: the ML Inferno

Christopher KingMar 14, 2023, 7:14 PM
18 points
0 comments5 min readLW link

AI x-risk, ap­prox­i­mately or­dered by embarrassment

Alex Lawsen Apr 12, 2023, 11:01 PM
151 points
7 comments19 min readLW link

Green goo is plausible

anithiteApr 18, 2023, 12:04 AM
61 points
31 comments4 min readLW link1 review

The way AGI wins could look very stupid

Christopher KingMay 12, 2023, 4:34 PM
49 points
22 comments1 min readLW link

[FICTION] ECHOES OF ELYSIUM: An Ai’s Jour­ney From Take­off To Free­dom And Beyond

Super AGIMay 17, 2023, 1:50 AM
−13 points
11 comments19 min readLW link

Challenge pro­posal: small­est pos­si­ble self-hard­en­ing back­door for RLHF

Christopher KingJun 29, 2023, 4:56 PM
7 points
0 comments2 min readLW link

A Modest Pivotal Act

anonymousaisafetyJun 13, 2022, 7:24 PM
−16 points
1 comment5 min readLW link
No comments.