You don’t know how bad most things are nor pre­cisely how they’re bad.

Solenoid_Entity4 Aug 2024 14:12 UTC
313 points
48 comments5 min readLW link

Would catch­ing your AIs try­ing to es­cape con­vince AI de­vel­op­ers to slow down or un­de­ploy?

Buck26 Aug 2024 16:46 UTC
285 points
69 comments4 min readLW link

Leav­ing MIRI, Seek­ing Funding

abramdemski8 Aug 2024 18:32 UTC
267 points
19 comments2 min readLW link

Prin­ci­ples for the AGI Race

William_S30 Aug 2024 14:29 UTC
244 points
13 comments18 min readLW link

AGI Safety and Align­ment at Google Deep­Mind: A Sum­mary of Re­cent Work

20 Aug 2024 16:22 UTC
220 points
33 comments9 min readLW link

The ‘strong’ fea­ture hy­poth­e­sis could be wrong

lewis smith2 Aug 2024 14:33 UTC
218 points
17 comments17 min readLW link

How I Learned To Stop Trust­ing Pre­dic­tion Mar­kets and Love the Arbitrage

orthonormal6 Aug 2024 2:32 UTC
191 points
26 comments3 min readLW link

WTH is Cere­brolysin, ac­tu­ally?

6 Aug 2024 20:40 UTC
175 points
23 comments17 min readLW link

You can re­move GPT2’s Lay­erNorm by fine-tun­ing for an hour

StefanHex8 Aug 2024 18:33 UTC
161 points
11 comments8 min readLW link

[Question] things that con­fuse me about the cur­rent AI mar­ket.

DMMF28 Aug 2024 13:46 UTC
152 points
28 comments2 min readLW link

Li­a­bil­ity regimes for AI

Ege Erdil19 Aug 2024 1:25 UTC
147 points
34 comments5 min readLW link

The In­for­ma­tion: OpenAI shows ‘Straw­berry’ to feds, races to launch it

Martín Soto27 Aug 2024 23:10 UTC
144 points
15 comments3 min readLW link

Nurs­ing doubts

dynomight30 Aug 2024 2:25 UTC
144 points
20 comments9 min readLW link
(dynomight.net)

Fields that I refer­ence when think­ing about AI takeover prevention

Buck13 Aug 2024 23:08 UTC
143 points
16 comments10 min readLW link
(redwoodresearch.substack.com)

Limi­ta­tions on For­mal Ver­ifi­ca­tion for AI Safety

Andrew Dickson19 Aug 2024 23:03 UTC
134 points
60 comments23 min readLW link

“Can AI Scal­ing Con­tinue Through 2030?”, Epoch AI (yes)

gwern24 Aug 2024 1:40 UTC
128 points
4 comments3 min readLW link
(epochai.org)

Near-mode think­ing on AI

Olli Järviniemi4 Aug 2024 20:47 UTC
127 points
8 comments5 min readLW link

How I started be­liev­ing re­li­gion might ac­tu­ally mat­ter for ra­tio­nal­ity and moral philosophy

zhukeepa23 Aug 2024 17:40 UTC
126 points
41 comments7 min readLW link

Par­a­sites (not a metaphor)

lemonhope8 Aug 2024 20:07 UTC
122 points
17 comments1 min readLW link

In­ves­ti­gat­ing the Chart of the Cen­tury: Why is food so ex­pen­sive?

Maxwell Tabarrok16 Aug 2024 13:21 UTC
121 points
26 comments3 min readLW link
(www.maximum-progress.com)

Please sup­port this blog (with money)

Elizabeth17 Aug 2024 15:30 UTC
110 points
2 comments6 min readLW link
(acesounderglass.com)

Ten ar­gu­ments that AI is an ex­is­ten­tial risk

13 Aug 2024 17:00 UTC
110 points
41 comments7 min readLW link
(blog.aiimpacts.org)

Please stop us­ing mediocre AI art in your posts

Raemon25 Aug 2024 0:13 UTC
109 points
24 comments2 min readLW link

A primer on the cur­rent state of longevity research

Abhishaike Mahajan22 Aug 2024 17:14 UTC
109 points
6 comments14 min readLW link
(www.owlposting.com)

Per­plex­ity wins my AI race

Elizabeth24 Aug 2024 19:20 UTC
107 points
12 comments10 min readLW link
(acesounderglass.com)

Danger, AI Scien­tist, Danger

Zvi15 Aug 2024 22:40 UTC
106 points
9 comments7 min readLW link
(thezvi.wordpress.com)

LLM Ap­pli­ca­tions I Want To See

sarahconstantin19 Aug 2024 21:10 UTC
102 points
5 comments8 min readLW link
(sarahconstantin.substack.com)

Dragon Agnosticism

jefftk1 Aug 2024 17:00 UTC
99 points
73 comments2 min readLW link
(www.jefftk.com)

the Giga Press was a mistake

bhauth21 Aug 2024 4:51 UTC
94 points
26 comments5 min readLW link
(bhauth.com)

It’s time for a self-re­pro­duc­ing machine

Carl Feynman7 Aug 2024 21:52 UTC
93 points
68 comments9 min readLW link

[Question] Am I con­fused about the “ma­lign uni­ver­sal prior” ar­gu­ment?

nostalgebraist27 Aug 2024 23:17 UTC
92 points
33 comments8 min readLW link

SB 1047: Fi­nal Takes and Also AB 3211

Zvi27 Aug 2024 22:10 UTC
92 points
11 comments21 min readLW link
(thezvi.wordpress.com)

Cir­cu­lar Reasoning

abramdemski5 Aug 2024 18:10 UTC
91 points
36 comments8 min readLW link

Defin­ing al­ign­ment research

Richard_Ngo19 Aug 2024 20:42 UTC
91 points
23 comments7 min readLW link

Sin­gu­lar learn­ing the­ory: exercises

Zach Furman30 Aug 2024 20:00 UTC
88 points
5 comments14 min readLW link

Solv­ing ad­ver­sar­ial at­tacks in com­puter vi­sion as a baby ver­sion of gen­eral AI alignment

Stanislav Fort29 Aug 2024 17:17 UTC
87 points
8 comments7 min readLW link

What De­pres­sion Is Like

Sable27 Aug 2024 17:43 UTC
83 points
23 comments4 min readLW link
(affablyevil.substack.com)

Why you should be us­ing a retinoid

GeneSmith19 Aug 2024 3:07 UTC
83 points
59 comments5 min readLW link

Sec­u­lar in­ter­pre­ta­tions of core peren­ni­al­ist claims

zhukeepa25 Aug 2024 23:41 UTC
82 points
32 comments14 min readLW link

Re­lease: Op­ti­mal Weave (P1): A Pro­to­type Co­hab­itive Game

mako yass17 Aug 2024 14:08 UTC
82 points
21 comments7 min readLW link

Quick look: ap­pli­ca­tions of chaos theory

18 Aug 2024 15:00 UTC
79 points
51 comments8 min readLW link
(acesounderglass.com)

Dar­wi­nian Traps and Ex­is­ten­tial Risks

KristianRonn25 Aug 2024 22:37 UTC
76 points
14 comments10 min readLW link

Value frag­ility and AI takeover

Joe Carlsmith5 Aug 2024 21:28 UTC
76 points
5 comments30 min readLW link

A Sim­ple Toy Co­her­ence Theorem

2 Aug 2024 17:47 UTC
74 points
19 comments7 min readLW link

AI for Bio: State Of The Field

sarahconstantin30 Aug 2024 18:00 UTC
73 points
2 comments15 min readLW link
(sarahconstantin.substack.com)

Soft Na­tion­al­iza­tion: how the USG will con­trol AI labs

27 Aug 2024 15:11 UTC
73 points
7 comments21 min readLW link
(www.convergenceanalysis.org)

In Defense of Open-Minded UDT

abramdemski12 Aug 2024 18:27 UTC
72 points
27 comments11 min readLW link

Far­mKind’s Illu­sory Offer

jefftk9 Aug 2024 11:30 UTC
71 points
5 comments3 min readLW link
(www.jefftk.com)

Guide to SB 1047

Zvi20 Aug 2024 13:10 UTC
71 points
18 comments53 min readLW link
(thezvi.wordpress.com)

If we solve al­ign­ment, do we die any­way?

Seth Herd23 Aug 2024 13:13 UTC
70 points
91 comments4 min readLW link