How I got 4.2M YouTube views with­out mak­ing a sin­gle video

Closed Limelike CurvesSep 3, 2024, 3:52 AM
392 points
36 comments1 min readLW link

The Great Data In­te­gra­tion Schlep

sarahconstantinSep 13, 2024, 3:40 PM
271 points
16 comments9 min readLW link
(sarahconstantin.substack.com)

Laz­i­ness death spirals

PatrickDFarleySep 19, 2024, 3:58 PM
262 points
40 comments8 min readLW link

the case for CoT un­faith­ful­ness is overstated

nostalgebraistSep 29, 2024, 10:07 PM
259 points
43 comments11 min readLW link

The Best Lay Ar­gu­ment is not a Sim­ple English Yud Essay

J BostockSep 10, 2024, 5:34 PM
253 points
15 comments5 min readLW link

Ex­plore More: A Bag of Tricks to Keep Your Life on the Rails

Shoshannah TekofskySep 28, 2024, 9:38 PM
235 points
19 comments11 min readLW link
(shoshanigans.substack.com)

“Slow” take­off is a ter­rible term for “maybe even faster take­off, ac­tu­ally”

RaemonSep 28, 2024, 11:38 PM
217 points
69 comments1 min readLW link

Pay Risk Eval­u­a­tors in Cash, Not Equity

Adam SchollSep 7, 2024, 2:37 AM
212 points
19 comments1 min readLW link

The Sun is big, but su­per­in­tel­li­gences will not spare Earth a lit­tle sunlight

Eliezer YudkowskySep 23, 2024, 3:39 AM
203 points
142 comments13 min readLW link

Cry­on­ics is free

Mati_RoySep 29, 2024, 5:58 PM
198 points
43 comments2 min readLW link

A ba­sic sys­tems ar­chi­tec­ture for AI agents that do au­tonomous research

BuckSep 23, 2024, 1:58 PM
189 points
16 comments8 min readLW link

Skills from a year of Pur­pose­ful Ra­tion­al­ity Practice

RaemonSep 18, 2024, 2:05 AM
187 points
18 comments7 min readLW link

Strug­gling like a Shadowmoth

RaemonSep 24, 2024, 12:47 AM
183 points
38 comments7 min readLW link

Con­tra pa­pers claiming su­per­hu­man AI forecasting

Sep 12, 2024, 6:10 PM
182 points
16 comments7 min readLW link

[Question] Why is o1 so de­cep­tive?

abramdemskiSep 27, 2024, 5:27 PM
180 points
24 comments3 min readLW link

Did Christo­pher Hitchens change his mind about wa­ter­board­ing?

Isaac KingSep 15, 2024, 8:28 AM
171 points
22 comments7 min readLW link

My takes on SB-1047

leogaoSep 9, 2024, 6:38 PM
151 points
8 comments4 min readLW link

OpenAI o1

Zach Stein-PerlmanSep 12, 2024, 5:30 PM
147 points
41 comments1 min readLW link

Stanis­lav Petrov Quar­terly Perfor­mance Review

Ricki HeicklenSep 26, 2024, 9:20 PM
147 points
3 comments5 min readLW link
(bayesshammai.substack.com)

That Alien Mes­sage—The Animation

WriterSep 7, 2024, 2:53 PM
144 points
10 comments8 min readLW link
(youtu.be)

The Check­list: What Suc­ceed­ing at AI Safety Will In­volve

Sam BowmanSep 3, 2024, 6:18 PM
142 points
49 comments22 min readLW link
(sleepinyourhat.github.io)

Sur­vey: How Do Elite Chi­nese Stu­dents Feel About the Risks of AI?

Nick CorvinoSep 2, 2024, 6:11 PM
141 points
13 comments10 min readLW link

[Com­pleted] The 2024 Petrov Day Scenario

Sep 26, 2024, 8:08 AM
136 points
114 comments5 min readLW link

My Num­ber 1 Episte­mol­ogy Book Recom­men­da­tion: In­vent­ing Temperature

adamShimiSep 8, 2024, 2:30 PM
121 points
18 comments3 min readLW link
(epistemologicalfascinations.substack.com)

Why I funded PIBBSS

Ryan KiddSep 15, 2024, 7:56 PM
115 points
21 comments3 min readLW link

You can, in fact, bam­boo­zle an un­al­igned AI into spar­ing your life

David MatolcsiSep 29, 2024, 4:59 PM
112 points
173 comments27 min readLW link

What hap­pens if you pre­sent 500 peo­ple with an ar­gu­ment that AI is risky?

Sep 4, 2024, 4:40 PM
109 points
8 comments3 min readLW link
(blog.aiimpacts.org)

Back­doors as an anal­ogy for de­cep­tive alignment

Sep 6, 2024, 3:30 PM
104 points
2 comments8 min readLW link
(www.alignment.org)

Re­fac­tor­ing cry­on­ics as struc­tural brain preservation

Andy_McKenzieSep 11, 2024, 6:36 PM
101 points
14 comments3 min readLW link

[Question] What are the best ar­gu­ments for/​against AIs be­ing “slightly ‘nice’”?

RaemonSep 24, 2024, 2:00 AM
99 points
61 comments31 min readLW link

The Oblique­ness Thesis

jessicataSep 19, 2024, 12:26 AM
95 points
19 comments17 min readLW link

2024 Petrov Day Retrospective

Sep 28, 2024, 9:30 PM
93 points
25 comments10 min readLW link

Ex­e­cutable philos­o­phy as a failed to­tal­iz­ing meta-worldview

jessicataSep 4, 2024, 10:50 PM
93 points
40 comments4 min readLW link
(unstableontology.com)

[In­tu­itive self-mod­els] 1. Preliminaries

Steven ByrnesSep 19, 2024, 1:45 PM
91 points
23 comments15 min readLW link

How to pre­vent col­lu­sion when us­ing un­trusted mod­els to mon­i­tor each other

BuckSep 25, 2024, 6:58 PM
88 points
11 comments22 min readLW link

GPT-o1

ZviSep 16, 2024, 1:40 PM
86 points
34 comments46 min readLW link
(thezvi.wordpress.com)

OpenAI o1, Llama 4, and AlphaZero of LLMs

Vladimir_NesovSep 14, 2024, 9:27 PM
83 points
25 comments1 min readLW link

[In­tu­itive self-mod­els] 2. Con­scious Awareness

Steven ByrnesSep 25, 2024, 1:29 PM
82 points
60 comments16 min readLW link

AI #83: The Mask Comes Off

ZviSep 26, 2024, 12:00 PM
82 points
20 comments36 min readLW link
(thezvi.wordpress.com)

Not ev­ery ac­com­mo­da­tion is a Curb Cut Effect: The Handi­capped Park­ing Effect, the Clap­per Effect, and more

Michael CohnSep 15, 2024, 5:27 AM
81 points
39 comments10 min readLW link
(perplexedguide.net)

AI, cen­tral­iza­tion, and the One Ring

owencbSep 13, 2024, 2:00 PM
80 points
12 comments8 min readLW link
(strangecities.substack.com)

Is “su­per­hu­man” AI fore­cast­ing BS? Some ex­per­i­ments on the “539″ bot from the Cen­tre for AI Safety

titotalSep 18, 2024, 1:07 PM
79 points
3 commentsLW link
(open.substack.com)

My 10-year ret­ro­spec­tive on try­ing SSRIs

Kaj_SotalaSep 22, 2024, 8:30 PM
77 points
9 comments2 min readLW link
(kajsotala.fi)

The case for a nega­tive al­ign­ment tax

Sep 18, 2024, 6:33 PM
75 points
20 comments7 min readLW link

[Paper] A is for Ab­sorp­tion: Study­ing Fea­ture Split­ting and Ab­sorp­tion in Sparse Autoencoders

Sep 25, 2024, 9:31 AM
73 points
16 comments3 min readLW link
(arxiv.org)

Ex­cerpts from “A Reader’s Man­i­festo”

Arjun PanicksserySep 6, 2024, 10:37 PM
72 points
1 comment13 min readLW link
(arjunpanickssery.substack.com)

Adam Op­ti­mizer Causes Priv­ileged Ba­sis in Trans­former LM Resi­d­ual Stream

Sep 6, 2024, 5:55 PM
70 points
7 comments4 min readLW link

In­ves­ti­gat­ing an in­surance-for-AI startup

Sep 21, 2024, 3:29 PM
70 points
0 comments16 min readLW link
(www.strataoftheworld.com)

Es­ti­mat­ing Tail Risk in Neu­ral Networks

Mark XuSep 13, 2024, 8:00 PM
68 points
9 comments23 min readLW link
(www.alignment.org)

How to Give in to Threats (with­out in­cen­tiviz­ing them)

Mikhail SaminSep 12, 2024, 3:55 PM
67 points
30 comments5 min readLW link