GPTs are Pre­dic­tors, not Imitators

Eliezer YudkowskyApr 8, 2023, 7:59 PM
416 points
100 comments3 min readLW link3 reviews

LW Team is ad­just­ing mod­er­a­tion policy

RaemonApr 4, 2023, 8:41 PM
304 points
185 comments3 min readLW link

Hooray for step­ping out of the limelight

So8resApr 1, 2023, 2:45 AM
284 points
26 comments1 min readLW link

Notes on Teach­ing in Prison

jsdApr 19, 2023, 1:53 AM
274 points
13 comments12 min readLW link

Paus­ing AI Devel­op­ments Isn’t Enough. We Need to Shut it All Down

Eliezer YudkowskyApr 8, 2023, 12:36 AM
268 points
44 comments12 min readLW link1 review

[SEE NEW EDITS] No, *You* Need to Write Clearer

Nicholas / Heather KrossApr 29, 2023, 5:04 AM
262 points
65 comments5 min readLW link
(www.thinkingmuchbetter.com)

My Assess­ment of the Chi­nese AI Safety Community

Lao MeinApr 25, 2023, 4:21 AM
252 points
94 comments3 min readLW link

My views on “doom”

paulfchristianoApr 27, 2023, 5:50 PM
250 points
37 comments2 min readLW link1 review
(ai-alignment.com)

On AutoGPT

ZviApr 13, 2023, 12:30 PM
248 points
47 comments20 min readLW link
(thezvi.wordpress.com)

Policy dis­cus­sions fol­low strong con­tex­tu­al­iz­ing norms

Richard_NgoApr 1, 2023, 11:51 PM
231 points
61 comments3 min readLW link

Catch­ing the Eye of Sauron

Casey B.Apr 7, 2023, 12:40 AM
221 points
68 comments4 min readLW link

Orthog­o­nal: A new agent foun­da­tions al­ign­ment organization

Tamsin LeakeApr 19, 2023, 8:17 PM
217 points
4 comments1 min readLW link
(orxl.org)

Eliezer Yud­kowsky’s Let­ter in Time Magazine

ZviApr 5, 2023, 6:00 PM
214 points
86 comments14 min readLW link
(thezvi.wordpress.com)

Gi­ant (In)scrutable Ma­tri­ces: (Maybe) the Best of All Pos­si­ble Worlds

1a3ornApr 4, 2023, 5:39 PM
208 points
38 comments5 min readLW link1 review

Evolu­tion pro­vides no ev­i­dence for the sharp left turn

Quintin PopeApr 11, 2023, 6:43 PM
206 points
65 comments15 min readLW link1 review

If in­ter­pretabil­ity re­search goes well, it may get dangerous

So8resApr 3, 2023, 9:48 PM
202 points
11 comments2 min readLW link

Killing Socrates

Duncan Sabien (Deactivated)Apr 11, 2023, 10:28 AM
195 points
146 comments8 min readLW link1 review

The ‘ pe­ter­todd’ phenomenon

mwatkinsApr 15, 2023, 12:59 AM
192 points
50 comments38 min readLW link1 review

Tran­script and Brief Re­sponse to Twit­ter Con­ver­sa­tion be­tween Yann LeCunn and Eliezer Yudkowsky

ZviApr 26, 2023, 1:10 PM
190 points
51 comments10 min readLW link
(thezvi.wordpress.com)

The ba­sic rea­sons I ex­pect AGI ruin

Rob BensingerApr 18, 2023, 3:37 AM
189 points
73 comments14 min readLW link

A re­port about LessWrong karma volatility from a differ­ent universe

Ben PaceApr 1, 2023, 9:48 PM
181 points
7 comments1 min readLW link

Talk­ing pub­li­cly about AI risk

Jan_KulveitApr 21, 2023, 11:28 AM
180 points
9 comments6 min readLW link

[April Fools’] Defini­tive con­fir­ma­tion of shard theory

TurnTroutApr 1, 2023, 7:27 AM
170 points
8 comments2 min readLW link

Tun­ing your Cog­ni­tive Strategies

Apr 27, 2023, 8:32 PM
169 points
59 comments9 min readLW link1 review
(bewelltuned.com)

Davi­dad’s Bold Plan for Align­ment: An In-Depth Explanation

Apr 19, 2023, 4:09 PM
168 points
40 comments21 min readLW link2 reviews

The Brain is Not Close to Ther­mo­dy­namic Limits on Computation

DaemonicSigilApr 24, 2023, 8:21 AM
167 points
58 comments5 min readLW link

Agen­tized LLMs will change the al­ign­ment landscape

Seth HerdApr 9, 2023, 2:29 AM
160 points
102 comments3 min readLW link1 review

AI doom from an LLM-plateau-ist perspective

Steven ByrnesApr 27, 2023, 1:58 PM
158 points
24 comments6 min readLW link

grey goo is unlikely

bhauthApr 17, 2023, 1:59 AM
157 points
123 comments9 min readLW link2 reviews
(bhauth.com)

A fresh­man year dur­ing the AI midgame: my ap­proach to the next year

BuckApr 14, 2023, 12:38 AM
154 points
15 commentsLW link1 review

The self-un­al­ign­ment problem

Apr 14, 2023, 12:10 PM
154 points
24 comments10 min readLW link

Shut­ting down AI is not enough. We need to de­stroy all tech­nol­ogy.

Matthew BarnettApr 1, 2023, 9:03 PM
152 points
36 comments1 min readLW link

AI x-risk, ap­prox­i­mately or­dered by embarrassment

Alex Lawsen Apr 12, 2023, 11:01 PM
151 points
7 comments19 min readLW link

Could a su­per­in­tel­li­gence de­duce gen­eral rel­a­tivity from a fal­ling ap­ple? An investigation

titotalApr 23, 2023, 12:49 PM
148 points
39 comments9 min readLW link

The Learn­ing-The­o­retic Agenda: Sta­tus 2023

Vanessa KosoyApr 19, 2023, 5:21 AM
144 points
21 comments56 min readLW link3 reviews

Con­sider The Hand Axe

ymeskhoutApr 8, 2023, 1:31 AM
142 points
16 comments6 min readLW link

Re­quest to AGI or­ga­ni­za­tions: Share your views on paus­ing AI progress

Apr 11, 2023, 5:30 PM
141 points
11 comments1 min readLW link

But why would the AI kill us?

So8resApr 17, 2023, 6:42 PM
138 points
96 comments2 min readLW link

Four mind­set dis­agree­ments be­hind ex­is­ten­tial risk dis­agree­ments in ML

Rob BensingerApr 11, 2023, 4:53 AM
136 points
12 commentsLW link

AI Sum­mer Harvest

Cleo NardoApr 4, 2023, 3:35 AM
130 points
10 comments1 min readLW link

Mis­gen­er­al­iza­tion as a misnomer

So8resApr 6, 2023, 8:43 PM
129 points
22 comments4 min readLW link

Good­hart’s Law in­side the hu­man mind

Kaj_SotalaApr 17, 2023, 1:48 PM
125 points
13 comments16 min readLW link

$250 prize for check­ing Jake Can­nell’s Brain Efficiency

Alexander Gietelink OldenzielApr 26, 2023, 4:21 PM
123 points
170 comments2 min readLW link

[New LW Fea­ture] “De­bates”

Apr 1, 2023, 7:00 AM
121 points
35 comments1 min readLW link

Deep learn­ing mod­els might be se­cretly (al­most) linear

berenApr 24, 2023, 6:43 PM
117 points
29 comments4 min readLW link

Fi­nan­cial Times: We must slow down the race to God-like AI

trevorApr 13, 2023, 7:55 PM
113 points
17 comments16 min readLW link
(www.ft.com)

10 rea­sons why lists of 10 rea­sons might be a win­ning strategy

trevorApr 6, 2023, 9:24 PM
110 points
7 comments1 min readLW link

Shap­ley Value At­tri­bu­tion in Chain of Thought

leogaoApr 14, 2023, 5:56 AM
106 points
7 comments4 min readLW link

Should we pub­lish mechanis­tic in­ter­pretabil­ity re­search?

Apr 21, 2023, 4:19 PM
106 points
40 comments13 min readLW link

How could you pos­si­bly choose what an AI wants?

So8resApr 19, 2023, 5:08 PM
106 points
19 comments1 min readLW link