Im­pe­rial Rus­sia was do­ing fine with­out the Soviets

Davis KedroskyJul 5, 2022, 10:24 PM
6 points
3 comments14 min readLW link
(daviskedrosky.substack.com)

A Pat­tern Lan­guage For Rationality

VaniverJul 5, 2022, 7:08 PM
75 points
14 comments15 min readLW link

How to de­stroy the uni­verse with a hypercomputer

Trevor CappalloJul 5, 2022, 7:05 PM
2 points
3 comments1 min readLW link

The cu­ri­ous case of Pretty Good hu­man in­ner/​outer alignment

PavleMihaJul 5, 2022, 7:04 PM
41 points
45 comments4 min readLW link

When is it ap­pro­pri­ate to use statis­ti­cal mod­els and prob­a­bil­ities for de­ci­sion mak­ing ?

Younes KamelJul 5, 2022, 12:34 PM
10 points
7 comments4 min readLW link
(youneskamel.substack.com)

Goal Factoring

CFAR!DuncanJul 5, 2022, 7:10 AM
92 points
2 comments8 min readLW link

As­sorted thoughts about ab­strac­tion

Adam ZernerJul 5, 2022, 6:40 AM
16 points
9 comments7 min readLW link

[AN #172] Sorry for the long hi­a­tus!

Rohin ShahJul 5, 2022, 6:20 AM
54 points
0 comments3 min readLW link
(mailchi.mp)

Out­line: The Rec­tify­ing of Maps

hamnoxJul 5, 2022, 5:14 AM
7 points
0 comments2 min readLW link

[Question] Seek­ing opinions on the cur­rent and for­ward state of cryp­tocur­ren­cies.

jmhJul 5, 2022, 5:01 AM
6 points
6 comments1 min readLW link

ITT-pass­ing and ci­vil­ity are good; “char­ity” is bad; steel­man­ning is niche

Rob BensingerJul 5, 2022, 12:15 AM
163 points
36 comments6 min readLW link1 review

Please help us com­mu­ni­cate AI xrisk. It could save the world.

otto.bartenJul 4, 2022, 9:47 PM
4 points
7 comments2 min readLW link

Bench­mark for suc­cess­ful con­cept ex­trap­o­la­tion/​avoid­ing goal misgeneralization

Stuart_ArmstrongJul 4, 2022, 8:48 PM
82 points
12 comments4 min readLW link

Pro­ce­du­ral Ex­ec­u­tive Func­tion, Part 1

DaystarEldJul 4, 2022, 6:51 PM
50 points
8 comments14 min readLW link
(daystareld.com)

An­thropic’s SoLU (Soft­max Lin­ear Unit)

Joel BurgetJul 4, 2022, 6:38 PM
21 points
1 comment4 min readLW link
(transformer-circuits.pub)

Book Re­view: The Righ­teous Mind

ErnestScribblerJul 4, 2022, 5:45 PM
33 points
8 comments35 min readLW link

My Most Likely Rea­son to Die Young is AI X-Risk

AISafetyIsNotLongtermistJul 4, 2022, 5:08 PM
61 points
24 comments4 min readLW link
(forum.effectivealtruism.org)

Is Gen­eral In­tel­li­gence “Com­pact”?

DragonGodJul 4, 2022, 1:27 PM
27 points
6 comments22 min readLW link

Re­mak­ing Effi­cien­tZero (as best I can)

HoagyJul 4, 2022, 11:03 AM
36 points
9 comments22 min readLW link

We Need a Con­soli­dated List of Bad AI Align­ment Solutions

DoubleJul 4, 2022, 6:54 AM
9 points
14 comments1 min readLW link

AI Fore­cast­ing: One Year In

jsteinhardtJul 4, 2022, 5:10 AM
132 points
12 comments6 min readLW link
(bounded-regret.ghost.io)

A com­pressed take on re­cent disagreements

kmanJul 4, 2022, 4:39 AM
33 points
9 comments1 min readLW link

New US Se­nate Bill on X-Risk Miti­ga­tion [Linkpost]

Evan R. MurphyJul 4, 2022, 1:25 AM
35 points
12 comments1 min readLW link
(www.hsgac.senate.gov)

Monthly Shorts 6/​22

CelerJul 3, 2022, 11:40 PM
5 points
2 comments5 min readLW link
(keller.substack.com)

De­ci­sion the­ory and dy­namic inconsistency

paulfchristianoJul 3, 2022, 10:20 PM
80 points
33 comments10 min readLW link
(sideways-view.com)

Five routes of ac­cess to sci­en­tific literature

DirectedEvolutionJul 3, 2022, 8:53 PM
13 points
4 comments6 min readLW link

Toni Kurz and the In­san­ity of Climb­ing Mountains

GeneSmithJul 3, 2022, 8:51 PM
270 points
67 comments11 min readLW link2 reviews

Won­der and The Golden AI Rule

JeffreyKJul 3, 2022, 6:21 PM
0 points
4 comments6 min readLW link

Na­ture ab­hors an im­mutable repli­ca­tor… usually

MSRayneJul 3, 2022, 3:08 PM
28 points
10 comments3 min readLW link

Post hoc jus­tifi­ca­tions as Com­pres­sion Algorithm

Johannes C. MayerJul 3, 2022, 5:02 AM
8 points
0 comments1 min readLW link

SOMA—A story about Consciousness

Johannes C. MayerJul 3, 2022, 4:46 AM
10 points
0 comments1 min readLW link
(www.youtube.com)

Sex­ual self-acceptance

Johannes C. MayerJul 3, 2022, 4:26 AM
11 points
6 comments1 min readLW link

Dono­hue, Le­vitt, Roe, and Wade: T-minus 20 years to a mas­sive crime wave?

Paul LoganJul 3, 2022, 3:03 AM
−24 points
6 comments3 min readLW link
(laulpogan.substack.com)

Can we achieve AGI Align­ment by bal­anc­ing mul­ti­ple hu­man ob­jec­tives?

Ben SmithJul 3, 2022, 2:51 AM
11 points
1 comment4 min readLW link

Trig­ger-Ac­tion Planning

CFAR!DuncanJul 3, 2022, 1:42 AM
86 points
14 comments13 min readLW link2 reviews

[Question] Which one of these two aca­demic routes should I take to end up in AI Safety?

Martín SotoJul 3, 2022, 1:05 AM
5 points
2 comments1 min readLW link

Naive Hy­pothe­ses on AI Alignment

Shoshannah TekofskyJul 2, 2022, 7:03 PM
98 points
29 comments5 min readLW link

The Tree of Life: Stan­ford AI Align­ment The­ory of Change

Gabe MJul 2, 2022, 6:36 PM
25 points
0 comments14 min readLW link

Fol­low along with Columbia EA’s Ad­vanced AI Safety Fel­low­ship!

RohanSJul 2, 2022, 5:45 PM
3 points
0 comments2 min readLW link
(forum.effectivealtruism.org)

Wel­come to Analo­gia! (Chap­ter 7)

Justin BullockJul 2, 2022, 5:04 PM
5 points
0 comments11 min readLW link

[Question] What about tran­shu­mans and be­yond?

AlignmentMirrorJul 2, 2022, 1:58 PM
7 points
6 comments1 min readLW link

Goal-di­rect­ed­ness: tack­ling complexity

Morgan_RogersJul 2, 2022, 1:51 PM
8 points
0 comments38 min readLW link

Liter­a­ture recom­men­da­tions July 2022

ChristianKlJul 2, 2022, 9:14 AM
17 points
9 comments1 min readLW link

Deon­tolog­i­cal Evil

lsusrJul 2, 2022, 6:57 AM
44 points
4 comments2 min readLW link

Could an AI Align­ment Sand­box be use­ful?

Michael SoareverixJul 2, 2022, 5:06 AM
2 points
1 comment1 min readLW link

Five views of Bayes’ Theorem

Adam ScherlisJul 2, 2022, 2:25 AM
38 points
4 comments1 min readLW link

[Linkpost] Ex­is­ten­tial Risk Anal­y­sis in Em­piri­cal Re­search Papers

Dan HJul 2, 2022, 12:09 AM
40 points
0 comments1 min readLW link
(arxiv.org)

Agenty AGI – How Tempt­ing?

PeterMcCluskeyJul 1, 2022, 11:40 PM
22 points
3 comments5 min readLW link
(www.bayesianinvestor.com)

AXRP Epi­sode 16 - Prepar­ing for De­bate AI with Ge­offrey Irving

DanielFilanJul 1, 2022, 10:20 PM
20 points
0 comments37 min readLW link

[Question] Ex­am­ples of prac­ti­cal im­pli­ca­tions of Judea Pearl’s Causal­ity work

ChristianKlJul 1, 2022, 8:58 PM
23 points
6 comments1 min readLW link