Book re­view: The Pas­sen­ger by Lisa Lutz

KatjaGraceJun 23, 2022, 11:10 PM
12 points
1 comment1 min readLW link
(worldspiritsockpuppet.com)

20 Cri­tiques of AI Safety That I Found on Twitter

dkirmaniJun 23, 2022, 7:23 PM
21 points
16 comments1 min readLW link

The Limits of Automation

milkandcigarettesJun 23, 2022, 6:03 PM
5 points
1 comment5 min readLW link
(milkandcigarettes.com)

[Question] Is CIRL a promis­ing agenda?

Chris_LeongJun 23, 2022, 5:12 PM
28 points
16 comments1 min readLW link

[Link] OpenAI: Learn­ing to Play Minecraft with Video PreTrain­ing (VPT)

Aryeh EnglanderJun 23, 2022, 4:29 PM
53 points
3 comments1 min readLW link

Half-baked AI Safety ideas thread

Aryeh EnglanderJun 23, 2022, 4:11 PM
64 points
63 comments1 min readLW link

Non­profit Boards are Weird

HoldenKarnofskyJun 23, 2022, 2:40 PM
156 points
26 comments20 min readLW link1 review
(www.cold-takes.com)

Covid 6/​23/​22: Un­der Five Alive

ZviJun 23, 2022, 2:00 PM
26 points
9 comments10 min readLW link
(thezvi.wordpress.com)

How do states re­spond to changes in nu­clear risk

NathanBarnardJun 23, 2022, 12:42 PM
8 points
2 comments5 min readLW link

[Question] What’s the con­tin­gency plan if we get AGI to­mor­row?

YitzJun 23, 2022, 3:10 AM
61 points
23 comments1 min readLW link

[Question] What are the best “policy” ap­proaches in wor­lds where al­ign­ment is difficult?

LHAJun 23, 2022, 1:53 AM
1 point
0 comments1 min readLW link

AI Train­ing Should Allow Opt-Out

alyssavanceJun 23, 2022, 1:33 AM
76 points
13 comments6 min readLW link

Loose thoughts on AGI risk

YitzJun 23, 2022, 1:02 AM
7 points
3 comments1 min readLW link

Air Con­di­tioner Test Re­sults & Discussion

johnswentworthJun 22, 2022, 10:26 PM
82 points
42 comments6 min readLW link

An­nounc­ing the LessWrong Cu­rated Podcast

Jun 22, 2022, 10:16 PM
137 points
27 comments1 min readLW link

Google’s new text-to-image model—Parti, a demon­stra­tion of scal­ing benefits

KaydenJun 22, 2022, 8:00 PM
32 points
4 comments1 min readLW link

Build­ing an Epistemic Sta­tus Tracker

rcuJun 22, 2022, 6:57 PM
7 points
8 comments1 min readLW link

Con­fu­sion about neu­ro­science/​cog­ni­tive sci­ence as a dan­ger for AI Alignment

Samuel NellessenJun 22, 2022, 5:59 PM
3 points
1 comment3 min readLW link
(snellessen.com)

[Question] How do I use caf­feine op­ti­mally?

randomstringJun 22, 2022, 5:59 PM
18 points
31 comments1 min readLW link

Make learn­ing a reality

Dalton MaberyJun 22, 2022, 3:58 PM
13 points
2 comments1 min readLW link

Reflec­tion Mechanisms as an Align­ment tar­get: A survey

Jun 22, 2022, 3:05 PM
32 points
1 comment14 min readLW link

House Phone

jefftkJun 22, 2022, 2:20 PM
15 points
2 comments1 min readLW link
(www.jefftk.com)

How to Vi­su­al­ize Bayesianism

David UdellJun 22, 2022, 1:57 PM
9 points
2 comments3 min readLW link

[Question] Are there spaces for ex­tremely short-form ra­tio­nal­ity con­tent?

Aleksi LiimatainenJun 22, 2022, 10:39 AM
5 points
1 comment1 min readLW link

Sols­tice Movie Re­view: Sum­mer Wars

SebastianG Jun 22, 2022, 1:09 AM
22 points
6 comments1 min readLW link

Se­cu­rity Mind­set: Les­sons from 20+ years of Soft­ware Se­cu­rity Failures Rele­vant to AGI Alignment

elspoodJun 21, 2022, 11:55 PM
362 points
42 comments7 min readLW link1 review

A Quick List of Some Prob­lems in AI Align­ment As A Field

Nicholas / Heather KrossJun 21, 2022, 11:23 PM
75 points
12 comments6 min readLW link
(www.thinkingmuchbetter.com)

[Question] What is the differ­ence be­tween AI mis­al­ign­ment and bad pro­gram­ming?

puzzleGuzzleJun 21, 2022, 9:52 PM
6 points
2 comments1 min readLW link

What I mean by the phrase “get­ting in­ti­mate with re­al­ity”

LuiseJun 21, 2022, 7:42 PM
6 points
0 comments2 min readLW link
(forum.effectivealtruism.org)

What I mean by the phrase “tak­ing ideas se­ri­ously”

LuiseJun 21, 2022, 7:42 PM
5 points
2 comments1 min readLW link
(forum.effectivealtruism.org)

Hy­dropho­bic Glasses Coat­ing Review

jefftkJun 21, 2022, 6:00 PM
16 points
6 comments1 min readLW link
(www.jefftk.com)

Progress links and tweets, 2022-06-20

jasoncrawfordJun 21, 2022, 5:12 PM
12 points
2 comments1 min readLW link
(rootsofprogress.org)

De­bat­ing Whether AI is Con­scious Is A Dis­trac­tion from Real Problems

sidhe_theyJun 21, 2022, 4:56 PM
2 points
10 comments1 min readLW link
(techpolicy.press)

Miti­gat­ing the dam­age from un­al­igned ASI by co­op­er­at­ing with aliens that don’t ex­ist yet

MSRayneJun 21, 2022, 4:12 PM
−8 points
7 comments6 min readLW link

The in­or­di­nately slow spread of good AGI con­ver­sa­tions in ML

Rob BensingerJun 21, 2022, 4:09 PM
173 points
62 comments8 min readLW link

Get­ting from an un­al­igned AGI to an al­igned AGI?

Tor Økland BarstadJun 21, 2022, 12:36 PM
13 points
7 comments9 min readLW link

Com­mon but ne­glected risk fac­tors that may let you get Paxlovid

DirectedEvolutionJun 21, 2022, 7:34 AM
29 points
8 comments4 min readLW link

Dag­ger of De­tect Evil

lsusrJun 21, 2022, 6:23 AM
45 points
22 comments3 min readLW link

[Question] How easy/​fast is it for a AGI to hack com­put­ers/​a hu­man brain?

Noosphere89Jun 21, 2022, 12:34 AM
0 points
1 comment1 min readLW link

[Question] What is the most prob­a­ble AI?

Zeruel017Jun 20, 2022, 11:26 PM
−2 points
0 comments3 min readLW link

Eval­u­at­ing a Corsi-Rosen­thal Filter Cube

jefftkJun 20, 2022, 7:40 PM
13 points
4 comments1 min readLW link
(www.jefftk.com)

Sur­vey re AIS/​LTism office in NYC

RyanCareyJun 20, 2022, 7:21 PM
7 points
0 comments1 min readLW link

Is This Thing Sen­tient, Y/​N?

Thane RuthenisJun 20, 2022, 6:37 PM
4 points
10 comments7 min readLW link

Steam

abramdemskiJun 20, 2022, 5:38 PM
149 points
13 comments5 min readLW link1 review

Parable: The Bomb that doesn’t Explode

Lone PineJun 20, 2022, 4:41 PM
14 points
5 comments2 min readLW link

On cor­rigi­bil­ity and its basin

Donald HobsonJun 20, 2022, 4:33 PM
16 points
3 comments2 min readLW link

An­nounc­ing the DWATV Discord

ZviJun 20, 2022, 3:50 PM
20 points
9 comments1 min readLW link
(thezvi.wordpress.com)

Key Papers in Lan­guage Model Safety

aogJun 20, 2022, 3:00 PM
40 points
1 comment22 min readLW link

Re­la­tion­ship Ad­vice Repository

RubyJun 20, 2022, 2:39 PM
109 points
36 comments38 min readLW link

Adap­ta­tion Ex­ecu­tors and the Telos Margin

PlinthistJun 20, 2022, 1:06 PM
2 points
8 comments5 min readLW link