My Mid-Ca­reer Tran­si­tion into Biosecurity

jefftk2 Oct 2023 21:20 UTC
26 points
4 comments2 min readLW link
(www.jefftk.com)

Dall-E 3

p.b.2 Oct 2023 20:33 UTC
37 points
9 comments1 min readLW link
(openai.com)

Thomas Kwa’s MIRI re­search experience

2 Oct 2023 16:42 UTC
172 points
53 comments1 min readLW link

Pop­u­la­tion After a Catastrophe

Stan Pinsent2 Oct 2023 16:06 UTC
3 points
5 comments14 min readLW link

Ex­pec­ta­tions for Gem­ini: hope­fully not a big deal

Maxime Riché2 Oct 2023 15:38 UTC
15 points
5 comments1 min readLW link

A coun­terex­am­ple for mea­surable fac­tor spaces

Matthias G. Mayer2 Oct 2023 15:16 UTC
14 points
0 comments3 min readLW link

Will early trans­for­ma­tive AIs pri­mar­ily use text? [Man­i­fold ques­tion]

Fabien Roger2 Oct 2023 15:05 UTC
16 points
0 comments3 min readLW link

en­ergy land­scapes of experts

bhauth2 Oct 2023 14:08 UTC
41 points
2 comments3 min readLW link
(www.bhauth.com)

Direc­tion of Fit

NicholasKees2 Oct 2023 12:34 UTC
34 points
0 comments3 min readLW link

The 99% prin­ci­ple for per­sonal problems

Kaj_Sotala2 Oct 2023 8:20 UTC
135 points
20 comments2 min readLW link
(kajsotala.fi)

Linkpost: They Stud­ied Dishon­esty. Was Their Work a Lie?

Linch2 Oct 2023 8:10 UTC
91 points
12 comments2 min readLW link
(www.newyorker.com)

Why I got the smal­l­pox vac­cine in 2023

joec2 Oct 2023 5:11 UTC
25 points
6 comments4 min readLW link

In­stru­men­tal Con­ver­gence and hu­man ex­tinc­tion.

Spiritus Dei2 Oct 2023 0:41 UTC
−10 points
3 comments7 min readLW link

Re­vis­it­ing the Man­i­fold Hypothesis

Aidan Rocke1 Oct 2023 23:55 UTC
13 points
19 comments4 min readLW link

AI Align­ment Break­throughs this Week [new sub­stack]

Logan Zoellner1 Oct 2023 22:13 UTC
0 points
8 comments2 min readLW link

[Question] Look­ing for study

Robert Feinstein1 Oct 2023 19:52 UTC
4 points
0 comments1 min readLW link

Join AISafety.info’s Distil­la­tion Hackathon (Oct 6-9th)

smallsilo1 Oct 2023 18:43 UTC
21 points
0 comments2 min readLW link
(forum.effectivealtruism.org)

Fifty Flips

abstractapplic1 Oct 2023 15:30 UTC
31 points
14 comments1 min readLW link
(h-b-p.github.io)

AI Safety Im­pact Mar­kets: Your Char­ity Eval­u­a­tor for AI Safety

Dawn Drescher1 Oct 2023 10:47 UTC
16 points
5 comments1 min readLW link
(impactmarkets.substack.com)

“Ab­sence of Ev­i­dence is Not Ev­i­dence of Ab­sence” As a Limit

transhumanist_atom_understander1 Oct 2023 8:15 UTC
16 points
1 comment2 min readLW link

New Tool: the Resi­d­ual Stream Viewer

AdamYedidia1 Oct 2023 0:49 UTC
32 points
7 comments4 min readLW link
(tinyurl.com)

My Effortless Weight­loss Story: A Quick Runthrough

CuoreDiVetro30 Sep 2023 23:02 UTC
123 points
78 comments9 min readLW link

Ar­gu­ments for moral indefinability

Richard_Ngo30 Sep 2023 22:40 UTC
47 points
16 comments7 min readLW link
(www.thinkingcomplete.com)

Con­di­tion­als All The Way Down

lunatic_at_large30 Sep 2023 21:06 UTC
33 points
2 comments3 min readLW link

Fo­cus­ing your im­pact on short vs long TAI timelines

kuhanj30 Sep 2023 19:34 UTC
4 points
0 comments10 min readLW link

How model edit­ing could help with the al­ign­ment prob­lem

Michael Ripa30 Sep 2023 17:47 UTC
12 points
1 comment15 min readLW link

My sub­mis­sion to the ALTER Prize

Lorxus30 Sep 2023 16:07 UTC
6 points
0 comments1 min readLW link
(www.docdroid.net)

Anki deck for learn­ing the main AI safety orgs, pro­jects, and programs

Bryce Robertson30 Sep 2023 16:06 UTC
2 points
0 comments1 min readLW link

The Lighthaven Cam­pus is open for bookings

habryka30 Sep 2023 1:08 UTC
209 points
18 comments5 min readLW link
(www.lighthaven.space)

Head­phones hook

philh29 Sep 2023 22:50 UTC
21 points
0 comments3 min readLW link
(reasonableapproximation.net)

Paul Chris­ti­ano’s views on “doom” (video ex­plainer)

Michaël Trazzi29 Sep 2023 21:56 UTC
15 points
0 comments1 min readLW link
(youtu.be)

The Retroac­tive Fund­ing Land­scape: In­no­va­tions for Donors and Grantmakers

Dawn Drescher29 Sep 2023 17:39 UTC
13 points
0 comments1 min readLW link
(impactmarkets.substack.com)

Bids To Defer On Value Judgements

johnswentworth29 Sep 2023 17:07 UTC
58 points
6 comments3 min readLW link

An­nounc­ing FAR Labs, an AI safety cowork­ing space

bgold29 Sep 2023 16:52 UTC
95 points
0 comments1 min readLW link

A tool for search­ing ra­tio­nal­ist & EA webs

Daniel_Friedrich29 Sep 2023 15:23 UTC
4 points
0 comments1 min readLW link
(ratsearch.blogspot.com)

Ba­sic Math­e­mat­ics of Pre­dic­tive Coding

Adam Shai29 Sep 2023 14:38 UTC
49 points
6 comments9 min readLW link

“Di­a­mon­doid bac­te­ria” nanobots: deadly threat or dead-end? A nan­otech in­ves­ti­ga­tion

titotal29 Sep 2023 14:01 UTC
154 points
79 comments1 min readLW link
(titotal.substack.com)

Steer­ing sub­sys­tems: ca­pa­bil­ities, agency, and alignment

Seth Herd29 Sep 2023 13:45 UTC
26 points
0 comments8 min readLW link

Ap­ply to Us­able Se­cu­rity Prize by Septem­ber 30

Allison Duettmann29 Sep 2023 13:39 UTC
4 points
0 comments1 min readLW link

List of how peo­ple have be­come more hard-working

Chi Nguyen29 Sep 2023 11:30 UTC
65 points
7 comments1 min readLW link

Re­solv­ing moral un­cer­tainty with randomization

29 Sep 2023 11:23 UTC
7 points
1 comment11 min readLW link

EA Ve­gan Ad­vo­cacy is not truth­seek­ing, and it’s ev­ery­one’s problem

Elizabeth28 Sep 2023 23:30 UTC
317 points
247 comments22 min readLW link
(acesounderglass.com)

Com­pet­i­tive, Co­op­er­a­tive, and Cohabitive

Screwtape28 Sep 2023 23:25 UTC
48 points
12 comments4 min readLW link

The Com­ing Wave

PeterMcCluskey28 Sep 2023 22:59 UTC
25 points
1 comment6 min readLW link
(bayesianinvestor.com)

High-level in­ter­pretabil­ity: de­tect­ing an AI’s objectives

28 Sep 2023 19:30 UTC
69 points
4 comments21 min readLW link

How to Catch an AI Liar: Lie De­tec­tion in Black-Box LLMs by Ask­ing Un­re­lated Questions

28 Sep 2023 18:53 UTC
185 points
38 comments3 min readLW link

Re­spon­si­ble scal­ing policy TLDR

lemonhope28 Sep 2023 18:51 UTC
9 points
0 comments1 min readLW link

Align­ment Work­shop talks

Richard_Ngo28 Sep 2023 18:26 UTC
37 points
1 comment1 min readLW link
(www.alignment-workshop.com)

My Cur­rent Thoughts on the AI Strate­gic Landscape

Jeffrey Heninger28 Sep 2023 17:59 UTC
11 points
28 comments14 min readLW link

My Ar­ro­gant Plan for Alignment

MrArrogant28 Sep 2023 17:51 UTC
2 points
6 comments6 min readLW link