[Question] What are your fa­vorite posts, pod­cast epi­sodes, and recorded talks, on AI timelines, or fac­tors that would in­fluence AI timelines?

nonzerosum2 Nov 2023 22:42 UTC
2 points
0 comments1 min readLW link

One Day Sooner

Screwtape2 Nov 2023 19:00 UTC
106 points
7 comments8 min readLW link

Pro­pa­ganda or Science: A Look at Open Source AI and Bioter­ror­ism Risk

1a3orn2 Nov 2023 18:20 UTC
193 points
79 comments23 min readLW link

AI #36: In the Background

Zvi2 Nov 2023 18:00 UTC
45 points
5 comments37 min readLW link
(thezvi.wordpress.com)

Doubt Certainty

RationalDino2 Nov 2023 17:43 UTC
4 points
13 comments3 min readLW link

Say­ing the quiet part out loud: trad­ing off x-risk for per­sonal immortality

disturbance2 Nov 2023 17:43 UTC
83 points
89 comments5 min readLW link

Mech In­terp Challenge: Novem­ber—De­ci­pher­ing the Cu­mu­la­tive Sum Model

CallumMcDougall2 Nov 2023 17:10 UTC
18 points
2 comments2 min readLW link

Es­ti­mat­ing effec­tive di­men­sion­al­ity of MNIST models

Arjun Panickssery2 Nov 2023 14:13 UTC
41 points
3 comments1 min readLW link

Aver­ages and sam­ple sizes

mruwnik2 Nov 2023 9:52 UTC
15 points
2 comments8 min readLW link

ACX/​LW/​EA crossover meetup

RasmusHB2 Nov 2023 5:57 UTC
2 points
0 comments1 min readLW link

Up­com­ing Feed­back Op­por­tu­nity on Dual-Use Foun­da­tion Models

Chris_Leong2 Nov 2023 4:28 UTC
3 points
0 comments1 min readLW link

Public Weights?

jefftk2 Nov 2023 2:50 UTC
49 points
19 comments3 min readLW link
(www.jefftk.com)

[Question] Should peo­ple build pro­duc­ti­za­tions of open source AI mod­els?

lc2 Nov 2023 1:26 UTC
23 points
0 comments1 min readLW link

Sin­gu­lar learn­ing the­ory and bridg­ing from ML to brain emulations

1 Nov 2023 21:31 UTC
26 points
16 comments29 min readLW link

My thoughts on the so­cial re­sponse to AI risk

Matthew Barnett1 Nov 2023 21:17 UTC
157 points
37 comments10 min readLW link

Re­ac­tions to the Ex­ec­u­tive Order

Zvi1 Nov 2023 20:40 UTC
77 points
4 comments29 min readLW link
(thezvi.wordpress.com)

Dario Amodei’s pre­pared re­marks from the UK AI Safety Sum­mit, on An­thropic’s Re­spon­si­ble Scal­ing Policy

Zac Hatfield-Dodds1 Nov 2023 18:10 UTC
85 points
1 comment4 min readLW link
(www.anthropic.com)

Book Re­view: Deter­mined by Sapolsky

Kailuo Wang1 Nov 2023 17:37 UTC
1 point
0 comments7 min readLW link

AI Align­ment: A Com­pre­hen­sive Survey

Stephen McAleer1 Nov 2023 17:35 UTC
15 points
1 comment1 min readLW link
(arxiv.org)

A list of all the dead­lines in Bi­den’s Ex­ec­u­tive Order on AI

Valentin Baltadzhiev1 Nov 2023 17:14 UTC
26 points
2 comments11 min readLW link

2023 LessWrong Com­mu­nity Cen­sus, Re­quest for Comments

Screwtape1 Nov 2023 16:32 UTC
43 points
37 comments2 min readLW link

[Question] Snap­shot of nar­ra­tives and frames against reg­u­lat­ing AI

Jan_Kulveit1 Nov 2023 16:30 UTC
36 points
19 comments3 min readLW link

Com­men­sal Institutions

Sable1 Nov 2023 16:01 UTC
8 points
12 comments4 min readLW link
(affablyevil.substack.com)

ChatGPT’s On­tolog­i­cal Land­scape

Bill Benzon1 Nov 2023 15:12 UTC
7 points
0 comments4 min readLW link

On the Ex­ec­u­tive Order

Zvi1 Nov 2023 14:20 UTC
100 points
4 comments30 min readLW link
(thezvi.wordpress.com)

Chi­nese sci­en­tists ac­knowl­edge xrisk & call for in­ter­na­tional reg­u­la­tory body [Linkpost]

Akash1 Nov 2023 13:28 UTC
44 points
4 comments1 min readLW link
(www.ft.com)

[Question] Fore­cast­ing Ques­tions: What do you want to pre­dict on AI?

Nathan Young1 Nov 2023 13:17 UTC
7 points
2 comments1 min readLW link

Mis­sion Im­pos­si­ble: Dead Reck­on­ing Part 1 AI Takeaways

Zvi1 Nov 2023 12:52 UTC
47 points
13 comments6 min readLW link

Ro­bust­ness of Con­trast-Con­sis­tent Search to Ad­ver­sar­ial Prompting

1 Nov 2023 12:46 UTC
18 points
1 comment7 min readLW link

The Bletch­ley Dec­la­ra­tion on AI Safety

Hauke Hillebrandt1 Nov 2023 11:44 UTC
17 points
0 comments1 min readLW link
(www.gov.uk)

Bay Win­ter Sols­tice 2023: Song & speech auditions

tcheasdfjkl1 Nov 2023 4:17 UTC
17 points
2 comments1 min readLW link

On Hav­ing No Clue

Chris_Leong1 Nov 2023 1:36 UTC
20 points
11 comments1 min readLW link

Balanc­ing Se­cu­rity Mind­set with Col­lab­o­ra­tive Re­search: A Proposal

MadHatter1 Nov 2023 0:46 UTC
9 points
3 comments4 min readLW link

Com­pu­ta­tional Ap­proaches to Pathogen Detection

jefftk1 Nov 2023 0:30 UTC
32 points
5 comments5 min readLW link
(www.jefftk.com)

Thoughts on the AI Safety Sum­mit com­pany policy re­quests and responses

So8res31 Oct 2023 23:54 UTC
169 points
14 comments10 min readLW link

AISN #25: White House Ex­ec­u­tive Order on AI, UK AI Safety Sum­mit, and Progress on Vol­un­tary Eval­u­a­tions of AI Risks

31 Oct 2023 19:34 UTC
35 points
1 comment6 min readLW link
(newsletter.safe.ai)

If AIs be­come self-aware, what re­li­gion will they have?

mnvr31 Oct 2023 17:29 UTC
−17 points
3 comments4 min readLW link

Self-Blinded L-Thea­nine RCT

niplav31 Oct 2023 15:24 UTC
53 points
12 comments3 min readLW link

AI Safety 101 - Chap­ter 5.2 - Un­re­stricted Ad­ver­sar­ial Training

Charbel-Raphaël31 Oct 2023 14:34 UTC
17 points
0 comments19 min readLW link

Prevent­ing Lan­guage Models from hid­ing their reasoning

31 Oct 2023 14:34 UTC
113 points
14 comments12 min readLW link

AI Safety 101 - Chap­ter 5.1 - Debate

Charbel-Raphaël31 Oct 2023 14:29 UTC
15 points
0 comments13 min readLW link

M&A in AI

Hauke Hillebrandt31 Oct 2023 12:20 UTC
2 points
0 comments1 min readLW link

Urg­ing an In­ter­na­tional AI Treaty: An Open Letter

Olli Järviniemi31 Oct 2023 11:26 UTC
48 points
2 comments1 min readLW link
(aitreaty.org)

[Closed] Agent Foun­da­tions track in MATS

Vanessa Kosoy31 Oct 2023 8:12 UTC
54 points
1 comment1 min readLW link
(www.matsprogram.org)

In­trin­sic Drives and Ex­trin­sic Mi­suse: Two In­ter­twined Risks of AI

jsteinhardt31 Oct 2023 5:10 UTC
40 points
0 comments12 min readLW link
(bounded-regret.ghost.io)

Fo­cus on ex­is­ten­tial risk is a dis­trac­tion from the real is­sues. A false fallacy

Nik Samoylov30 Oct 2023 23:42 UTC
−19 points
11 comments2 min readLW link

Will re­leas­ing the weights of large lan­guage mod­els grant wide­spread ac­cess to pan­demic agents?

jefftk30 Oct 2023 18:22 UTC
46 points
25 comments1 min readLW link
(arxiv.org)

[Linkpost] Two ma­jor an­nounce­ments in AI gov­er­nance today

Angélina30 Oct 2023 17:28 UTC
1 point
1 comment1 min readLW link
(www.whitehouse.gov)

Grokking Beyond Neu­ral Networks

Jack Miller30 Oct 2023 17:28 UTC
10 points
0 comments2 min readLW link
(arxiv.org)

Re­sponse to “Co­or­di­nated paus­ing: An eval­u­a­tion-based co­or­di­na­tion scheme for fron­tier AI de­vel­op­ers”

Matthew Wearden30 Oct 2023 17:27 UTC
5 points
2 comments6 min readLW link
(matthewwearden.co.uk)