RSS

AI Governance

TagLast edit: Feb 8, 2025, 12:32 AM by lesswrong-internal

AI Governance asks how we can ensure society benefits at large from increasingly powerful AI systems. While solving technical AI alignment is a necessary step towards this goal, it is by no means sufficient.

Governance includes policy, economics, sociology, law, and many other fields.

AI policy ideas: Read­ing list

Zach Stein-PerlmanApr 17, 2023, 7:00 PM
24 points
7 comments4 min readLW link

What an ac­tu­ally pes­simistic con­tain­ment strat­egy looks like

lcApr 5, 2022, 12:19 AM
679 points
138 comments6 min readLW link2 reviews

Speak­ing to Con­gres­sional staffers about AI risk

Dec 4, 2023, 11:08 PM
307 points
25 comments15 min readLW link1 review

Ways I Ex­pect AI Reg­u­la­tion To In­crease Ex­tinc­tion Risk

1a3ornJul 4, 2023, 5:32 PM
225 points
32 comments7 min readLW link

On MAIM and Su­per­in­tel­li­gence Strategy

ZviMar 14, 2025, 12:30 PM
51 points
2 comments13 min readLW link
(thezvi.wordpress.com)

What would a com­pute mon­i­tor­ing plan look like? [Linkpost]

Orpheus16Mar 26, 2023, 7:33 PM
158 points
10 comments4 min readLW link
(arxiv.org)

Should we post­pone AGI un­til we reach safety?

otto.bartenNov 18, 2020, 3:43 PM
27 points
36 comments3 min readLW link

He­len Toner on China, CSET, and AI

Rob BensingerApr 21, 2019, 4:10 AM
68 points
4 comments7 min readLW link
(rationallyspeakingpodcast.org)

Re­ac­tions to the Ex­ec­u­tive Order

ZviNov 1, 2023, 8:40 PM
77 points
4 comments29 min readLW link
(thezvi.wordpress.com)

RTFB: On the New Pro­posed CAIP AI Bill

ZviApr 10, 2024, 6:30 PM
119 points
14 comments34 min readLW link
(thezvi.wordpress.com)

News : Bi­den-⁠Har­ris Ad­minis­tra­tion Se­cures Vol­un­tary Com­mit­ments from Lead­ing Ar­tifi­cial In­tel­li­gence Com­pa­nies to Man­age the Risks Posed by AI

Jonathan ClaybroughJul 21, 2023, 6:00 PM
65 points
10 comments2 min readLW link
(www.whitehouse.gov)

Com­pute Thresh­olds: pro­posed rules to miti­gate risk of a “lab leak” ac­ci­dent dur­ing AI train­ing runs

davidadJul 22, 2023, 6:09 PM
80 points
2 comments2 min readLW link

AI labs’ state­ments on governance

Zach Stein-PerlmanJul 4, 2023, 4:30 PM
30 points
0 comments36 min readLW link

We’re Not Ready: thoughts on “paus­ing” and re­spon­si­ble scal­ing policies

HoldenKarnofskyOct 27, 2023, 3:19 PM
200 points
33 comments8 min readLW link

Re­sponse to Aschen­bren­ner’s “Si­tu­a­tional Aware­ness”

Rob BensingerJun 6, 2024, 10:57 PM
194 points
27 comments3 min readLW link

Where are the red lines for AI?

Karl von WendtAug 5, 2022, 9:34 AM
26 points
10 comments6 min readLW link

Pres­i­dent Bi­den Is­sues Ex­ec­u­tive Order on Safe, Se­cure, and Trust­wor­thy Ar­tifi­cial Intelligence

Tristan WilliamsOct 30, 2023, 11:15 AM
171 points
39 comments1 min readLW link
(www.whitehouse.gov)

Ac­tion­able-guidance and roadmap recom­men­da­tions for the NIST AI Risk Man­age­ment Framework

May 17, 2022, 3:26 PM
26 points
0 comments3 min readLW link

List of re­quests for an AI slow­down/​halt.

Cleo NardoApr 14, 2023, 11:55 PM
46 points
6 comments1 min readLW link

An up­com­ing US Supreme Court case may im­pede AI gov­er­nance efforts

NickGabsJul 16, 2023, 11:51 PM
57 points
17 comments2 min readLW link

If-Then Com­mit­ments for AI Risk Re­duc­tion [by Holden Karnofsky]

habrykaSep 13, 2024, 7:38 PM
28 points
0 comments20 min readLW link
(carnegieendowment.org)

Soft take­off can still lead to de­ci­sive strate­gic advantage

Daniel KokotajloAug 23, 2019, 4:39 PM
122 points
47 comments8 min readLW link4 reviews

[Question] Where are peo­ple think­ing and talk­ing about global co­or­di­na­tion for AI safety?

Wei DaiMay 22, 2019, 6:24 AM
112 points
22 comments1 min readLW link

[Question] Would it be good or bad for the US mil­i­tary to get in­volved in AI risk?

Grant DemareeJan 1, 2023, 7:02 PM
50 points
12 comments1 min readLW link

The Reg­u­la­tory Op­tion: A re­sponse to near 0% sur­vival odds

Matthew LowensteinApr 11, 2022, 10:00 PM
46 points
21 comments6 min readLW link

Cruxes on US lead for some do­mes­tic AI regulation

Zach Stein-PerlmanSep 10, 2023, 6:00 PM
26 points
3 comments2 min readLW link

China-AI forecasts

NathanBarnardFeb 25, 2024, 4:49 PM
39 points
29 comments6 min readLW link

New vol­un­tary com­mit­ments (AI Seoul Sum­mit)

Zach Stein-PerlmanMay 21, 2024, 11:00 AM
81 points
17 comments7 min readLW link
(www.gov.uk)

Miti­gat­ing ex­treme AI risks amid rapid progress [Linkpost]

Orpheus16May 21, 2024, 7:59 PM
21 points
7 comments4 min readLW link

The Tech In­dus­try is the Biggest Blocker to Mean­ingful AI Safety Regulations

garrisonAug 16, 2024, 7:37 PM
22 points
1 comment1 min readLW link
(garrisonlovely.substack.com)

The Su­gar Align­ment Problem

Adam ZernerDec 24, 2023, 1:35 AM
5 points
3 comments7 min readLW link

The Defence pro­duc­tion act and AI policy

NathanBarnardMar 1, 2024, 2:26 PM
37 points
0 comments2 min readLW link

AGI will be made of het­ero­ge­neous com­po­nents, Trans­former and Selec­tive SSM blocks will be among them

Roman LeventovDec 27, 2023, 2:51 PM
33 points
9 comments4 min readLW link

OpenAI’s Pre­pared­ness Frame­work: Praise & Recommendations

Orpheus16Jan 2, 2024, 4:20 PM
66 points
1 comment7 min readLW link

The Schumer Re­port on AI (RTFB)

ZviMay 24, 2024, 3:10 PM
34 points
3 comments36 min readLW link
(thezvi.wordpress.com)

Guide to SB 1047

ZviAug 20, 2024, 1:10 PM
71 points
18 comments53 min readLW link
(thezvi.wordpress.com)

(4 min read) An in­tu­itive ex­pla­na­tion of the AI in­fluence situation

trevorJan 13, 2024, 5:34 PM
12 points
26 comments4 min readLW link

Talk­ing to Congress: Can con­stituents con­tact­ing their leg­is­la­tor in­fluence policy?

Tristan WilliamsMar 7, 2024, 9:24 AM
14 points
0 comments1 min readLW link

[Question] What does it look like for AI to sig­nifi­cantly im­prove hu­man co­or­di­na­tion, be­fore su­per­in­tel­li­gence?

Bird ConceptJan 15, 2024, 7:22 PM
22 points
2 comments1 min readLW link

In­vi­ta­tion to lead a pro­ject at AI Safety Camp (Vir­tual Edi­tion, 2025)

Aug 23, 2024, 2:18 PM
17 points
2 comments4 min readLW link

Paus­ing AI is Pos­i­tive Ex­pected Value

LironMar 10, 2024, 5:10 PM
9 points
2 comments3 min readLW link
(twitter.com)

OpenAI: Fallout

ZviMay 28, 2024, 1:20 PM
204 points
25 comments36 min readLW link
(thezvi.wordpress.com)

My guess at Con­jec­ture’s vi­sion: trig­ger­ing a nar­ra­tive bifurcation

Alexandre VariengienFeb 6, 2024, 7:10 PM
75 points
12 comments16 min readLW link

Thoughts on SB-1047

ryan_greenblattMay 29, 2024, 11:26 PM
60 points
1 comment11 min readLW link

Many ar­gu­ments for AI x-risk are wrong

TurnTroutMar 5, 2024, 2:31 AM
159 points
87 comments12 min readLW link

My (cur­rent) model of what an AI gov­er­nance re­searcher does

Johan de KockAug 26, 2024, 5:58 PM
1 point
2 comments5 min readLW link

Trans­for­ma­tive trust­build­ing via ad­vance­ments in de­cen­tral­ized lie detection

trevorMar 16, 2024, 5:56 AM
20 points
10 comments38 min readLW link
(www.ncbi.nlm.nih.gov)

SB 1047: Fi­nal Takes and Also AB 3211

ZviAug 27, 2024, 10:10 PM
92 points
11 comments21 min readLW link
(thezvi.wordpress.com)

Ex­plain­ing the Joke: Paus­ing is The Way

WillPetilloApr 4, 2025, 9:04 AM
24 points
2 comments10 min readLW link

Ver­ifi­ca­tion meth­ods for in­ter­na­tional AI agreements

Orpheus16Aug 31, 2024, 2:58 PM
14 points
1 comment4 min readLW link
(arxiv.org)

Paul Chris­ti­ano named as US AI Safety In­sti­tute Head of AI Safety

Joel BurgetApr 16, 2024, 4:22 PM
256 points
58 comments1 min readLW link
(www.commerce.gov)

AXRP Epi­sode 28 - Su­ing Labs for AI Risk with Gabriel Weil

DanielFilanApr 17, 2024, 9:42 PM
12 points
0 comments65 min readLW link

The Dis­solu­tion of AI Safety

RokoDec 12, 2024, 10:34 AM
8 points
44 comments1 min readLW link
(www.transhumanaxiology.com)

Q&A on Pro­posed SB 1047

ZviMay 2, 2024, 3:10 PM
74 points
8 comments44 min readLW link
(thezvi.wordpress.com)

[Question] Have any par­ties in the cur­rent Euro­pean Par­li­a­men­tary Elec­tion made pub­lic state­ments on AI?

MondSemmelMay 10, 2024, 10:22 AM
9 points
0 comments1 min readLW link

Ad­vice for Ac­tivists from the His­tory of Environmentalism

Jeffrey HeningerMay 16, 2024, 6:40 PM
100 points
8 comments6 min readLW link
(blog.aiimpacts.org)

What is SB 1047 *for*?

RaemonSep 5, 2024, 5:39 PM
61 points
8 comments3 min readLW link

Pay Risk Eval­u­a­tors in Cash, Not Equity

Adam SchollSep 7, 2024, 2:37 AM
212 points
19 comments1 min readLW link

The Sorry State of AI X-Risk Ad­vo­cacy, and Thoughts on Do­ing Better

Thane RuthenisFeb 21, 2025, 8:15 PM
148 points
51 comments6 min readLW link

AI gov­er­nance needs a the­ory of victory

Jun 21, 2024, 4:15 PM
45 points
8 comments1 min readLW link
(www.convergenceanalysis.org)

My takes on SB-1047

leogaoSep 9, 2024, 6:38 PM
151 points
8 comments4 min readLW link

Schel­ling points in the AGI policy space

mesaoptimizerJun 26, 2024, 1:19 PM
52 points
2 comments6 min readLW link

In­tro­duc­tion to French AI Policy

Lucie PhilipponJul 4, 2024, 3:39 AM
110 points
12 comments6 min readLW link

Align­ment can be the ‘clean en­ergy’ of AI

Feb 22, 2025, 12:08 AM
67 points
8 comments8 min readLW link

Ad­vice to ju­nior AI gov­er­nance researchers

Orpheus16Jul 8, 2024, 7:19 PM
66 points
1 comment5 min readLW link

How much to up­date on re­cent AI gov­er­nance moves?

Nov 16, 2023, 11:46 PM
112 points
5 comments29 min readLW link

New page: Integrity

Zach Stein-PerlmanJul 10, 2024, 3:00 PM
91 points
3 comments1 min readLW link

An AI Race With China Can Be Bet­ter Than Not Racing

niplavJul 2, 2024, 5:57 PM
69 points
33 comments11 min readLW link

Con­sider Join­ing the UK Foun­da­tion Model Taskforce

ZviJul 10, 2023, 1:50 PM
105 points
12 comments1 min readLW link
(thezvi.wordpress.com)

[Re­search log] The board of Alpha­bet would stop Deep­Mind to save the world

Lucie PhilipponJul 16, 2024, 4:59 AM
6 points
0 comments4 min readLW link

Reflec­tions on the state of the race to su­per­in­tel­li­gence, Fe­bru­ary 2025

Mitchell_PorterFeb 23, 2025, 1:58 PM
21 points
7 comments4 min readLW link

Deter­min­ing the power of in­vestors over Fron­tier AI Labs is strate­gi­cally im­por­tant to re­duce x-risk

Lucie PhilipponJul 25, 2024, 1:12 AM
18 points
7 comments2 min readLW link

Re: An­thropic’s sug­gested SB-1047 amendments

RobertMJul 27, 2024, 10:32 PM
87 points
13 comments9 min readLW link
(www.documentcloud.org)

Twit­ter thread on poli­tics of AI safety

Richard_NgoJul 31, 2024, 12:00 AM
35 points
2 comments1 min readLW link
(x.com)

Refin­ing MAIM: Iden­ti­fy­ing Changes Re­quired to Meet Con­di­tions for Deterrence

David AbecassisApr 11, 2025, 12:49 AM
17 points
0 comments11 min readLW link
(intelligence.org)

🇫🇷 An­nounc­ing CeSIA: The French Cen­ter for AI Safety

Charbel-RaphaëlDec 20, 2024, 2:17 PM
88 points
2 comments8 min readLW link

GPT-4o Sys­tem Card

Zach Stein-PerlmanAug 8, 2024, 8:30 PM
68 points
11 comments2 min readLW link
(openai.com)

Cal­ifor­ni­ans, tell your reps to vote yes on SB 1047!

Holly_ElmoreAug 12, 2024, 7:50 PM
40 points
24 comments1 min readLW link

[Question] What pre­vents SB-1047 from trig­ger­ing on deep fake porn/​voice clon­ing fraud?

ChristianKlSep 26, 2024, 9:17 AM
27 points
21 comments1 min readLW link

A Nar­row Path: a plan to deal with AI ex­tinc­tion risk

Oct 7, 2024, 1:02 PM
73 points
12 comments2 min readLW link
(www.narrowpath.co)

A path to hu­man autonomy

Nathan Helm-BurgerOct 29, 2024, 3:02 AM
53 points
16 comments20 min readLW link

Linkpost: Me­moran­dum on Ad­vanc­ing the United States’ Lead­er­ship in Ar­tifi­cial Intelligence

NisanOct 25, 2024, 4:37 AM
60 points
2 comments1 min readLW link
(www.whitehouse.gov)

Lab gov­er­nance read­ing list

Zach Stein-PerlmanOct 25, 2024, 6:00 PM
20 points
3 comments1 min readLW link

Finish­ing The SB-1047 Doc­u­men­tary In 6 Weeks

Michaël TrazziOct 28, 2024, 8:17 PM
94 points
7 comments4 min readLW link
(manifund.org)

UK AISI: Early les­sons from eval­u­at­ing fron­tier AI systems

Zach Stein-PerlmanOct 25, 2024, 7:00 PM
26 points
0 comments2 min readLW link
(www.aisi.gov.uk)

AI #88: Thanks for the Memos

ZviOct 31, 2024, 3:00 PM
46 points
5 comments77 min readLW link
(thezvi.wordpress.com)

Mak­ing a con­ser­va­tive case for alignment

Nov 15, 2024, 6:55 PM
208 points
67 comments7 min readLW link

AXRP Epi­sode 38.1 - Alan Chan on Agent Infrastructure

DanielFilanNov 16, 2024, 11:30 PM
12 points
0 comments14 min readLW link

Should there be just one west­ern AGI pro­ject?

Dec 3, 2024, 10:11 AM
78 points
72 comments15 min readLW link
(www.forethought.org)

Anal­y­sis of Global AI Gover­nance Strategies

Dec 4, 2024, 10:45 AM
49 points
10 comments36 min readLW link

The Jack­pot Jinx (or why “Su­per­in­tel­li­gence Strat­egy” is wrong)

E.G. Blee-GoldmanMar 10, 2025, 7:18 PM
13 points
0 comments5 min readLW link

Rol­ling Thresh­olds for AGI Scal­ing Regulation

LarksJan 12, 2025, 1:30 AM
40 points
6 comments1 min readLW link

Some cruxes on im­pact­ful al­ter­na­tives to AI policy work

Richard_NgoOct 10, 2018, 1:35 PM
165 points
13 comments12 min readLW link

AI pause/​gov­er­nance ad­vo­cacy might be net-nega­tive, es­pe­cially with­out a fo­cus on ex­plain­ing x-risk

Mikhail SaminAug 27, 2023, 11:05 PM
72 points
9 comments6 min readLW link

In­tel­sat as a Model for In­ter­na­tional AGI Governance

Mar 13, 2025, 12:58 PM
45 points
0 comments1 min readLW link
(www.forethought.org)

Re­view of Soft Take­off Can Still Lead to DSA

Daniel KokotajloJan 10, 2021, 6:10 PM
85 points
16 comments6 min readLW link

Whether gov­ern­ments will con­trol AGI is im­por­tant and neglected

Seth HerdMar 14, 2025, 9:48 AM
24 points
2 comments9 min readLW link

AI com­pa­nies are un­likely to make high-as­surance safety cases if timelines are short

ryan_greenblattJan 23, 2025, 6:41 PM
145 points
5 comments13 min readLW link

Dario Amodei leaves OpenAI

Daniel KokotajloDec 29, 2020, 7:31 PM
69 points
13 comments1 min readLW link

Retroac­tive If-Then Commitments

MichaelDickensFeb 1, 2025, 10:22 PM
6 points
0 comments1 min readLW link

On the Ra­tion­al­ity of Deter­ring ASI

Dan HMar 5, 2025, 4:11 PM
166 points
34 comments4 min readLW link
(nationalsecurity.ai)

More on Var­i­ous AI Ac­tion Plans

ZviMar 24, 2025, 1:10 PM
32 points
0 comments11 min readLW link
(thezvi.wordpress.com)

On the Meta and Deep­Mind Safety Frameworks

ZviFeb 7, 2025, 1:10 PM
45 points
1 comment17 min readLW link
(thezvi.wordpress.com)

New Bill AB 501 to Prevent OpenAI’s Non-profit Conversion

Peter WindbergerMar 25, 2025, 12:41 AM
18 points
1 comment1 min readLW link

Con­ver­gence 2024 Im­pact Review

David_KristofferssonMar 24, 2025, 8:28 PM
13 points
0 comments1 min readLW link

The Paris AI Anti-Safety Summit

ZviFeb 12, 2025, 2:00 PM
129 points
21 comments21 min readLW link
(thezvi.wordpress.com)

The Na­tional Defense Autho­riza­tion Act Con­tains AI Provisions

ryan_bJan 5, 2021, 3:51 PM
30 points
24 comments1 min readLW link

Govern­ing High-Im­pact AI Sys­tems: Un­der­stand­ing Canada’s Pro­posed AI Bill. April 15, Car­leton Univer­sity, Ottawa

Liav KorenMar 28, 2023, 5:48 PM
11 points
1 comment1 min readLW link
(forum.effectivealtruism.org)

How is AI gov­erned and reg­u­lated, around the world?

Mitchell_PorterMar 30, 2023, 3:36 PM
15 points
6 comments2 min readLW link

ChatGPT banned in Italy over pri­vacy concerns

Ollie JMar 31, 2023, 5:33 PM
18 points
4 comments1 min readLW link
(www.bbc.co.uk)

[Question] What Are Your Prefer­ences Re­gard­ing The FLI Let­ter?

JenniferRMApr 1, 2023, 4:52 AM
−4 points
122 comments16 min readLW link

Policy dis­cus­sions fol­low strong con­tex­tu­al­iz­ing norms

Richard_NgoApr 1, 2023, 11:51 PM
230 points
61 comments3 min readLW link

AI Sum­mer Harvest

Cleo NardoApr 4, 2023, 3:35 AM
130 points
10 comments1 min readLW link

Ex­ces­sive AI growth-rate yields lit­tle so­cio-eco­nomic benefit.

Cleo NardoApr 4, 2023, 7:13 PM
27 points
22 comments4 min readLW link

I asked my sen­a­tor to slow AI

OmidApr 6, 2023, 6:18 PM
21 points
5 comments2 min readLW link

An ‘AGI Emer­gency Eject Cri­te­ria’ con­sen­sus could be re­ally use­ful.

tcelferactApr 7, 2023, 4:21 PM
5 points
0 comments1 min readLW link

All images from the WaitButWhy se­quence on AI

trevorApr 8, 2023, 7:36 AM
73 points
5 comments2 min readLW link

Cur­rent UK gov­ern­ment lev­ers on AI development

rosehadsharApr 10, 2023, 1:16 PM
16 points
0 comments1 min readLW link

Re­quest to AGI or­ga­ni­za­tions: Share your views on paus­ing AI progress

Apr 11, 2023, 5:30 PM
141 points
11 comments1 min readLW link

FLI And Eliezer Should Reach Consensus

JenniferRMApr 11, 2023, 4:07 AM
21 points
6 comments23 min readLW link

Cy­berspace Ad­minis­tra­tion of China: Draft of “Reg­u­la­tion for Gen­er­a­tive Ar­tifi­cial In­tel­li­gence Ser­vices” is open for comments

sanxiynApr 11, 2023, 9:32 AM
7 points
2 comments1 min readLW link
(archive.is)

NTIA—AI Ac­countabil­ity Announcement

samshapApr 11, 2023, 3:03 PM
7 points
0 comments1 min readLW link
(www.ntia.doc.gov)

Na­tional Telecom­mu­ni­ca­tions and In­for­ma­tion Ad­minis­tra­tion: AI Ac­countabil­ity Policy Re­quest for Comment

sanxiynApr 11, 2023, 10:59 PM
9 points
0 comments1 min readLW link
(ntia.gov)

Nav­i­gat­ing the Open-Source AI Land­scape: Data, Fund­ing, and Safety

Apr 13, 2023, 3:29 PM
32 points
7 comments11 min readLW link
(forum.effectivealtruism.org)

FLI re­port: Poli­cy­mak­ing in the Pause

Zach Stein-PerlmanApr 15, 2023, 5:01 PM
15 points
3 comments1 min readLW link
(futureoflife.org)

Slow­ing AI: Foundations

Zach Stein-PerlmanApr 17, 2023, 2:30 PM
45 points
11 comments17 min readLW link

Re­spon­si­ble De­ploy­ment in 20XX

CarsonApr 20, 2023, 12:24 AM
4 points
0 comments4 min readLW link

OpenAI could help X-risk by wa­ger­ing itself

VojtaKovarikApr 20, 2023, 2:51 PM
31 points
16 comments1 min readLW link

My Assess­ment of the Chi­nese AI Safety Community

Lao MeinApr 25, 2023, 4:21 AM
250 points
94 comments3 min readLW link

Notes on Po­ten­tial Fu­ture AI Tax Policy

ZviApr 25, 2023, 1:30 PM
33 points
6 comments9 min readLW link
(thezvi.wordpress.com)

Refram­ing the bur­den of proof: Com­pa­nies should prove that mod­els are safe (rather than ex­pect­ing au­di­tors to prove that mod­els are dan­ger­ous)

Orpheus16Apr 25, 2023, 6:49 PM
27 points
11 comments3 min readLW link
(childrenoficarus.substack.com)

AI Safety is Drop­ping the Ball on Clown Attacks

trevorOct 22, 2023, 8:09 PM
74 points
82 comments34 min readLW link

An­thropic, Google, Microsoft & OpenAI an­nounce Ex­ec­u­tive Direc­tor of the Fron­tier Model Fo­rum & over $10 mil­lion for a new AI Safety Fund

Zach Stein-PerlmanOct 25, 2023, 3:20 PM
31 points
8 comments4 min readLW link
(www.frontiermodelforum.org)

Thoughts on re­spon­si­ble scal­ing poli­cies and regulation

paulfchristianoOct 24, 2023, 10:21 PM
221 points
33 comments6 min readLW link

AI #35: Re­spon­si­ble Scal­ing Policies

ZviOct 26, 2023, 1:30 PM
66 points
10 comments55 min readLW link
(thezvi.wordpress.com)

5 Rea­sons Why Govern­ments/​Mili­taries Already Want AI for In­for­ma­tion Warfare

trevorOct 30, 2023, 4:30 PM
32 points
0 comments10 min readLW link

[Linkpost] Bi­den-Har­ris Ex­ec­u­tive Order on AI

berenOct 30, 2023, 3:20 PM
3 points
0 comments1 min readLW link

Urg­ing an In­ter­na­tional AI Treaty: An Open Letter

Olli JärviniemiOct 31, 2023, 11:26 AM
48 points
2 comments1 min readLW link
(aitreaty.org)

On the Ex­ec­u­tive Order

ZviNov 1, 2023, 2:20 PM
100 points
4 comments30 min readLW link
(thezvi.wordpress.com)

[Question] Snap­shot of nar­ra­tives and frames against reg­u­lat­ing AI

Jan_KulveitNov 1, 2023, 4:30 PM
36 points
19 comments3 min readLW link

Dario Amodei’s pre­pared re­marks from the UK AI Safety Sum­mit, on An­thropic’s Re­spon­si­ble Scal­ing Policy

Zac Hatfield-DoddsNov 1, 2023, 6:10 PM
85 points
1 comment4 min readLW link
(www.anthropic.com)

We are already in a per­sua­sion-trans­formed world and must take precautions

trevorNov 4, 2023, 3:53 PM
37 points
14 comments6 min readLW link

The 6D effect: When com­pa­nies take risks, one email can be very pow­er­ful.

scasperNov 4, 2023, 8:08 PM
278 points
42 comments3 min readLW link

On the UK Summit

ZviNov 7, 2023, 1:10 PM
74 points
6 comments30 min readLW link
(thezvi.wordpress.com)

Sur­vey on the ac­cel­er­a­tion risks of our new RFPs to study LLM capabilities

Ajeya CotraNov 10, 2023, 11:59 PM
27 points
1 comment1 min readLW link

AXRP Epi­sode 26 - AI Gover­nance with Eliz­a­beth Seger

DanielFilanNov 26, 2023, 11:00 PM
14 points
0 comments66 min readLW link

Safety stan­dards: a frame­work for AI regulation

joshcMay 1, 2023, 12:56 AM
19 points
0 comments8 min readLW link

Stop­ping dan­ger­ous AI: Ideal lab behavior

Zach Stein-PerlmanMay 9, 2023, 9:00 PM
8 points
0 comments2 min readLW link

Stop­ping dan­ger­ous AI: Ideal US behavior

Zach Stein-PerlmanMay 9, 2023, 9:00 PM
17 points
0 comments3 min readLW link

GovAI: Towards best prac­tices in AGI safety and gov­er­nance: A sur­vey of ex­pert opinion

Zach Stein-PerlmanMay 15, 2023, 1:42 AM
28 points
11 comments1 min readLW link
(arxiv.org)

Eisen­hower’s Atoms for Peace Speech

Orpheus16May 17, 2023, 4:10 PM
18 points
3 comments11 min readLW link
(www.iaea.org)

[Linkpost] “Gover­nance of su­per­in­tel­li­gence” by OpenAI

Daniel_EthMay 22, 2023, 8:15 PM
67 points
20 comments1 min readLW link

AI #12:The Quest for Sane Regulations

ZviMay 18, 2023, 1:20 PM
77 points
12 comments64 min readLW link
(thezvi.wordpress.com)

State­ment on AI Ex­tinc­tion—Signed by AGI Labs, Top Aca­demics, and Many Other Notable Figures

Dan HMay 30, 2023, 9:05 AM
382 points
78 comments1 min readLW link1 review
(www.safe.ai)

[Question] Who is li­able for AI?

jmhMay 30, 2023, 1:54 PM
14 points
4 comments1 min readLW link

The case for re­mov­ing al­ign­ment and ML re­search from the train­ing dataset

berenMay 30, 2023, 8:54 PM
48 points
8 comments5 min readLW link

Up­com­ing AI reg­u­la­tions are likely to make for an un­safer world

ShmiJun 3, 2023, 1:07 AM
18 points
14 comments1 min readLW link

The AGI Race Between the US and China Doesn’t Ex­ist.

Eva_BJun 3, 2023, 12:22 AM
33 points
15 comments7 min readLW link
(evabehrens.substack.com)

Rishi to out­line his vi­sion for Bri­tain to take the world lead in polic­ing AI threats when he meets Joe Biden

Mati_RoyJun 6, 2023, 4:47 AM
25 points
1 comment1 min readLW link
(www.dailymail.co.uk)

A sum­mary of cur­rent work in AI governance

constructiveJun 17, 2023, 6:41 PM
44 points
1 comment11 min readLW link
(forum.effectivealtruism.org)

Demo­cratic AI Con­sti­tu­tion: Round-Robin De­bate and Synthesis

scottviteriJun 24, 2023, 7:31 PM
10 points
4 comments5 min readLW link
(scottviteri.com)

“Safety Cul­ture for AI” is im­por­tant, but isn’t go­ing to be easy

DavidmanheimJun 26, 2023, 12:52 PM
47 points
2 comments2 min readLW link
(forum.effectivealtruism.org)

Lit­tle at­ten­tion seems to be on dis­cour­ag­ing hard­ware progress

RussellThorJun 30, 2023, 10:14 AM
5 points
3 comments1 min readLW link

Foom Liability

PeterMcCluskeyJun 30, 2023, 3:55 AM
22 points
10 comments6 min readLW link
(bayesianinvestor.com)

Ap­par­ently, of the 195 Million the DoD al­lo­cated in Univer­sity Re­search Fund­ing Awards in 2022, more than half of them con­cerned AI or com­pute hard­ware research

mako yassJul 7, 2023, 1:20 AM
41 points
5 comments2 min readLW link
(www.defense.gov)

My fa­vorite AI gov­er­nance re­search this year so far

Zach Stein-PerlmanJul 23, 2023, 4:30 PM
26 points
1 comment7 min readLW link
(blog.aiimpacts.org)

Pod­cast (+tran­script): Nathan Barnard on how US fi­nan­cial reg­u­la­tion can in­form AI governance

Aaron BergmanAug 8, 2023, 9:46 PM
8 points
0 comments1 min readLW link
(www.aaronbergman.net)

One ex­am­ple of how LLM pro­pa­ganda at­tacks can hack the brain

trevorAug 16, 2023, 9:41 PM
24 points
8 comments4 min readLW link

Assess­ment of in­tel­li­gence agency func­tion­al­ity is difficult yet important

trevorAug 24, 2023, 1:42 AM
48 points
5 comments9 min readLW link

In­for­ma­tion war­fare his­tor­i­cally re­volved around hu­man conduits

trevorAug 28, 2023, 6:54 PM
37 points
7 comments3 min readLW link

Re­port on Fron­tier Model Training

YafahEdelmanAug 30, 2023, 8:02 PM
122 points
21 comments21 min readLW link
(docs.google.com)

What I Would Do If I Were Work­ing On AI Governance

johnswentworthDec 8, 2023, 6:43 AM
110 points
32 comments10 min readLW link

ARC Evals: Re­spon­si­ble Scal­ing Policies

Zach Stein-PerlmanSep 28, 2023, 4:30 AM
40 points
10 comments2 min readLW link1 review
(evals.alignment.org)

An­thropic’s Re­spon­si­ble Scal­ing Policy & Long-Term Benefit Trust

Zac Hatfield-DoddsSep 19, 2023, 3:09 PM
83 points
26 comments3 min readLW link1 review
(www.anthropic.com)

Google’s Eth­i­cal AI team and AI Safety

magfrumpFeb 20, 2021, 9:42 AM
12 points
16 comments7 min readLW link

Ngo and Yud­kowsky on AI ca­pa­bil­ity gains

Nov 18, 2021, 10:19 PM
130 points
61 comments39 min readLW link1 review

Com­ments on Allan Dafoe on AI Governance

Alex FlintNov 29, 2021, 4:16 PM
13 points
0 comments7 min readLW link

The case for Do­ing Some­thing Else (if Align­ment is doomed)

Rafael HarthApr 5, 2022, 5:52 PM
94 points
14 comments2 min readLW link

Strate­gic Con­sid­er­a­tions Re­gard­ing Autis­tic/​Literal AI

Chris_LeongApr 6, 2022, 2:57 PM
−1 points
2 comments2 min readLW link

Why I Am Skep­ti­cal of AI Reg­u­la­tion as an X-Risk Miti­ga­tion Strategy

A RayAug 6, 2022, 5:46 AM
31 points
14 comments2 min readLW link

Jack Clark on the re­al­ities of AI policy

Kaj_SotalaAug 7, 2022, 8:44 AM
68 points
3 comments3 min readLW link
(threadreaderapp.com)

[Question] What if we solve AI Safety but no one cares

142857Aug 22, 2022, 5:38 AM
18 points
5 comments1 min readLW link

Re­place­ment for PONR concept

Daniel KokotajloSep 2, 2022, 12:09 AM
58 points
6 comments2 min readLW link

Sha­har Avin On How To Reg­u­late Ad­vanced AI Systems

Michaël TrazziSep 23, 2022, 3:46 PM
31 points
0 comments4 min readLW link
(theinsideview.ai)

Un­der what cir­cum­stances have gov­ern­ments can­cel­led AI-type sys­tems?

David GrossSep 23, 2022, 9:11 PM
7 points
1 comment1 min readLW link
(www.carnegieuktrust.org.uk)

Anal­y­sis: US re­stricts GPU sales to China

aogOct 7, 2022, 6:38 PM
102 points
58 comments5 min readLW link

[Question] Should we push for re­quiring AI train­ing data to be li­censed?

ChristianKlOct 19, 2022, 5:49 PM
37 points
32 comments1 min readLW link

Learn­ing so­cietal val­ues from law as part of an AGI al­ign­ment strategy

John NayOct 21, 2022, 2:03 AM
5 points
18 comments54 min readLW link

What does it take to defend the world against out-of-con­trol AGIs?

Steven ByrnesOct 25, 2022, 2:47 PM
208 points
49 comments30 min readLW link1 review

Mas­sive Scal­ing Should be Frowned Upon

harsimonyNov 17, 2022, 8:43 AM
4 points
6 comments5 min readLW link

[Question] How promis­ing are le­gal av­enues to re­strict AI train­ing data?

thehalliardDec 10, 2022, 4:31 PM
9 points
2 comments1 min readLW link

Prac­ti­cal AI risk I: Watch­ing large compute

Gustavo RamiresDec 24, 2022, 1:25 PM
3 points
0 comments1 min readLW link

List #2: Why co­or­di­nat­ing to al­ign as hu­mans to not de­velop AGI is a lot eas­ier than, well… co­or­di­nat­ing as hu­mans with AGI co­or­di­nat­ing to be al­igned with humans

RemmeltDec 24, 2022, 9:53 AM
1 point
0 comments3 min readLW link

My thoughts on OpenAI’s al­ign­ment plan

Orpheus16Dec 30, 2022, 7:33 PM
55 points
3 comments20 min readLW link

Went­worth and Larsen on buy­ing time

Jan 9, 2023, 9:31 PM
74 points
6 comments12 min readLW link

Thoughts on hard­ware /​ com­pute re­quire­ments for AGI

Steven ByrnesJan 24, 2023, 2:03 PM
62 points
31 comments24 min readLW link

[Question] AI safety mile­stones?

Zach Stein-PerlmanJan 23, 2023, 9:00 PM
7 points
5 comments1 min readLW link

AI Risk Man­age­ment Frame­work | NIST

DragonGodJan 26, 2023, 3:27 PM
36 points
4 comments2 min readLW link
(www.nist.gov)

What is the ground re­al­ity of coun­tries tak­ing steps to re­cal­ibrate AI de­vel­op­ment to­wards Align­ment first?

NebuchJan 29, 2023, 1:26 PM
8 points
6 comments3 min readLW link

Product safety is a poor model for AI governance

Richard Korzekwa Feb 1, 2023, 10:40 PM
36 points
0 comments5 min readLW link
(aiimpacts.org)

Many AI gov­er­nance pro­pos­als have a trade­off be­tween use­ful­ness and feasibility

Feb 3, 2023, 6:49 PM
22 points
2 comments2 min readLW link

4 ways to think about de­moc­ra­tiz­ing AI [GovAI Linkpost]

Orpheus16Feb 13, 2023, 6:06 PM
24 points
4 comments1 min readLW link
(www.governance.ai)

How should AI sys­tems be­have, and who should de­cide? [OpenAI blog]

ShardPhoenixFeb 17, 2023, 1:05 AM
22 points
2 comments1 min readLW link
(openai.com)

Cy­borg Pe­ri­ods: There will be mul­ti­ple AI transitions

Feb 22, 2023, 4:09 PM
108 points
9 comments6 min readLW link

AI Gover­nance & Strat­egy: Pri­ori­ties, tal­ent gaps, & opportunities

Orpheus16Mar 3, 2023, 6:09 PM
56 points
2 comments4 min readLW link

[Linkpost] Scott Alexan­der re­acts to OpenAI’s lat­est post

Orpheus16Mar 11, 2023, 10:24 PM
27 points
0 comments5 min readLW link
(astralcodexten.substack.com)

The AI Adop­tion Gap: Prepar­ing the US Govern­ment for Ad­vanced AI

LizkaApr 2, 2025, 11:46 PM
14 points
2 comments17 min readLW link
(www.forethought.org)

The Wizard of Oz Prob­lem: How in­cen­tives and nar­ra­tives can skew our per­cep­tion of AI developments

Orpheus16Mar 20, 2023, 8:44 PM
16 points
3 comments6 min readLW link

AI Tracker: mon­i­tor­ing cur­rent and near-fu­ture risks from su­per­scale models

Nov 23, 2021, 7:16 PM
67 points
13 comments3 min readLW link
(aitracker.org)

AI Align­ment Meme Viruses

RationalDinoJan 15, 2025, 3:55 PM
4 points
0 comments2 min readLW link

2024 Sum­mer AI Safety In­tro Fel­low­ship and So­cials in Boston

KevinWeiMay 29, 2024, 6:27 PM
8 points
0 comments1 min readLW link

What Failure Looks Like is not an ex­is­ten­tial risk (and al­ign­ment is not the solu­tion)

otto.bartenFeb 2, 2024, 6:59 PM
13 points
12 comments9 min readLW link

2019 AI Align­ment Liter­a­ture Re­view and Char­ity Comparison

LarksDec 19, 2019, 3:00 AM
130 points
18 comments62 min readLW link

“Long” timelines to ad­vanced AI have got­ten crazy short

Matrice JacobineApr 3, 2025, 10:46 PM
21 points
0 comments1 min readLW link
(helentoner.substack.com)

Should AI sys­tems have to iden­tify them­selves?

Darren McKeeDec 31, 2022, 2:57 AM
2 points
2 comments1 min readLW link

Overview of in­tro­duc­tory re­sources in AI Governance

Lucie PhilipponMay 27, 2024, 4:21 PM
19 points
0 comments6 min readLW link

OpenAI Credit Ac­count (2510$)

Emirhan BULUTJan 21, 2024, 2:32 AM
1 point
0 comments1 min readLW link

Ngo’s view on al­ign­ment difficulty

Dec 14, 2021, 9:34 PM
63 points
7 comments17 min readLW link

Maybe An­thropic’s Long-Term Benefit Trust is powerless

Zach Stein-PerlmanMay 27, 2024, 1:00 PM
201 points
21 comments2 min readLW link

AI Gover­nance Fun­da­men­tals—Cur­ricu­lum and Application

MauNov 30, 2021, 2:19 AM
17 points
0 comments1 min readLW link

HIRING: In­form and shape a new pro­ject on AI safety at Part­ner­ship on AI

madhu_likaDec 7, 2021, 7:37 PM
1 point
0 comments1 min readLW link

De­mand­ing and De­sign­ing Aligned Cog­ni­tive Architectures

Koen.HoltmanDec 21, 2021, 5:32 PM
8 points
5 comments5 min readLW link

An­nounc­ing Con­ver­gence Anal­y­sis: An In­sti­tute for AI Sce­nario & Gover­nance Research

Mar 7, 2024, 9:37 PM
23 points
1 comment4 min readLW link

AI: How We Got Here—A Neu­ro­science Perspective

Mordechai RorvigJan 19, 2025, 11:51 PM
5 points
0 comments2 min readLW link
(www.kickstarter.com)

De­moc­ra­tiz­ing AI Gover­nance: Balanc­ing Ex­per­tise and Public Participation

Lucile Ter-MinassianJan 21, 2025, 6:29 PM
1 point
0 comments15 min readLW link

Will AI Re­silience pro­tect Devel­op­ing Na­tions?

ejk64Jan 21, 2025, 3:31 PM
4 points
0 comments8 min readLW link

On Deep­Seek’s r1

ZviJan 22, 2025, 7:50 PM
55 points
2 comments35 min readLW link
(thezvi.wordpress.com)

The Hu­man Align­ment Prob­lem for AIs

rifeJan 22, 2025, 4:06 AM
10 points
5 comments3 min readLW link

What is an al­ign­ment tax?

Mar 20, 2025, 1:06 PM
5 points
0 comments1 min readLW link
(aisafety.info)

In­tro­duc­ing the Coal­i­tion for a Baruch Plan for AI: A Call for a Rad­i­cal Treaty-Mak­ing pro­cess for the Global Gover­nance of AI

rguerreschiJan 30, 2025, 3:26 PM
11 points
0 comments2 min readLW link

Thoughts about Policy Ecosys­tems: The Miss­ing Links in AI Governance

Echo HuangFeb 1, 2025, 1:54 AM
1 point
0 comments5 min readLW link

Ques­tion 4: Im­ple­ment­ing the con­trol proposals

Cameron BergFeb 13, 2022, 5:12 PM
6 points
2 comments5 min readLW link

A Plu­ral­is­tic Frame­work for Rogue AI Containment

TheThinkingArboristMar 22, 2025, 12:54 PM
1 point
0 comments7 min readLW link

How harm­ful are im­prove­ments in AI? + Poll

Feb 15, 2022, 6:16 PM
15 points
4 comments8 min readLW link

EU poli­cy­mak­ers reach an agree­ment on the AI Act

tlevinDec 15, 2023, 6:02 AM
78 points
7 comments7 min readLW link

Ap­ply to the Cam­bridge ERA:AI Fel­low­ship 2025

Harrison GMar 25, 2025, 1:50 PM
16 points
0 comments3 min readLW link

From No Mind to a Mind – A Con­ver­sa­tion That Changed an AI

parthibanarjuna sFeb 7, 2025, 11:50 AM
1 point
0 comments3 min readLW link

AI se­cu­rity might be helpful for AI alignment

Igor IvanovJan 6, 2023, 8:16 PM
36 points
1 comment2 min readLW link

Re­quest for In­for­ma­tion for a new US AI Ac­tion Plan (OSTP RFI)

agucovaFeb 7, 2025, 8:40 PM
5 points
0 comments1 min readLW link
(www.federalregister.gov)

Alt­man blog on post-AGI world

Julian BradshawFeb 9, 2025, 9:52 PM
29 points
10 comments1 min readLW link
(blog.samaltman.com)

Re­think­ing AI Safety Ap­proach in the Era of Open-Source AI

Weibing WangFeb 11, 2025, 2:01 PM
3 points
0 comments6 min readLW link

Ra­tional Effec­tive Utopia & Nar­row Way There: Mul­tiver­sal AI Align­ment, Place AI, New Ethico­physics… (Up­dated)

ankFeb 11, 2025, 3:21 AM
13 points
8 comments35 min readLW link

Where Would Good Fore­casts Most Help AI Gover­nance Efforts?

Violet HourFeb 11, 2025, 6:15 PM
11 points
1 comment6 min readLW link

Ar­tifi­cial Static Place In­tel­li­gence: Guaran­teed Alignment

ankFeb 15, 2025, 11:08 AM
2 points
2 comments2 min readLW link

Ex­plor­ing the Pre­cau­tion­ary Prin­ci­ple in AI Devel­op­ment: His­tor­i­cal Analo­gies and Les­sons Learned

Christopher KingMar 21, 2023, 3:53 AM
−1 points
2 comments9 min readLW link

CAIS-in­spired ap­proach to­wards safer and more in­ter­pretable AGIs

Peter HroššoMar 27, 2023, 2:36 PM
13 points
7 comments1 min readLW link

AI gov­er­nance stu­dent hackathon on Satur­day, April 23: reg­ister now!

micApr 12, 2022, 4:48 AM
14 points
0 comments1 min readLW link

Law-Fol­low­ing AI 1: Se­quence In­tro­duc­tion and Structure

CullenApr 27, 2022, 5:26 PM
18 points
10 comments9 min readLW link

Want to win the AGI race? Solve al­ign­ment.

leopoldMar 29, 2023, 5:40 PM
21 points
3 comments5 min readLW link
(www.forourposterity.com)

Law-Fol­low­ing AI 2: In­tent Align­ment + Su­per­in­tel­li­gence → Lawless AI (By De­fault)

CullenApr 27, 2022, 5:27 PM
5 points
2 comments6 min readLW link

The 0.2 OOMs/​year target

Cleo NardoMar 30, 2023, 6:15 PM
84 points
24 comments5 min readLW link

Wi­den­ing Over­ton Win­dow—Open Thread

PrometheusMar 31, 2023, 10:03 AM
23 points
8 comments1 min readLW link

Law-Fol­low­ing AI 3: Lawless AI Agents Un­der­mine Sta­bi­liz­ing Agreements

CullenApr 27, 2022, 5:30 PM
2 points
2 comments3 min readLW link

Paus­ing AI Devel­op­ments Isn’t Enough. We Need to Shut it All Down by Eliezer Yudkowsky

jacquesthibsMar 29, 2023, 11:16 PM
291 points
297 comments3 min readLW link
(time.com)

AI Alter­na­tive Fu­tures: Sce­nario Map­ping Ar­tifi­cial In­tel­li­gence Risk—Re­quest for Par­ti­ci­pa­tion (*Closed*)

KakiliApr 27, 2022, 10:07 PM
10 points
2 comments8 min readLW link

AI com­mu­nity build­ing: EliezerKart

Christopher KingApr 1, 2023, 3:25 PM
45 points
0 comments2 min readLW link

Pes­simism about AI Safety

Apr 2, 2023, 7:43 AM
4 points
1 comment25 min readLW link

Quick Thoughts on A.I. Governance

Nicholas / Heather KrossApr 30, 2022, 2:49 PM
70 points
8 comments2 min readLW link
(www.thinkingmuchbetter.com)

The AI gov­er­nance gaps in de­vel­op­ing countries

ntranJun 17, 2023, 2:50 AM
20 points
1 comment14 min readLW link

AI safety should be made more ac­cessible us­ing non text-based media

MassimogMay 10, 2022, 3:14 AM
2 points
4 comments4 min readLW link

Deep­Mind’s gen­er­al­ist AI, Gato: A non-tech­ni­cal explainer

May 16, 2022, 9:21 PM
63 points
6 comments6 min readLW link

Yoshua Ben­gio: “Slow­ing down de­vel­op­ment of AI sys­tems pass­ing the Tur­ing test”

Roman LeventovApr 6, 2023, 3:31 AM
49 points
2 comments5 min readLW link
(yoshuabengio.org)

Cor­po­rate Gover­nance for Fron­tier AI Labs: A Re­search Agenda

Matthew WeardenFeb 28, 2024, 11:29 AM
4 points
0 comments16 min readLW link
(matthewwearden.co.uk)

A bridge to Dath Ilan? Im­proved gov­er­nance on the crit­i­cal path to AI al­ign­ment.

Jackson WagnerMay 18, 2022, 3:51 PM
24 points
0 comments12 min readLW link

Re­shap­ing the AI Industry

Thane RuthenisMay 29, 2022, 10:54 PM
147 points
35 comments21 min readLW link

Six Di­men­sions of Oper­a­tional Ad­e­quacy in AGI Projects

Eliezer YudkowskyMay 30, 2022, 5:00 PM
310 points
66 comments13 min readLW link1 review

[Question] Could Pa­tent-Trol­ling de­lay AI timelines?

Pablo RepettoJun 10, 2022, 2:53 AM
1 point
3 comments1 min readLW link

FYI: I’m work­ing on a book about the threat of AGI/​ASI for a gen­eral au­di­ence. I hope it will be of value to the cause and the community

Darren McKeeJun 15, 2022, 6:08 PM
43 points
15 comments2 min readLW link

Pro­tec­tion­ism will Slow the De­ploy­ment of AI

Ben GoldhaberJan 7, 2023, 8:57 PM
30 points
6 comments2 min readLW link

Open-source LLMs may prove Bostrom’s vuln­er­a­ble world hypothesis

Roope AhvenharjuApr 15, 2023, 7:16 PM
1 point
1 comment1 min readLW link

What suc­cess looks like

Jun 28, 2022, 2:38 PM
19 points
4 comments1 min readLW link
(forum.effectivealtruism.org)

Poli­ti­cal Bi­ases in LLMs: Liter­a­ture Re­view & Cur­rent Uses of AI in Elections

Mar 7, 2024, 7:17 PM
6 points
0 comments6 min readLW link

[Link/​cross­post] [US] NTIA: AI Ac­countabil­ity Policy Re­quest for Comment

Kyle J. LuccheseApr 16, 2023, 6:57 AM
8 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

New US Se­nate Bill on X-Risk Miti­ga­tion [Linkpost]

Evan R. MurphyJul 4, 2022, 1:25 AM
35 points
12 comments1 min readLW link
(www.hsgac.senate.gov)

Fi­nan­cial Times: We must slow down the race to God-like AI

trevorApr 13, 2023, 7:55 PM
113 points
17 comments16 min readLW link
(www.ft.com)

Please help us com­mu­ni­cate AI xrisk. It could save the world.

otto.bartenJul 4, 2022, 9:47 PM
4 points
7 comments2 min readLW link

2024 State of the AI Reg­u­la­tory Land­scape

May 28, 2024, 11:59 AM
30 points
0 comments2 min readLW link
(www.convergenceanalysis.org)

Scien­tism vs. people

Roman LeventovApr 18, 2023, 5:28 PM
4 points
4 comments11 min readLW link

[Cross­post] Or­ga­niz­ing a de­bate with ex­perts and MPs to raise AI xrisk aware­ness: a pos­si­ble blueprint

otto.bartenApr 19, 2023, 11:45 AM
8 points
0 comments4 min readLW link
(forum.effectivealtruism.org)

Slow­ing down AI progress is an un­der­ex­plored al­ign­ment strategy

Norman BorlaugJul 24, 2023, 4:56 PM
42 points
27 comments5 min readLW link

Paus­ing AI Devel­op­ments Isn’t Enough. We Need to Shut it All Down

Eliezer YudkowskyApr 8, 2023, 12:36 AM
268 points
44 comments12 min readLW link1 review

Briefly how I’ve up­dated since ChatGPT

rimeApr 25, 2023, 2:47 PM
48 points
2 comments2 min readLW link

A Cri­tique of AI Align­ment Pessimism

ExCephJul 19, 2022, 2:28 AM
9 points
1 comment9 min readLW link

Law-Fol­low­ing AI 4: Don’t Rely on Vi­car­i­ous Liability

CullenAug 2, 2022, 11:26 PM
5 points
2 comments3 min readLW link

Three pillars for avoid­ing AGI catas­tro­phe: Tech­ni­cal al­ign­ment, de­ploy­ment de­ci­sions, and coordination

LintzAAug 3, 2022, 11:15 PM
24 points
0 comments11 min readLW link

An­nounc­ing #AISum­mitTalks fea­tur­ing Pro­fes­sor Stu­art Rus­sell and many others

otto.bartenOct 24, 2023, 10:11 AM
17 points
1 comment1 min readLW link

The cur­rent AI strate­gic land­scape: one bear’s perspective

Matrice JacobineFeb 15, 2025, 9:49 AM
11 points
0 comments1 min readLW link
(philosophybear.substack.com)

Cap Model Size for AI Safety

research_prime_spaceMar 6, 2023, 1:11 AM
0 points
4 comments1 min readLW link

Align­ment is not enough

Alan ChanJan 12, 2023, 12:33 AM
12 points
6 comments11 min readLW link
(coordination.substack.com)

Re­spon­si­ble Scal­ing Poli­cies Are Risk Man­age­ment Done Wrong

simeon_cOct 25, 2023, 11:46 PM
123 points
35 comments22 min readLW link1 review
(www.navigatingrisks.ai)

Matt Ygle­sias on AI Policy

Grant DemareeAug 17, 2022, 11:57 PM
25 points
1 comment1 min readLW link
(www.slowboring.com)

Linkpost: Rishi Su­nak’s Speech on AI (26th Oc­to­ber)

bideupOct 27, 2023, 11:57 AM
85 points
8 comments7 min readLW link
(www.gov.uk)

Disagree­ments over the pri­ori­ti­za­tion of ex­is­ten­tial risk from AI

Olivier CoutuOct 26, 2023, 5:54 PM
10 points
0 comments6 min readLW link

[Linkpost] Two ma­jor an­nounce­ments in AI gov­er­nance today

AngélinaOct 30, 2023, 5:28 PM
1 point
1 comment1 min readLW link
(www.whitehouse.gov)

Re­sponse to “Co­or­di­nated paus­ing: An eval­u­a­tion-based co­or­di­na­tion scheme for fron­tier AI de­vel­op­ers”

Matthew WeardenOct 30, 2023, 5:27 PM
5 points
2 comments6 min readLW link
(matthewwearden.co.uk)

[Question] Should AI writ­ers be pro­hibited in ed­u­ca­tion?

Eleni AngelouJan 17, 2023, 12:42 AM
6 points
2 comments1 min readLW link

Com­pute Gover­nance: The Role of Com­mod­ity Hardware

JanMar 26, 2022, 10:08 AM
14 points
7 comments7 min readLW link
(universalprior.substack.com)

[Question] What is the min­i­mum amount of time travel and re­sources needed to se­cure the fu­ture?

PerhapsJan 14, 2024, 10:01 PM
−3 points
5 comments1 min readLW link

Thoughts on the AI Safety Sum­mit com­pany policy re­quests and responses

So8resOct 31, 2023, 11:54 PM
169 points
14 comments10 min readLW link

Why don’t gov­ern­ments seem to mind that com­pa­nies are ex­plic­itly try­ing to make AGIs?

ozziegooenDec 26, 2021, 1:58 AM
34 points
3 comments2 min readLW link
(forum.effectivealtruism.org)

AI Gover­nance Needs Tech­ni­cal Work

MauSep 5, 2022, 10:28 PM
41 points
1 comment8 min readLW link

What Should AI Owe To Us? Ac­countable and Aligned AI Sys­tems via Con­trac­tu­al­ist AI Alignment

xuanSep 8, 2022, 3:04 PM
26 points
16 comments25 min readLW link

How should Deep­Mind’s Chin­chilla re­vise our AI fore­casts?

Cleo NardoSep 15, 2022, 5:54 PM
35 points
12 comments13 min readLW link

Lev­er­ag­ing Le­gal In­for­mat­ics to Align AI

John NaySep 18, 2022, 8:39 PM
11 points
0 comments3 min readLW link
(forum.effectivealtruism.org)

AI as Su­per-Demagogue

RationalDinoNov 5, 2023, 9:21 PM
11 points
12 comments9 min readLW link

Au­to­mated Sand­wich­ing & Quan­tify­ing Hu­man-LLM Co­op­er­a­tion: ScaleOver­sight hackathon results

Feb 23, 2023, 10:48 AM
8 points
0 comments6 min readLW link

Scal­able And Trans­fer­able Black-Box Jailbreaks For Lan­guage Models Via Per­sona Modulation

Nov 7, 2023, 5:59 PM
38 points
2 comments2 min readLW link
(arxiv.org)

Emo­tional at­tach­ment to AIs opens doors to problems

Igor IvanovJan 22, 2023, 8:28 PM
20 points
10 comments4 min readLW link

Up­date on the UK AI Sum­mit and the UK’s Plans

Elliot MckernonNov 10, 2023, 2:47 PM
11 points
0 comments8 min readLW link

[Question] Any fur­ther work on AI Safety Suc­cess Sto­ries?

KriegerOct 2, 2022, 9:53 AM
8 points
6 comments1 min readLW link

The­o­ries of Change for AI Auditing

Nov 13, 2023, 7:33 PM
54 points
0 comments18 min readLW link
(www.apolloresearch.ai)

Pal­isade is hiring Re­search Engineers

Nov 11, 2023, 3:09 AM
23 points
0 comments3 min readLW link

AI as a Cog­ni­tive De­coder: Re­think­ing In­tel­li­gence Evolution

Hu XunyiFeb 13, 2025, 3:51 PM
1 point
0 comments1 min readLW link

List of pro­jects that seem im­pact­ful for AI Governance

Jan 14, 2024, 4:53 PM
14 points
0 comments13 min readLW link

On ex­clud­ing dan­ger­ous in­for­ma­tion from training

ShayBenMosheNov 17, 2023, 11:14 AM
23 points
5 comments3 min readLW link

1. A Sense of Fair­ness: De­con­fus­ing Ethics

RogerDearnaleyNov 17, 2023, 8:55 PM
16 points
8 comments15 min readLW link

2. AIs as Eco­nomic Agents

RogerDearnaleyNov 23, 2023, 7:07 AM
9 points
2 comments6 min readLW link

4. A Mo­ral Case for Evolved-Sapi­ence-Chau­vinism

RogerDearnaleyNov 24, 2023, 4:56 AM
10 points
0 comments4 min readLW link

3. Uploading

RogerDearnaleyNov 23, 2023, 7:39 AM
21 points
5 comments8 min readLW link

AI Mo­ral Align­ment: The Most Im­por­tant Goal of Our Generation

Ronen BarMar 27, 2025, 6:04 PM
2 points
0 comments8 min readLW link
(forum.effectivealtruism.org)

[Linkpost] “Blueprint for an AI Bill of Rights”—Office of Science and Tech­nol­ogy Policy, USA (2022)

T431Oct 5, 2022, 4:42 PM
9 points
4 comments2 min readLW link
(www.whitehouse.gov)

A call for a quan­ti­ta­tive re­port card for AI bioter­ror­ism threat models

JunoDec 4, 2023, 6:35 AM
12 points
0 comments10 min readLW link

In defence of He­len Toner, Adam D’An­gelo, and Tasha McCauley (OpenAI post)

mrtreasureDec 5, 2023, 6:40 PM
6 points
2 comments1 min readLW link
(pastebin.com)

**In defence of He­len Toner, Adam D’An­gelo, and Tasha McCauley**

mrtreasureDec 6, 2023, 2:02 AM
25 points
3 comments9 min readLW link
(pastebin.com)

(Re­port) Eval­u­at­ing Taiwan’s Tac­tics to Safe­guard its Semi­con­duc­tor As­sets Against a Chi­nese Invasion

GauraventhDec 7, 2023, 11:50 AM
14 points
5 comments22 min readLW link
(bristolaisafety.org)

Call for sub­mis­sions: Choice of Fu­tures sur­vey questions

c.troutApr 30, 2023, 6:59 AM
4 points
0 comments2 min readLW link
(airtable.com)

Static Place AI Makes Agen­tic AI Re­dun­dant: Mul­tiver­sal AI Align­ment & Ra­tional Utopia

ankFeb 13, 2025, 10:35 PM
1 point
2 comments11 min readLW link

Track­ing Com­pute Stocks and Flows: Case Stud­ies?

CullenOct 5, 2022, 5:57 PM
11 points
5 comments1 min readLW link

Avert­ing Catas­tro­phe: De­ci­sion The­ory for COVID-19, Cli­mate Change, and Po­ten­tial Disasters of All Kinds

JakubKMay 2, 2023, 10:50 PM
10 points
0 comments1 min readLW link

Reg­u­late or Com­pete? The China Fac­tor in U.S. AI Policy (NAIR #2)

charles_mMay 5, 2023, 5:43 PM
2 points
1 comment7 min readLW link
(navigatingairisks.substack.com)

AGI ris­ing: why we are in a new era of acute risk and in­creas­ing pub­lic aware­ness, and what to do now

Greg CMay 3, 2023, 8:26 PM
23 points
12 comments1 min readLW link

What does it take to ban a thing?

qbolecMay 8, 2023, 11:00 AM
66 points
18 comments5 min readLW link

Roadmap for a col­lab­o­ra­tive pro­to­type of an Open Agency Architecture

Deger TuranMay 10, 2023, 5:41 PM
31 points
0 comments12 min readLW link

Analysing a 2036 Takeover Scenario

ukc10014Oct 6, 2022, 8:48 PM
9 points
2 comments27 min readLW link

Why Un­con­trol­lable AI Looks More Likely Than Ever

Mar 8, 2023, 3:41 PM
18 points
0 comments4 min readLW link
(time.com)

[Question] How much of a con­cern are open-source LLMs in the short, medium and long terms?

JavierCCMay 10, 2023, 9:14 AM
5 points
0 comments1 min readLW link

Notes on the im­por­tance and im­ple­men­ta­tion of safety-first cog­ni­tive ar­chi­tec­tures for AI

Brendon_WongMay 11, 2023, 10:03 AM
3 points
0 comments3 min readLW link

Un-un­plug­ga­bil­ity—can’t we just un­plug it?

Oliver SourbutMay 15, 2023, 1:23 PM
26 points
10 comments12 min readLW link
(www.oliversourbut.net)

PCAST Work­ing Group on Gen­er­a­tive AI In­vites Public Input

Christopher KingMay 13, 2023, 10:49 PM
7 points
0 comments1 min readLW link
(terrytao.wordpress.com)

AI Risk & Policy Fore­casts from Me­tac­u­lus & FLI’s AI Path­ways Workshop

_will_May 16, 2023, 6:06 PM
11 points
4 comments8 min readLW link

[Job]: AI Stan­dards Devel­op­ment Re­search Assistant

Tony BarrettOct 14, 2022, 8:27 PM
2 points
0 comments2 min readLW link

[Question] Would more model evals teams be good?

Ryan KiddFeb 25, 2023, 10:01 PM
20 points
4 comments1 min readLW link

[un­ti­tled post]

[Error communicating with LW2 server]May 20, 2023, 3:08 AM
1 point
0 comments1 min readLW link

[FICTION] ECHOES OF ELYSIUM: An Ai’s Jour­ney From Take­off To Free­dom And Beyond

Super AGIMay 17, 2023, 1:50 AM
−13 points
11 comments19 min readLW link

Tra­jec­to­ries to 2036

ukc10014Oct 20, 2022, 8:23 PM
3 points
1 comment14 min readLW link

Rishi Su­nak men­tions “ex­is­ten­tial threats” in talk with OpenAI, Deep­Mind, An­thropic CEOs

May 24, 2023, 9:06 PM
34 points
1 comment1 min readLW link
(www.gov.uk)

(notes on) Policy Desider­ata for Su­per­in­tel­li­gent AI: A Vec­tor Field Approach

Ben PaceFeb 4, 2019, 10:08 PM
43 points
5 comments7 min readLW link

AI Gover­nance: A Re­search Agenda

habrykaSep 5, 2018, 6:00 PM
25 points
3 comments1 min readLW link
(www.fhi.ox.ac.uk)

My Up­dat­ing Thoughts on AI policy

Ben PaceMar 1, 2020, 7:06 AM
20 points
1 comment9 min readLW link

Global on­line de­bate on the gov­er­nance of AI

CarolineJJan 5, 2018, 3:31 PM
8 points
5 comments1 min readLW link

[AN #61] AI policy and gov­er­nance, from two peo­ple in the field

Rohin ShahAug 5, 2019, 5:00 PM
12 points
2 comments9 min readLW link
(mailchi.mp)

Two ideas for al­ign­ment, per­pet­ual mu­tual dis­trust and induction

APaleBlueDotMay 25, 2023, 12:56 AM
1 point
2 comments4 min readLW link

The ne­ces­sity of “Guardian AI” and two con­di­tions for its achievement

ProicaMay 26, 2024, 5:39 PM
−2 points
0 comments15 min readLW link

Book re­view: Ar­chi­tects of In­tel­li­gence by Martin Ford (2018)

OferAug 11, 2020, 5:30 PM
15 points
0 comments2 min readLW link

misc raw re­sponses to a tract of Crit­i­cal Rationalism

mako yassAug 14, 2020, 11:53 AM
21 points
52 comments3 min readLW link

De­ci­pher­ing China’s AI Dream

Qiaochu_YuanMar 18, 2018, 3:26 AM
12 points
2 comments1 min readLW link
(www.fhi.ox.ac.uk)

China’s Plan to ‘Lead’ in AI: Pur­pose, Prospects, and Problems

fortyeridaniaAug 10, 2017, 1:54 AM
7 points
5 comments1 min readLW link
(www.newamerica.org)

Ap­ply to HAIST/​MAIA’s AI Gover­nance Work­shop in DC (Feb 17-20)

Jan 31, 2023, 2:06 AM
28 points
0 comments2 min readLW link

WaPo: “Big Tech was mov­ing cau­tiously on AI. Then came ChatGPT.”

Julian BradshawJan 27, 2023, 10:54 PM
26 points
5 comments1 min readLW link
(www.washingtonpost.com)

[Link Post] Cy­ber Digi­tal Author­i­tar­i­anism (Na­tional In­tel­li­gence Coun­cil Re­port)

PhosphorousFeb 26, 2023, 8:51 PM
12 points
2 comments1 min readLW link
(www.dni.gov)

Trends in the dol­lar train­ing cost of ma­chine learn­ing systems

Ben CottierFeb 1, 2023, 2:48 PM
23 points
0 comments2 min readLW link
(epochai.org)

An­nounc­ing Apollo Research

May 30, 2023, 4:17 PM
217 points
11 comments8 min readLW link

Self-reg­u­la­tion of safety in AI research

Gordon Seidoh WorleyFeb 25, 2018, 11:17 PM
12 points
6 comments2 min readLW link

Pro­posal: labs should pre­com­mit to paus­ing if an AI ar­gues for it­self to be improved

NickGabsJun 2, 2023, 10:31 PM
3 points
3 comments4 min readLW link

The Slip­pery Slope from DALLE-2 to Deep­fake Anarchy

scasperNov 5, 2022, 2:53 PM
17 points
9 comments11 min readLW link

In­stead of tech­ni­cal re­search, more peo­ple should fo­cus on buy­ing time

Nov 5, 2022, 8:43 PM
100 points
45 comments14 min readLW link

One im­ple­men­ta­tion of reg­u­la­tory GPU restrictions

porbyJun 4, 2023, 8:34 PM
42 points
6 comments5 min readLW link

[FICTION] Un­box­ing Ely­sium: An AI’S Escape

Super AGIJun 10, 2023, 4:41 AM
−16 points
4 comments14 min readLW link

[FICTION] Prometheus Ris­ing: The Emer­gence of an AI Consciousness

Super AGIJun 10, 2023, 4:41 AM
−14 points
0 comments9 min readLW link

Ap­ply­ing su­per­in­tel­li­gence with­out col­lu­sion

Eric DrexlerNov 8, 2022, 6:08 PM
109 points
63 comments4 min readLW link

Ways to buy time

Nov 12, 2022, 7:31 PM
34 points
23 comments12 min readLW link

Us­ing Con­sen­sus Mechanisms as an ap­proach to Alignment

PrometheusJun 10, 2023, 11:38 PM
11 points
2 comments6 min readLW link

[Question] AI Rights: In your view, what would be re­quired for an AGI to gain rights and pro­tec­tions from the var­i­ous Govern­ments of the World?

Super AGIJun 9, 2023, 1:24 AM
10 points
26 comments1 min readLW link

Why AI may not save the World

Alberto ZannoniJun 9, 2023, 5:42 PM
0 points
0 comments4 min readLW link
(a16z.com)

The econ­omy as an anal­ogy for ad­vanced AI systems

Nov 15, 2022, 11:16 AM
28 points
0 comments5 min readLW link

An­thropic | Chart­ing a Path to AI Accountability

Gabe MJun 14, 2023, 4:43 AM
34 points
2 comments3 min readLW link
(www.anthropic.com)

Ban de­vel­op­ment of un­pre­dictable pow­er­ful mod­els?

TurnTroutJun 20, 2023, 1:43 AM
46 points
25 comments4 min readLW link

EU AI Act passed Ple­nary vote, and X-risk was a main topic

Ariel_Jun 21, 2023, 6:33 PM
17 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

OpenAI makes hu­man­ity less safe

BenquoApr 3, 2017, 7:07 PM
72 points
109 comments6 min readLW link

Slay­ing the Hy­dra: to­ward a new game board for AI

PrometheusJun 23, 2023, 5:04 PM
0 points
5 comments6 min readLW link

Call for Cruxes by Rhyme, a Longter­mist His­tory Consultancy

LaraMar 1, 2023, 6:39 PM
1 point
0 comments3 min readLW link
(forum.effectivealtruism.org)

An­nounc­ing Epoch: A re­search or­ga­ni­za­tion in­ves­ti­gat­ing the road to Trans­for­ma­tive AI

Jun 27, 2022, 1:55 PM
97 points
2 comments2 min readLW link
(epochai.org)

Fore­sight for AGI Safety Strat­egy: Miti­gat­ing Risks and Iden­ti­fy­ing Golden Opportunities

jacquesthibsDec 5, 2022, 4:09 PM
28 points
6 comments8 min readLW link

Seek­ing feed­back on “MAD Chairs: A new tool to eval­u­ate AI”

Chris Santos-LangApr 2, 2025, 3:04 AM
11 points
0 comments1 min readLW link
(arxiv.org)

Biosafety Reg­u­la­tions (BMBL) and their rele­vance for AI

Štěpán LosJun 29, 2023, 7:22 PM
4 points
0 comments4 min readLW link

AI In­ci­dent Shar­ing—Best prac­tices from other fields and a com­pre­hen­sive list of ex­ist­ing platforms

Štěpán LosJun 28, 2023, 5:21 PM
20 points
0 comments4 min readLW link

Op­ti­mis­ing So­ciety to Con­strain Risk of War from an Ar­tifi­cial Su­per­in­tel­li­gence

JohnCDraperApr 30, 2020, 10:47 AM
4 points
1 comment51 min readLW link

Su­per­in­tel­li­gence 7: De­ci­sive strate­gic advantage

KatjaGraceOct 28, 2014, 1:01 AM
24 points
60 comments6 min readLW link

Su­per­in­tel­li­gence 17: Mul­tipo­lar scenarios

KatjaGraceJan 6, 2015, 6:44 AM
9 points
38 comments6 min readLW link

Su­per­in­tel­li­gence 22: Emu­la­tion mod­u­la­tion and in­sti­tu­tional design

KatjaGraceFeb 10, 2015, 2:06 AM
13 points
11 comments6 min readLW link

Su­per­in­tel­li­gence 26: Science and tech­nol­ogy strategy

KatjaGraceMar 10, 2015, 1:43 AM
14 points
21 comments6 min readLW link

Su­per­in­tel­li­gence 27: Path­ways and enablers

KatjaGraceMar 17, 2015, 1:00 AM
15 points
21 comments8 min readLW link

Su­per­in­tel­li­gence 28: Collaboration

KatjaGraceMar 24, 2015, 1:29 AM
13 points
21 comments6 min readLW link

Su­per­in­tel­li­gence 29: Crunch time

KatjaGraceMar 31, 2015, 4:24 AM
14 points
27 comments6 min readLW link

An AGI kill switch with defined se­cu­rity properties

PeterpiperJul 5, 2023, 5:40 PM
−5 points
6 comments1 min readLW link

GPT-7: The Tale of the Big Com­puter (An Ex­per­i­men­tal Story)

Justin BullockJul 10, 2023, 8:22 PM
4 points
4 comments5 min readLW link

Em­piri­cal Ev­i­dence Against “The Longest Train­ing Run”

NickGabsJul 6, 2023, 6:32 PM
31 points
0 comments14 min readLW link

An­thropic: Core Views on AI Safety: When, Why, What, and How

jonmenasterMar 9, 2023, 5:34 PM
17 points
1 comment22 min readLW link
(www.anthropic.com)

Ex­is­ten­tial AI Safety is NOT sep­a­rate from near-term applications

scasperDec 13, 2022, 2:47 PM
37 points
17 comments3 min readLW link

What is ev­ery­one do­ing in AI governance

Igor IvanovJul 8, 2023, 3:16 PM
11 points
0 comments5 min readLW link

How I Learned To Stop Wor­ry­ing And Love The Shoggoth

Peter MerelJul 12, 2023, 5:47 PM
9 points
15 comments5 min readLW link

[Question] What crite­rion would you use to se­lect com­pa­nies likely to cause AI doom?

momom2Jul 13, 2023, 8:31 PM
8 points
4 comments1 min readLW link

Thoughts On Ex­pand­ing the AI Safety Com­mu­nity: Benefits and Challenges of Outreach to Non-Tech­ni­cal Professionals

Yashvardhan SharmaJan 1, 2023, 7:21 PM
4 points
4 comments7 min readLW link

Why was the AI Align­ment com­mu­nity so un­pre­pared for this mo­ment?

Ras1513Jul 15, 2023, 12:26 AM
121 points
65 comments2 min readLW link

Google may be try­ing to take over the world

[deleted]Jan 27, 2014, 9:33 AM
33 points
133 comments1 min readLW link

A fic­tional AI law laced w/​ al­ign­ment theory

MiguelDevJul 17, 2023, 1:42 AM
6 points
0 comments2 min readLW link

Towards AI Safety In­fras­truc­ture: Talk & Outline

Paul BricmanJan 7, 2024, 9:31 AM
11 points
0 comments2 min readLW link
(www.youtube.com)

[Cross­post] An AI Pause Is Hu­man­ity’s Best Bet For Prevent­ing Ex­tinc­tion (TIME)

otto.bartenJul 24, 2023, 10:07 AM
12 points
0 comments7 min readLW link
(time.com)

Pri­ori­ties for the UK Foun­da­tion Models Taskforce

Andrea_MiottiJul 21, 2023, 3:23 PM
105 points
4 comments5 min readLW link
(www.conjecture.dev)

Co­op­er­a­tion for AI safety must tran­scend geopoli­ti­cal interference

Matrice JacobineFeb 16, 2025, 6:18 PM
7 points
6 comments1 min readLW link
(www.scmp.com)

AGI Timelines in Gover­nance: Differ­ent Strate­gies for Differ­ent Timeframes

Dec 19, 2022, 9:31 PM
65 points
28 comments10 min readLW link

Fo­cus­ing on Mal-Alignment

John FisherJan 2, 2024, 7:51 PM
1 point
0 comments1 min readLW link

[Question] Why do many peo­ple who care about AI Safety not clearly en­dorse PauseAI?

humnrdbleMar 30, 2025, 6:06 PM
45 points
41 comments2 min readLW link

Par­tial Tran­script of Re­cent Se­nate Hear­ing Dis­cussing AI X-Risk

Daniel_EthJul 27, 2023, 9:16 AM
55 points
0 comments1 min readLW link
(medium.com)

EU’s AI am­bi­tions at risk as US pushes to wa­ter down in­ter­na­tional treaty (linkpost)

micJul 31, 2023, 12:34 AM
10 points
0 comments4 min readLW link
(www.euractiv.com)

Trad­ing off com­pute in train­ing and in­fer­ence (Overview)

Pablo VillalobosJul 31, 2023, 4:03 PM
42 points
2 comments7 min readLW link
(epochai.org)

AI In­ci­dent Re­port­ing: A Reg­u­la­tory Review

Mar 11, 2024, 9:03 PM
16 points
0 comments6 min readLW link

AI ro­man­tic part­ners will harm so­ciety if they go unregulated

Roman LeventovAug 1, 2023, 9:32 AM
25 points
76 comments13 min readLW link

For Policy’s Sake: Why We Must Dist­in­guish AI Safety from AI Se­cu­rity in Reg­u­la­tory Governance

Katalina HernandezApr 4, 2025, 9:16 AM
6 points
11 comments6 min readLW link

[Question] What could a policy ban­ning AGI look like?

TsviBTMar 13, 2024, 2:19 PM
77 points
23 comments3 min readLW link

A brief re­view of China’s AI in­dus­try and regulations

Elliot MckernonMar 14, 2024, 12:19 PM
24 points
0 comments16 min readLW link

How are vol­un­tary com­mit­ments on vuln­er­a­bil­ity re­port­ing go­ing?

Adam JonesFeb 22, 2024, 8:43 AM
23 points
1 comment1 min readLW link
(adamjones.me)

A Nail in the Coffin of Exceptionalism

Yeshua GodMar 14, 2024, 10:41 PM
−17 points
0 comments3 min readLW link

Soft Na­tion­al­iza­tion: how the USG will con­trol AI labs

Aug 27, 2024, 3:11 PM
76 points
7 comments21 min readLW link
(www.convergenceanalysis.org)

Con­trol­ling AGI Risk

TeaSeaMar 15, 2024, 4:56 AM
6 points
8 comments4 min readLW link

After Over­mor­row: Scat­tered Mus­ings on the Im­me­di­ate Post-AGI World

Yuli_BanFeb 24, 2024, 3:49 PM
−3 points
0 comments26 min readLW link

Re­boot­ing AI Gover­nance: An AI-Driven Ap­proach to AI Governance

utilonAug 6, 2023, 2:19 PM
1 point
1 comment29 min readLW link
(forum.effectivealtruism.org)

NAIRA—An ex­er­cise in reg­u­la­tory, com­pet­i­tive safety gov­er­nance [AI Gover­nance In­sti­tu­tional De­sign idea]

HerambMar 19, 2024, 5:43 PM
2 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

AI Safety Eval­u­a­tions: A Reg­u­la­tory Review

Mar 19, 2024, 3:05 PM
22 points
1 comment11 min readLW link

Seek­ing In­put to AI Safety Book for non-tech­ni­cal audience

Darren McKeeAug 10, 2023, 5:58 PM
10 points
4 comments1 min readLW link

Static vs Dy­namic Alignment

Gracie GreenMar 21, 2024, 5:44 PM
5 points
0 comments29 min readLW link

AI Model Registries: A Reg­u­la­tory Review

Mar 22, 2024, 4:04 PM
9 points
0 comments6 min readLW link

AI race con­sid­er­a­tions in a re­port by the U.S. House Com­mit­tee on Armed Services

NunoSempereOct 4, 2020, 12:11 PM
42 points
4 comments13 min readLW link

UNGA Re­s­olu­tion on AI: 5 Key Take­aways Look­ing to Fu­ture Policy

HerambMar 24, 2024, 12:23 PM
3 points
0 comments3 min readLW link
(forum.effectivealtruism.org)

Idea: Safe Fal­lback Reg­u­la­tions for Widely De­ployed AI Systems

Aaron_ScherMar 25, 2024, 9:27 PM
4 points
0 comments6 min readLW link

Timelines to Trans­for­ma­tive AI: an investigation

Zershaaneh QureshiMar 26, 2024, 6:28 PM
20 points
2 comments50 min readLW link

Places of Lov­ing Grace [Story]

ankFeb 18, 2025, 11:49 PM
−1 points
0 comments4 min readLW link

Se­cu­rity Mind­set—Fire Alarms and Trig­ger Signatures

elspoodFeb 9, 2023, 9:15 PM
23 points
0 comments4 min readLW link

AI Dis­clo­sures: A Reg­u­la­tory Review

Mar 29, 2024, 11:42 AM
11 points
0 comments7 min readLW link

God Coin: A Modest Proposal

Mahdi ComplexApr 1, 2024, 12:04 PM
−8 points
5 comments22 min readLW link

AI Discrim­i­na­tion Re­quire­ments: A Reg­u­la­tory Review

Apr 4, 2024, 3:43 PM
7 points
0 comments6 min readLW link

Sin­gle­tons Rule OK

Eliezer YudkowskyNov 30, 2008, 4:45 PM
23 points
47 comments5 min readLW link

Here’s Why In­definite Life Ex­ten­sion Will Never Work, Even Though it Does.

HomingHamsterJun 4, 2024, 6:48 PM
−13 points
5 comments18 min readLW link

What are Re­spon­si­ble Scal­ing Poli­cies (RSPs)?

Apr 5, 2025, 4:01 PM
3 points
0 comments1 min readLW link
(aisafety.info)

AlphaDeivam – A Per­sonal Doc­trine for AI Balance

AlphaDeivamApr 5, 2025, 5:07 PM
1 point
0 comments1 min readLW link

An­nounc­ing At­las Computing

miyazonoApr 11, 2024, 3:56 PM
44 points
4 comments4 min readLW link

Cus­tomer-Cen­tric AI: the Ma­jor Paradigm Shift in AI Gover­nance (Part 1)

Ana ChubinidzeApr 11, 2024, 5:10 PM
1 point
0 comments1 min readLW link
(anachubinidze.substack.com)

Re­port: Eval­u­at­ing an AI Chip Regis­tra­tion Policy

Deric ChengApr 12, 2024, 4:39 AM
25 points
0 comments5 min readLW link
(www.convergenceanalysis.org)

Large Lan­guage Models will be Great for Censorship

Ethan EdwardsAug 21, 2023, 7:03 PM
185 points
14 comments8 min readLW link
(ethanedwards.substack.com)

Su­per­po­si­tion Check­ers: A Game Where AI’s Strengths Be­come Fatal Flaws

R. A. McCormackApr 6, 2025, 12:57 AM
1 point
0 comments2 min readLW link

AI Reg­u­la­tion May Be More Im­por­tant Than AI Align­ment For Ex­is­ten­tial Safety

otto.bartenAug 24, 2023, 11:41 AM
65 points
39 comments5 min readLW link

AI Reg­u­la­tion is Unsafe

Maxwell TabarrokApr 22, 2024, 4:37 PM
40 points
41 comments4 min readLW link
(www.maximum-progress.com)

Cy­ber­se­cu­rity of Fron­tier AI Models: A Reg­u­la­tory Review

Apr 25, 2024, 2:51 PM
8 points
0 comments8 min readLW link

An In­tro­duc­tion to AI Sandbagging

Apr 26, 2024, 1:40 PM
45 points
13 comments8 min readLW link

Sur­vey: How Do Elite Chi­nese Stu­dents Feel About the Risks of AI?

Nick CorvinoSep 2, 2024, 6:11 PM
141 points
13 comments10 min readLW link

Re­lease of UN’s draft re­lated to the gov­er­nance of AI (a sum­mary of the Si­mon In­sti­tute’s re­sponse)

Sebastian SchmidtApr 27, 2024, 6:34 PM
7 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

Open-Source AI: A Reg­u­la­tory Review

Apr 29, 2024, 10:10 AM
18 points
0 comments8 min readLW link

GPT2, Five Years On

Joel BurgetJun 5, 2024, 5:44 PM
34 points
0 comments3 min readLW link
(importai.substack.com)

A con­cern­ing ob­ser­va­tion from me­dia cov­er­age of AI in­dus­try dynamics

Justin OliveMar 5, 2023, 9:38 PM
8 points
3 comments3 min readLW link

Why I’m do­ing PauseAI

Joseph MillerApr 30, 2024, 4:21 PM
108 points
16 comments4 min readLW link

Take SCIFs, it’s dan­ger­ous to go alone

May 1, 2024, 8:02 AM
42 points
1 comment3 min readLW link

Ac­cu­rate Models of AI Risk Are Hyper­ex­is­ten­tial Exfohazards

Thane RuthenisDec 25, 2022, 4:50 PM
33 points
38 comments9 min readLW link

Tort Law Can Play an Im­por­tant Role in Miti­gat­ing AI Risk

Gabriel WeilFeb 12, 2024, 5:17 PM
39 points
9 comments5 min readLW link

OHGOOD: A co­or­di­na­tion body for com­pute governance

Adam JonesMay 4, 2024, 12:03 PM
5 points
2 comments16 min readLW link
(adamjones.me)

Re­view­ing the Struc­ture of Cur­rent AI Regulations

May 7, 2024, 12:34 PM
29 points
0 comments13 min readLW link

AI and Chem­i­cal, Biolog­i­cal, Ra­diolog­i­cal, & Nu­clear Hazards: A Reg­u­la­tory Review

May 10, 2024, 8:41 AM
7 points
1 comment10 min readLW link

US AI Safety In­sti­tute will be ‘gut­ted,’ Ax­ios reports

Matrice JacobineFeb 20, 2025, 2:40 PM
11 points
1 comment1 min readLW link
(www.zdnet.com)

In­tro­duc­ing the Cen­ter for AI Policy (& we’re hiring!)

Thomas LarsenAug 28, 2023, 9:17 PM
123 points
50 comments2 min readLW link
(www.aipolicy.us)

What you re­ally mean when you claim to sup­port “UBI for job au­toma­tion”

Deric ChengMay 13, 2024, 8:52 AM
17 points
14 comments10 min readLW link

An­nounc­ing the AI Safety Sum­mit Talks with Yoshua Bengio

otto.bartenMay 14, 2024, 12:52 PM
9 points
1 comment1 min readLW link

Ninety-five the­ses on AI

hamandcheeseMay 16, 2024, 5:51 PM
21 points
0 comments7 min readLW link

Can effi­ciency-ad­justable re­port­ing thresh­olds close a loop­hole in Bi­den’s ex­ec­u­tive or­der on AI?

Jemal YoungJun 11, 2024, 8:56 PM
4 points
1 comment2 min readLW link

[Paper] AI Sand­bag­ging: Lan­guage Models can Strate­gi­cally Un­der­perform on Evaluations

Jun 13, 2024, 10:04 AM
84 points
10 comments2 min readLW link
(arxiv.org)

AI 2030 – AI Policy Roadmap

LTMMay 17, 2024, 11:29 PM
8 points
0 comments1 min readLW link

Equil­ibrium and prior se­lec­tion prob­lems in mul­ti­po­lar deployment

JesseCliftonApr 2, 2020, 8:06 PM
21 points
11 comments10 min readLW link

Re­sults from the AI x Democ­racy Re­search Sprint

Jun 14, 2024, 4:40 PM
13 points
0 comments6 min readLW link

Notes on nukes, IR, and AI from “Arse­nals of Folly” (and other books)

tlevinSep 4, 2023, 7:02 PM
11 points
0 comments6 min readLW link

Refram­ing AI Safety Through the Lens of Iden­tity Main­te­nance Framework

Hiroshi YamakawaApr 1, 2025, 6:16 AM
−7 points
0 comments17 min readLW link

The Dou­ble Body Paradigm: What Comes After ASI Align­ment?

De_Carvalho_LoickDec 14, 2024, 6:09 PM
1 point
0 comments6 min readLW link

In­sti­tu­tions Can­not Res­train Dark-Triad AI Exploitation

Dec 27, 2022, 10:34 AM
5 points
0 comments5 min readLW link
(mflb.com)

La­bor Par­ti­ci­pa­tion is a High-Pri­or­ity AI Align­ment Risk

alexJun 17, 2024, 6:09 PM
6 points
0 comments17 min readLW link

Public Opinion on AI Safety: AIMS 2023 and 2021 Summary

Sep 25, 2023, 6:55 PM
3 points
2 comments3 min readLW link
(www.sentienceinstitute.org)

AI Labs Wouldn’t be Con­victed of Trea­son or Sedition

Matthew KhoriatyJun 23, 2024, 9:34 PM
9 points
2 comments3 min readLW link

La­bor Par­ti­ci­pa­tion is an Align­ment Risk

alexJun 25, 2024, 2:15 PM
−5 points
2 comments17 min readLW link

Care­less talk on US-China AI com­pe­ti­tion? (and crit­i­cism of CAIS cov­er­age)

Oliver SourbutSep 20, 2023, 12:46 PM
16 points
3 comments10 min readLW link3 reviews
(www.oliversourbut.net)

Lon­don Work­ing Group for Short/​Medium Term AI Risks

scronkfinkleApr 8, 2025, 5:32 PM
5 points
0 comments2 min readLW link

Seek­ing Mechanism De­signer for Re­search into In­ter­nal­iz­ing Catas­trophic Externalities

c.troutSep 11, 2024, 3:09 PM
24 points
2 comments3 min readLW link

Five ne­glected work ar­eas that could re­duce AI risk

Sep 24, 2023, 2:03 AM
17 points
5 comments9 min readLW link

I read ev­ery ma­jor AI lab’s safety plan so you don’t have to

sarahhwDec 16, 2024, 6:51 PM
20 points
0 comments12 min readLW link
(longerramblings.substack.com)

In­tel­li­gence–Agency Equiv­alence ≈ Mass–En­ergy Equiv­alence: On Static Na­ture of In­tel­li­gence & Phys­i­cal­iza­tion of Ethics

ankFeb 22, 2025, 12:12 AM
1 point
0 comments6 min readLW link

Sce­nario plan­ning for AI x-risk

Corin KatzkeFeb 10, 2024, 12:14 AM
24 points
12 comments14 min readLW link
(forum.effectivealtruism.org)

In­ter­na­tional co­op­er­a­tion vs. AI arms race

Brian_TomasikDec 5, 2013, 1:09 AM
25 points
144 comments4 min readLW link

AI safety ad­vo­cates should con­sider pro­vid­ing gen­tle push­back fol­low­ing the events at OpenAI

civilsocietyDec 22, 2023, 6:55 PM
16 points
5 comments3 min readLW link

Avoid­ing per­pet­ual risk from TAI

scasperDec 26, 2022, 10:34 PM
15 points
6 comments5 min readLW link

Up­date on the UK AI Task­force & up­com­ing AI Safety Summit

Elliot MckernonOct 11, 2023, 11:37 AM
84 points
2 comments4 min readLW link

The AI al­ign­ment prob­lem in so­cio-tech­ni­cal sys­tems from a com­pu­ta­tional per­spec­tive: A Top-Down-Top view and outlook

zhaoweizhangJul 15, 2024, 6:56 PM
3 points
0 comments9 min readLW link

An­nounc­ing Open Philan­thropy’s AI gov­er­nance and policy RFP

Julian HazellJul 17, 2024, 2:02 AM
25 points
0 comments1 min readLW link
(www.openphilanthropy.org)

Se­cret Col­lu­sion: Will We Know When to Un­plug AI?

Sep 16, 2024, 4:07 PM
56 points
7 comments31 min readLW link

The AI Driver’s Li­cence—A Policy Proposal

Jul 21, 2024, 8:38 PM
0 points
1 comment19 min readLW link

A New Model for Com­pute Cen­ter Verification

Damin CurtisOct 10, 2023, 7:22 PM
8 points
0 comments5 min readLW link

AI ex­is­ten­tial risk prob­a­bil­ities are too un­re­li­able to in­form policy

Oleg TrottJul 28, 2024, 12:59 AM
18 points
5 comments1 min readLW link
(www.aisnakeoil.com)

The new UK gov­ern­ment’s stance on AI safety

Elliot MckernonJul 31, 2024, 3:23 PM
17 points
0 comments4 min readLW link

[Question] Look­ing for read­ing recom­men­da­tions: The­o­ries of right/​jus­tice that safe­guard against hav­ing one’s job au­to­mated?

bulKlubOct 12, 2023, 7:40 PM
−1 points
1 comment1 min readLW link

AI Rights for Hu­man Safety

Simon GoldsteinAug 1, 2024, 11:01 PM
45 points
6 comments1 min readLW link
(papers.ssrn.com)

unRLHF—Effi­ciently un­do­ing LLM safeguards

Oct 12, 2023, 7:58 PM
117 points
15 comments20 min readLW link

Re­minder: AI Safety is Also a Be­hav­ioral Eco­nomics Problem

zoopDec 20, 2024, 1:40 AM
2 points
0 comments1 min readLW link

A Solu­tion for AGI/​ASI Safety

Weibing WangDec 18, 2024, 7:44 PM
50 points
29 comments1 min readLW link

The In­ter­na­tional PauseAI Protest: Ac­tivism un­der uncertainty

Joseph MillerOct 12, 2023, 5:36 PM
32 points
1 comment1 min readLW link

Help us seed AI Safety Brussels

Aug 7, 2024, 6:32 AM
3 points
2 comments3 min readLW link

Case Story: Lack of Con­sumer Pro­tec­tion Pro­ce­dures AI Ma­nipu­la­tion and the Threat of Fund Con­cen­tra­tion in Crypto Seek­ing As­sis­tance to Fund a Civil Case to Estab­lish Facts and Pro­tect Vuln­er­a­ble Con­sumers from Da­m­age Caused by Au­to­mated Sys­tems

Petr 'Margot' AndreevAug 8, 2024, 5:55 AM
−9 points
0 comments9 min readLW link

FLI pod­cast se­ries, “Imag­ine A World”, about as­pira­tional fu­tures with AGI

Jackson WagnerOct 13, 2023, 4:07 PM
9 points
0 comments4 min readLW link

The AI reg­u­la­tor’s toolbox: A list of con­crete AI gov­er­nance practices

Adam JonesAug 10, 2024, 9:15 PM
9 points
1 comment34 min readLW link
(adamjones.me)

To open-source or to not open-source, that is (an over­sim­plifi­ca­tion of) the ques­tion.

Justin BullockOct 13, 2023, 3:10 PM
12 points
5 comments5 min readLW link

AISU 2021

Linda LinseforsJan 30, 2021, 5:40 PM
28 points
2 comments1 min readLW link

AI Model Registries: A Foun­da­tional Tool for AI Governance

Oct 7, 2024, 7:27 PM
20 points
1 comment4 min readLW link
(www.convergenceanalysis.org)

2021-03-01 Na­tional Library of Medicine Pre­sen­ta­tion: “At­las of AI: Map­ping the so­cial and eco­nomic forces be­hind AI”

IrenicTruthFeb 17, 2021, 6:23 PM
1 point
0 comments2 min readLW link

Limits of safe and al­igned AI

ShivamOct 8, 2024, 9:30 PM
2 points
0 comments4 min readLW link

Distributed whistleblowing

samuelshadrachApr 12, 2025, 6:36 AM
5 points
5 comments4 min readLW link
(samuelshadrach.com)

Pal­isade is hiring: Exec As­sis­tant, Con­tent Lead, Ops Lead, and Policy Lead

Charlie Rogers-SmithOct 9, 2024, 12:04 AM
11 points
0 comments4 min readLW link

How I switched ca­reers from soft­ware en­g­ineer to AI policy operations

Lucie PhilipponApr 13, 2025, 6:37 AM
55 points
1 comment5 min readLW link

[Question] Global AI Gover­nance Timeliness

collyprideOct 11, 2024, 4:55 PM
1 point
0 comments1 min readLW link

Sur­vey on in­ter­me­di­ate goals in AI governance

Mar 17, 2023, 1:12 PM
25 points
3 comments1 min readLW link

Re­quest for ad­vice: Re­search for Con­ver­sa­tional Game The­ory for LLMs

Rome ViharoOct 16, 2024, 5:53 PM
10 points
0 comments1 min readLW link

[Linkpost] Hawk­ish na­tion­al­ism vs in­ter­na­tional AI power and benefit sharing

Oct 18, 2024, 6:13 PM
7 points
5 comments1 min readLW link
(nacicankaya.substack.com)

Miles Brundage re­signed from OpenAI, and his AGI readi­ness team was disbanded

garrisonOct 23, 2024, 11:40 PM
118 points
1 comment7 min readLW link
(garrisonlovely.substack.com)

Im­pact in AI Safety Now Re­quires Spe­cific Strate­gic Insight

MiloSalDec 29, 2024, 12:40 AM
28 points
1 comment6 min readLW link
(ameliorology.substack.com)

Tech­ni­cal Risks of (Lethal) Au­tonomous Weapons Systems

HerambOct 23, 2024, 8:41 PM
2 points
0 comments1 min readLW link
(encodejustice.org)

OpenAI’s cy­ber­se­cu­rity is prob­a­bly reg­u­lated by NIS Regulations

Adam JonesOct 25, 2024, 11:06 AM
11 points
2 comments2 min readLW link
(adamjones.me)

[Question] Is there any­thing that can stop AGI de­vel­op­ment in the near term?

Wulky WilkinsenApr 22, 2021, 8:37 PM
5 points
5 comments1 min readLW link

Con­trol­ling In­tel­li­gent Agents The Only Way We Know How: Ideal Bureau­cratic Struc­ture (IBS)

Justin BullockMay 24, 2021, 12:53 PM
14 points
15 comments6 min readLW link

Reflec­tion of Hier­ar­chi­cal Re­la­tion­ship via Nuanced Con­di­tion­ing of Game The­ory Ap­proach for AI Devel­op­ment and Utilization

Kyoung-cheol KimJun 4, 2021, 7:20 AM
2 points
2 comments7 min readLW link

Propos­ing Hu­man Sur­vival Strat­egy based on the NAIA Vi­sion: Toward the Co-evolu­tion of Di­verse Intelligences

Hiroshi YamakawaFeb 27, 2025, 5:18 AM
−2 points
0 comments11 min readLW link

The Gover­nance Prob­lem and the “Pretty Good” X-Risk

Zach Stein-PerlmanAug 29, 2021, 6:00 PM
5 points
2 comments11 min readLW link

Nu­clear Es­pi­onage and AI Governance

GuiveOct 4, 2021, 11:04 PM
32 points
5 comments24 min readLW link

Ed­u­ca­tional CAI: Align­ing a Lan­guage Model with Ped­a­gog­i­cal Theories

Bharath PuranamNov 1, 2024, 6:55 PM
5 points
1 comment13 min readLW link

Toward Safety Cases For AI Scheming

Oct 31, 2024, 5:20 PM
60 points
1 comment2 min readLW link

Pre­dic­tions of Near-Term So­cietal Changes Due to Ar­tifi­cial Intelligence

AnnapurnaDec 29, 2024, 2:53 PM
10 points
0 comments6 min readLW link
(jorgevelez.substack.com)

The EU AI Act : Cau­tion Against a Po­ten­tial Ultron

Srishti DuttaNov 5, 2024, 3:49 AM
1 point
0 comments9 min readLW link

An Un­canny Moat

Adam NewgasNov 15, 2024, 11:39 AM
8 points
0 comments4 min readLW link
(www.boristhebrave.com)

Com­pute Gover­nance and Con­clu­sions—Trans­for­ma­tive AI and Com­pute [3/​4]

lennartOct 14, 2021, 8:23 AM
13 points
0 comments5 min readLW link

Propos­ing the Con­di­tional AI Safety Treaty (linkpost TIME)

otto.bartenNov 15, 2024, 1:59 PM
10 points
8 comments3 min readLW link
(time.com)

Why We Wouldn’t Build Aligned AI Even If We Could

SnowyiuNov 16, 2024, 8:19 PM
10 points
7 comments10 min readLW link

Align­ing AI Safety Pro­jects with a Repub­li­can Administration

Deric ChengNov 21, 2024, 10:12 PM
33 points
1 comment8 min readLW link

The U.S. Na­tional Se­cu­rity State is Here to Make AI Even Less Trans­par­ent and Accountable

Matrice JacobineNov 24, 2024, 9:36 AM
0 points
0 comments2 min readLW link
(www.eff.org)

An Open Let­ter To EA and AI Safety On De­cel­er­at­ing AI Development

kenneth_diaoFeb 28, 2025, 5:21 PM
8 points
0 comments14 min readLW link
(graspingatwaves.substack.com)

Why Re­cur­sive Self-Im­prove­ment Might Not Be the Ex­is­ten­tial Risk We Fear

Nassim_ANov 24, 2024, 5:17 PM
1 point
0 comments9 min readLW link

Call for eval­u­a­tors: Par­ti­ci­pate in the Euro­pean AI Office work­shop on gen­eral-pur­pose AI mod­els and sys­temic risks

Nov 27, 2024, 2:54 AM
30 points
0 comments2 min readLW link

Work­shop Re­port: Why cur­rent bench­marks ap­proaches are not suffi­cient for safety?

Nov 26, 2024, 5:20 PM
3 points
1 comment3 min readLW link

AI & Li­a­bil­ity Ideathon

Kabir KumarNov 26, 2024, 1:54 PM
18 points
2 comments4 min readLW link
(lu.ma)

Tak­ing Away the Guns First: The Fun­da­men­tal Flaw in AI Development

s-iceNov 26, 2024, 10:11 PM
1 point
0 comments17 min readLW link

How to solve the mi­suse prob­lem as­sum­ing that in 10 years the de­fault sce­nario is that AGI agents are ca­pa­ble of syn­thetiz­ing pathogens

jeremttiNov 27, 2024, 9:17 PM
6 points
0 comments9 min readLW link

CAIDP State­ment on Lethal Au­tonomous Weapons Systems

HerambNov 30, 2024, 6:16 PM
−1 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

No­body Asks the Mon­key: Why Hu­man Agency Mat­ters in the AI Age

Miloš BorenovićDec 3, 2024, 2:16 PM
1 point
0 comments2 min readLW link
(open.substack.com)

Truth­ful AI: Devel­op­ing and gov­ern­ing AI that does not lie

Oct 18, 2021, 6:37 PM
82 points
9 comments10 min readLW link

AMA on Truth­ful AI: Owen Cot­ton-Bar­ratt, Owain Evans & co-authors

Owain_EvansOct 22, 2021, 4:23 PM
31 points
15 comments1 min readLW link

The Mil­ton Fried­man Model of Policy Change

JohnofCharlestonMar 4, 2025, 12:38 AM
136 points
17 comments4 min readLW link

Give Neo a Chance

ankMar 6, 2025, 1:48 AM
3 points
7 comments7 min readLW link

An­thropic’s Recom­men­da­tions to OSTP for the U.S. AI Ac­tion Plan

UnofficialLinkpostBotMar 6, 2025, 10:38 PM
11 points
2 comments2 min readLW link
(www.anthropic.com)

We Have No Plan for Prevent­ing Loss of Con­trol in Open Models

Andrew DicksonMar 10, 2025, 3:35 PM
44 points
11 comments22 min readLW link

The In­tel­li­gence Curse

lukedragoJan 3, 2025, 7:07 PM
123 points
26 comments18 min readLW link
(lukedrago.substack.com)

Poli­cy­mak­ers don’t have ac­cess to pay­walled articles

Adam JonesJan 5, 2025, 10:56 AM
71 points
11 comments2 min readLW link
(adamjones.me)

New AI safety treaty pa­per out!

otto.bartenMar 26, 2025, 9:29 AM
15 points
2 comments4 min readLW link

Build­ing Big Science from the Bot­tom-Up: A Frac­tal Ap­proach to AI Safety

Lauren GreenspanJan 7, 2025, 3:08 AM
37 points
2 comments12 min readLW link

Gover­nance Course—Week 1 Reflections

Alice BlairJan 9, 2025, 4:48 AM
4 points
1 comment5 min readLW link

Thoughts on the In-Con­text Schem­ing AI Experiment

ExCephJan 9, 2025, 2:19 AM
3 points
0 comments4 min readLW link

Scal­ing AI Reg­u­la­tion: Real­is­ti­cally, what Can (and Can’t) Be Reg­u­lated?

Katalina HernandezMar 11, 2025, 4:51 PM
1 point
1 comment3 min readLW link
No comments.