A trans­parency and in­ter­pretabil­ity tech tree

evhubJun 16, 2022, 11:44 PM
163 points
11 comments18 min readLW link1 review

BBC Fu­ture cov­ers progress studies

jasoncrawfordJun 16, 2022, 10:44 PM
21 points
6 comments3 min readLW link
(rootsofprogress.org)

Hu­mans are very re­li­able agents

alyssavanceJun 16, 2022, 10:02 PM
269 points
35 comments3 min readLW link

Towards Gears-Level Un­der­stand­ing of Agency

Thane RuthenisJun 16, 2022, 10:00 PM
25 points
4 comments18 min readLW link

A pos­si­ble AI-in­oc­u­la­tion due to early “robot up­ris­ing”

ShmiJun 16, 2022, 9:21 PM
16 points
2 comments1 min readLW link

AI Risk, as Seen on Snapchat

dkirmaniJun 16, 2022, 7:31 PM
23 points
8 comments1 min readLW link

[Link] “The mad­ness of re­duced med­i­cal di­ag­nos­tics” by Dynomight

KennyJun 16, 2022, 7:20 PM
16 points
25 comments1 min readLW link

Break­ing Down Goal-Directed Behaviour

Oliver SourbutJun 16, 2022, 6:45 PM
11 points
1 comment2 min readLW link

Per­ils of op­ti­miz­ing in so­cial contexts

owencbJun 16, 2022, 5:40 PM
50 points
1 comment2 min readLW link

Don’t Over-Op­ti­mize Things

owencbJun 16, 2022, 4:33 PM
27 points
6 comments4 min readLW link

[Question] Se­cu­rity anal­y­sis of ‘cloud chem­istry labs’?

KennyJun 16, 2022, 4:06 PM
6 points
2 comments1 min readLW link

Covid 6/​16/​22: Do Not Hand it to Them

ZviJun 16, 2022, 2:40 PM
29 points
5 comments7 min readLW link
(thezvi.wordpress.com)

[Question] Is there a worked ex­am­ple of Ge­or­gian taxes?

DagonJun 16, 2022, 2:07 PM
8 points
12 comments1 min readLW link

Against Ac­tive Shooter Drills

ZviJun 16, 2022, 1:40 PM
91 points
30 comments7 min readLW link
(thezvi.wordpress.com)

Ten ex­per­i­ments in mod­u­lar­ity, which we’d like you to run!

Jun 16, 2022, 9:17 AM
62 points
3 comments9 min readLW link

[Question] What if LaMDA is in­deed sen­tient /​ self-aware /​ worth hav­ing rights?

RomanSJun 16, 2022, 9:10 AM
22 points
13 comments1 min readLW link

Lifeguards

Orpheus16Jun 15, 2022, 11:03 PM
12 points
3 comments2 min readLW link
(forum.effectivealtruism.org)

Ra­tion­al­ity Vienna Hike

Laszlo_TreszkaiJun 15, 2022, 10:11 PM
3 points
0 comments1 min readLW link

Con­tra Hofs­tadter on GPT-3 Nonsense

ricticJun 15, 2022, 9:53 PM
237 points
24 comments2 min readLW link

Progress links and tweets, 2022-06-13

jasoncrawfordJun 15, 2022, 7:47 PM
12 points
0 comments1 min readLW link
(rootsofprogress.org)

I ap­plied for a MIRI job in 2020. Here’s what hap­pened next.

ViktoriaMalyasovaJun 15, 2022, 7:37 PM
86 points
17 comments7 min readLW link

Con­tex­tual Evil

ACrackedPotJun 15, 2022, 7:32 PM
1 point
12 comments2 min readLW link

Multi­gate Priors

Adam JermynJun 15, 2022, 7:30 PM
4 points
0 comments3 min readLW link

FYI: I’m work­ing on a book about the threat of AGI/​ASI for a gen­eral au­di­ence. I hope it will be of value to the cause and the community

Darren McKeeJun 15, 2022, 6:08 PM
43 points
15 comments2 min readLW link

[Question] What are all the AI Align­ment and AI Safety Com­mu­ni­ca­tion Hubs?

Gunnar_ZarnckeJun 15, 2022, 4:16 PM
27 points
5 comments1 min readLW link

Ge­or­gism, in theory

Stuart_ArmstrongJun 15, 2022, 3:20 PM
40 points
22 comments4 min readLW link

Ber­lin AI Safety Open Meetup June 2022

pranomostroJun 15, 2022, 2:33 PM
12 points
0 comments1 min readLW link

A cen­tral AI al­ign­ment prob­lem: ca­pa­bil­ities gen­er­al­iza­tion, and the sharp left turn

So8resJun 15, 2022, 1:10 PM
272 points
55 comments10 min readLW link1 review

Our men­tal build­ing blocks are more differ­ent than I thought

Marius HobbhahnJun 15, 2022, 11:07 AM
50 points
11 comments14 min readLW link

[Question] Has there been any work on at­tempt­ing to use Pas­cal’s Mug­ging to make an AGI be­have?

Chris_LeongJun 15, 2022, 8:33 AM
7 points
17 comments1 min readLW link

Align­ment Risk Doesn’t Re­quire Superintelligence

JustisMillsJun 15, 2022, 3:12 AM
35 points
4 comments2 min readLW link

A But­terfly’s View of Probability

Gabriel WuJun 15, 2022, 2:14 AM
29 points
17 comments11 min readLW link

[Question] Favourite new AI pro­duc­tivity tools?

Gabe MJun 15, 2022, 1:08 AM
14 points
5 comments1 min readLW link

Will vague “AI sen­tience” con­cerns do more for AI safety than any­thing else we might do?

Aryeh EnglanderJun 14, 2022, 11:53 PM
15 points
2 comments1 min readLW link

Yes, AI re­search will be sub­stan­tially cur­tailed if a lab causes a ma­jor disaster

lcJun 14, 2022, 10:17 PM
103 points
31 comments2 min readLW link

Slow mo­tion videos as AI risk in­tu­ition pumps

Andrew_CritchJun 14, 2022, 7:31 PM
241 points
41 comments2 min readLW link1 review

Cryp­to­graphic Life: How to tran­scend in a sub-light­speed world via Ho­mo­mor­phic encryption

GololJun 14, 2022, 7:22 PM
1 point
0 comments3 min readLW link

Blake Richards on Why he is Skep­ti­cal of Ex­is­ten­tial Risk from AI

Michaël TrazziJun 14, 2022, 7:09 PM
41 points
12 comments4 min readLW link
(theinsideview.ai)

[Question] How Do You Quan­tify [Physics In­ter­fac­ing] Real World Ca­pa­bil­ities?

DragonGodJun 14, 2022, 2:49 PM
17 points
1 comment4 min readLW link

Was the In­dus­trial Revolu­tion The In­dus­trial Revolu­tion?

Davis KedroskyJun 14, 2022, 2:48 PM
29 points
0 comments12 min readLW link
(daviskedrosky.substack.com)

In­ves­ti­gat­ing causal un­der­stand­ing in LLMs

Jun 14, 2022, 1:57 PM
28 points
6 comments13 min readLW link

Why multi-agent safety is im­por­tant

Akbir KhanJun 14, 2022, 9:23 AM
10 points
2 comments10 min readLW link

[Question] Was Eliezer Yud­kowsky right to give him­self 10% to suc­ceed with HPMoR in 2010?

momom2Jun 14, 2022, 7:00 AM
2 points
2 comments1 min readLW link

Re­sources I send to AI re­searchers about AI safety

Vael GatesJun 14, 2022, 2:24 AM
69 points
12 comments1 min readLW link

Vael Gates: Risks from Ad­vanced AI (June 2022)

Vael GatesJun 14, 2022, 12:54 AM
38 points
2 comments30 min readLW link

Cam­bridge LW Meetup: Per­sonal Finance

Tony WangJun 14, 2022, 12:12 AM
3 points
0 comments1 min readLW link

OpenAI: GPT-based LLMs show abil­ity to dis­crim­i­nate be­tween its own wrong an­swers, but in­abil­ity to ex­plain how/​why it makes that dis­crim­i­na­tion, even as model scales

Aditya JainJun 13, 2022, 11:33 PM
14 points
5 comments1 min readLW link
(openai.com)

[Question] Who said some­thing like “The fact that putting 2 ap­ples next to 2 other ap­ples leads to there be­ing 4 ap­ples there has noth­ing to do with the fact that 2 + 2 = 4”?

hunterglenn13 Jun 2022 22:23 UTC
1 point
2 comments1 min readLW link

Con­ti­nu­ity Assumptions

Jan_Kulveit13 Jun 2022 21:31 UTC
44 points
13 comments4 min readLW link

Crypto-fed Computation

aaguirre13 Jun 2022 21:20 UTC
24 points
7 comments7 min readLW link