Will vague “AI sen­tience” con­cerns do more for AI safety than any­thing else we might do?

Aryeh EnglanderJun 14, 2022, 11:53 PM
15 points
2 comments1 min readLW link

Yes, AI re­search will be sub­stan­tially cur­tailed if a lab causes a ma­jor disaster

lcJun 14, 2022, 10:17 PM
103 points
31 comments2 min readLW link

Slow mo­tion videos as AI risk in­tu­ition pumps

Andrew_CritchJun 14, 2022, 7:31 PM
241 points
41 comments2 min readLW link1 review

Cryp­to­graphic Life: How to tran­scend in a sub-light­speed world via Ho­mo­mor­phic encryption

GololJun 14, 2022, 7:22 PM
1 point
0 comments3 min readLW link

Blake Richards on Why he is Skep­ti­cal of Ex­is­ten­tial Risk from AI

Michaël TrazziJun 14, 2022, 7:09 PM
41 points
12 comments4 min readLW link
(theinsideview.ai)

[Question] How Do You Quan­tify [Physics In­ter­fac­ing] Real World Ca­pa­bil­ities?

DragonGodJun 14, 2022, 2:49 PM
17 points
1 comment4 min readLW link

Was the In­dus­trial Revolu­tion The In­dus­trial Revolu­tion?

Davis KedroskyJun 14, 2022, 2:48 PM
29 points
0 comments12 min readLW link
(daviskedrosky.substack.com)

In­ves­ti­gat­ing causal un­der­stand­ing in LLMs

Jun 14, 2022, 1:57 PM
28 points
6 comments13 min readLW link

Why multi-agent safety is im­por­tant

Akbir KhanJun 14, 2022, 9:23 AM
10 points
2 comments10 min readLW link

[Question] Was Eliezer Yud­kowsky right to give him­self 10% to suc­ceed with HPMoR in 2010?

momom2Jun 14, 2022, 7:00 AM
2 points
2 comments1 min readLW link

Re­sources I send to AI re­searchers about AI safety

Vael GatesJun 14, 2022, 2:24 AM
69 points
12 comments1 min readLW link

Vael Gates: Risks from Ad­vanced AI (June 2022)

Vael GatesJun 14, 2022, 12:54 AM
38 points
2 comments30 min readLW link

Cam­bridge LW Meetup: Per­sonal Finance

Tony WangJun 14, 2022, 12:12 AM
3 points
0 comments1 min readLW link

OpenAI: GPT-based LLMs show abil­ity to dis­crim­i­nate be­tween its own wrong an­swers, but in­abil­ity to ex­plain how/​why it makes that dis­crim­i­na­tion, even as model scales

Aditya JainJun 13, 2022, 11:33 PM
14 points
5 comments1 min readLW link
(openai.com)

[Question] Who said some­thing like “The fact that putting 2 ap­ples next to 2 other ap­ples leads to there be­ing 4 ap­ples there has noth­ing to do with the fact that 2 + 2 = 4”?

hunterglennJun 13, 2022, 10:23 PM
1 point
2 comments1 min readLW link

Con­ti­nu­ity Assumptions

Jan_KulveitJun 13, 2022, 9:31 PM
44 points
13 comments4 min readLW link

Crypto-fed Computation

aaguirreJun 13, 2022, 9:20 PM
24 points
7 comments7 min readLW link

A Modest Pivotal Act

anonymousaisafetyJun 13, 2022, 7:24 PM
−16 points
1 comment5 min readLW link

Con­tra EY: Can AGI de­stroy us with­out trial & er­ror?

nsokolskyJun 13, 2022, 6:26 PM
137 points
72 comments15 min readLW link

What are some smaller-but-con­crete challenges re­lated to AI safety that are im­pact­ing peo­ple to­day?

nonzerosumJun 13, 2022, 5:36 PM
4 points
3 comments1 min readLW link

[Link] New SEP ar­ti­cle on Bayesian Epistemology

Aryeh EnglanderJun 13, 2022, 3:03 PM
6 points
0 comments1 min readLW link

Train­ing Trace Priors

Adam JermynJun 13, 2022, 2:22 PM
12 points
17 comments4 min readLW link

[Question] Can you MRI a deep learn­ing model?

Yair HalberstadtJun 13, 2022, 1:43 PM
3 points
3 comments1 min readLW link

On A List of Lethalities

ZviJun 13, 2022, 12:30 PM
165 points
50 comments54 min readLW link1 review
(thezvi.wordpress.com)

D&D.Sci June 2022 Eval­u­a­tion and Ruleset

abstractapplicJun 13, 2022, 10:31 AM
34 points
11 comments4 min readLW link

[Question] What’s the “This AI is of moral con­cern.” fire alarm?

Quintin PopeJun 13, 2022, 8:05 AM
37 points
56 comments2 min readLW link

The beau­tiful mag­i­cal en­chanted golden Dall-e Mini is underrated

p.b.Jun 13, 2022, 7:58 AM
14 points
0 comments1 min readLW link

Why so lit­tle AI risk on ra­tio­nal­ist-ad­ja­cent blogs?

Grant DemareeJun 13, 2022, 6:31 AM
46 points
23 comments8 min readLW link

Code Qual­ity and Rule Consequentialism

Adam ZernerJun 13, 2022, 3:12 AM
17 points
13 comments6 min readLW link

Grokking “Semi-in­for­ma­tive pri­ors over AI timelines”

anson.hoJun 12, 2022, 10:17 PM
15 points
7 comments14 min readLW link

[Question] How much does cy­ber­se­cu­rity re­duce AI risk?

DarmaniJun 12, 2022, 10:13 PM
34 points
23 comments1 min readLW link

[Question] How are com­pute as­sets dis­tributed in the world?

Chris van MerwijkJun 12, 2022, 10:13 PM
30 points
7 comments1 min readLW link

In­tu­itive Ex­pla­na­tion of AIXI

Thomas LarsenJun 12, 2022, 9:41 PM
22 points
1 comment5 min readLW link

Why all the fuss about re­cur­sive self-im­prove­ment?

So8resJun 12, 2022, 8:53 PM
158 points
62 comments7 min readLW link1 review

Why the Kal­dor-Hicks crite­rion can be non-transitive

RupertJun 12, 2022, 5:26 PM
4 points
10 comments2 min readLW link

[Question] How do you post links here?

skybrianJun 12, 2022, 4:23 PM
1 point
1 comment1 min readLW link

[Question] Filter out tags from the front page?

jaspaxJun 12, 2022, 10:59 AM
9 points
2 comments1 min readLW link

How To: A Work­shop (or any­thing)

Duncan Sabien (Inactive)Jun 12, 2022, 8:00 AM
53 points
13 comments37 min readLW link1 review

A claim that Google’s LaMDA is sentient

Ben LivengoodJun 12, 2022, 4:18 AM
31 points
133 comments1 min readLW link

[Question] How much stupi­der than hu­mans can AI be and still kill us all through sheer num­bers and re­source ac­cess?

ShmiJun 12, 2022, 1:01 AM
11 points
11 comments1 min readLW link

ELK Pro­posal—Make the Re­porter care about the Pre­dic­tor’s beliefs

Jun 11, 2022, 10:53 PM
8 points
0 comments6 min readLW link

[Question] Why has no per­son /​ group ever taken over the world?

Aryeh EnglanderJun 11, 2022, 8:51 PM
25 points
19 comments1 min readLW link

[Question] Are there English-speak­ing mee­tups in Frank­furt/​Mu­nich/​Zurich?

Grant DemareeJun 11, 2022, 8:02 PM
6 points
2 comments1 min readLW link

Beauty and the Beast

Tomás B.11 Jun 2022 18:59 UTC
38 points
8 comments6 min readLW link

Poorly-Aimed Death Rays

Thane Ruthenis11 Jun 2022 18:29 UTC
48 points
5 comments4 min readLW link

AGI Safety Com­mu­ni­ca­tions Initiative

ines11 Jun 2022 17:34 UTC
7 points
0 comments1 min readLW link

A gam­ing group for ra­tio­nal­ity-aware people

dhatas11 Jun 2022 16:04 UTC
7 points
0 comments1 min readLW link

[Question] Why don’t you in­tro­duce re­ally im­pres­sive peo­ple you per­son­ally know to AI al­ign­ment (more of­ten)?

Verden11 Jun 2022 15:59 UTC
33 points
14 comments1 min readLW link

Godzilla Strategies

johnswentworth11 Jun 2022 15:44 UTC
159 points
72 comments3 min readLW link

Steganog­ra­phy and the Cy­cleGAN—al­ign­ment failure case study

Jan Czechowski11 Jun 2022 9:41 UTC
34 points
0 comments4 min readLW link