Law-Fol­low­ing AI 4: Don’t Rely on Vi­car­i­ous Liability

CullenAug 2, 2022, 11:26 PM
5 points
2 comments3 min readLW link

Two-year up­date on my per­sonal AI timelines

Ajeya CotraAug 2, 2022, 11:07 PM
293 points
60 comments16 min readLW link

What are the Red Flags for Neu­ral Net­work Suffer­ing? - Seeds of Science call for reviewers

rogersbaconAug 2, 2022, 10:37 PM
24 points
6 comments1 min readLW link

Againstness

CFAR!DuncanAug 2, 2022, 7:29 PM
50 points
8 comments9 min readLW link

(Sum­mary) Se­quence High­lights—Think­ing Bet­ter on Purpose

qazzquimbyAug 2, 2022, 5:45 PM
33 points
3 comments11 min readLW link

Progress links and tweets, 2022-08-02

jasoncrawfordAug 2, 2022, 5:03 PM
9 points
0 comments1 min readLW link
(rootsofprogress.org)

[Question] I want to donate some money (not much, just what I can af­ford) to AGI Align­ment re­search, to what­ever or­ga­ni­za­tion has the best chance of mak­ing sure that AGI goes well and doesn’t kill us all. What are my best op­tions, where can I make the most differ­ence per dol­lar?

lumenwritesAug 2, 2022, 12:08 PM
15 points
9 comments1 min readLW link

Think­ing with­out pri­ors?

Q HomeAug 2, 2022, 9:17 AM
7 points
0 comments9 min readLW link

[Question] Would quan­tum im­mor­tal­ity mean sub­jec­tive im­mor­tal­ity?

n0ahAug 2, 2022, 4:54 AM
2 points
10 comments1 min readLW link

Turbocharging

CFAR!DuncanAug 2, 2022, 12:01 AM
52 points
4 comments9 min readLW link

Let­ter from lead­ing Soviet Aca­demi­ci­ans to party and gov­ern­ment lead­ers of the Soviet Union re­gard­ing signs of de­cline and struc­tural prob­lems of the eco­nomic-poli­ti­cal sys­tem (1970)

M. Y. ZuoAug 1, 2022, 10:35 PM
20 points
10 comments16 min readLW link

Tech­ni­cal AI Align­ment Study Group

Eric KAug 1, 2022, 6:33 PM
5 points
0 comments1 min readLW link

[Question] Is there any writ­ing about prompt en­g­ineer­ing for hu­mans?

Alex HollowAug 1, 2022, 12:52 PM
18 points
8 comments1 min readLW link

Med­i­ta­tion course claims 65% en­light­en­ment rate: my review

KatWoodsAug 1, 2022, 11:25 AM
111 points
35 comments14 min readLW link

[Question] Which in­tro-to-AI-risk text would you recom­mend to...

SherrinfordAug 1, 2022, 9:36 AM
12 points
1 comment1 min readLW link

Po­laris, Five-Se­cond Ver­sions, and Thought Lengths

CFAR!DuncanAug 1, 2022, 7:14 AM
50 points
12 comments8 min readLW link

A Word is Worth 1,000 Pictures

KullyAug 1, 2022, 4:08 AM
1 point
0 comments2 min readLW link

On akra­sia: start­ing at the bottom

seecrowAug 1, 2022, 4:08 AM
37 points
2 comments3 min readLW link

[Question] How likely do you think worse-than-ex­tinc­tion type fates to be?

span1Aug 1, 2022, 4:08 AM
3 points
3 comments1 min readLW link

Ab­strac­tion sac­ri­fices causal clarity

Marv KJul 31, 2022, 7:24 PM
2 points
0 comments3 min readLW link

Time-log­ging pro­grams and/​or spread­sheets (2022)

mikbpJul 31, 2022, 6:18 PM
3 points
3 comments1 min readLW link

Con­ser­vatism is a ra­tio­nal re­sponse to epistemic uncertainty

contrarianbritJul 31, 2022, 6:04 PM
2 points
11 comments9 min readLW link
(thomasprosser.substack.com)

South Bay ACX/​LW Meetup

ISJul 31, 2022, 3:30 PM
2 points
0 comments1 min readLW link

Per­verse In­de­pen­dence Incentives

jefftkJul 31, 2022, 2:40 PM
61 points
3 comments1 min readLW link
(www.jefftk.com)

Wolfram Re­search v Cook

KennyJul 31, 2022, 1:35 PM
7 points
3 comments8 min readLW link

Wanted: No­ta­tion for credal resilience

PeterHJul 31, 2022, 7:35 AM
21 points
12 comments1 min readLW link

Anatomy of a Dat­ing Document

squidiousJul 31, 2022, 2:40 AM
29 points
24 comments4 min readLW link
(opalsandbonobos.blogspot.com)

chin­chilla’s wild implications

nostalgebraistJul 31, 2022, 1:18 AM
424 points
128 comments10 min readLW link1 review

AGI-level rea­soner will ap­pear sooner than an agent; what the hu­man­ity will do with this rea­soner is critical

Roman LeventovJul 30, 2022, 8:56 PM
24 points
10 comments1 min readLW link

[Question] What job should I do?

Tom PaineJul 30, 2022, 9:15 AM
2 points
8 comments1 min readLW link

How trans­parency changed over time

ViktoriaMalyasovaJul 30, 2022, 4:36 AM
21 points
0 comments6 min readLW link

Trans­lat­ing be­tween La­tent Spaces

Jul 30, 2022, 3:25 AM
27 points
2 comments8 min readLW link

Drexler’s Nan­otech Forecast

PeterMcCluskeyJul 30, 2022, 12:45 AM
25 points
28 comments3 min readLW link
(www.bayesianinvestor.com)

Hu­mans Reflect­ing on HRH

leogaoJul 29, 2022, 9:56 PM
26 points
4 comments2 min readLW link

Com­par­ing Four Ap­proaches to In­ner Alignment

Lucas TeixeiraJul 29, 2022, 9:06 PM
38 points
1 comment9 min readLW link

Ques­tions for a The­ory of Narratives

Marv KJul 29, 2022, 7:31 PM
5 points
4 comments4 min readLW link

Focusing

CFAR!DuncanJul 29, 2022, 7:15 PM
114 points
23 comments14 min readLW link

Con­jec­ture: In­ter­nal In­fo­haz­ard Policy

Jul 29, 2022, 7:07 PM
131 points
6 comments19 min readLW link

Ab­stract­ing The Hard­ness of Align­ment: Un­bounded Atomic Optimization

adamShimiJul 29, 2022, 6:59 PM
72 points
3 comments16 min readLW link

Bucket Errors

CFAR!DuncanJul 29, 2022, 6:50 PM
43 points
7 comments11 min readLW link

Distil­la­tion Con­test—Re­sults and Recap

ArisJul 29, 2022, 5:40 PM
34 points
0 comments7 min readLW link

The gen­er­al­ized Sier­pin­ski-Mazurk­iewicz the­o­rem.

Donald HobsonJul 29, 2022, 12:12 AM
11 points
4 comments1 min readLW link

The Con­ver­sa­tions We Make Space For

Severin T. SeehrichJul 28, 2022, 9:37 PM
21 points
0 comments3 min readLW link

An­nounc­ing the AI Safety Field Build­ing Hub, a new effort to provide AISFB pro­jects, men­tor­ship, and funding

Vael GatesJul 28, 2022, 9:29 PM
49 points
3 comments6 min readLW link

Defin­ing Op­ti­miza­tion in a Deeper Way Part 4

J BostockJul 28, 2022, 5:02 PM
7 points
0 comments5 min readLW link

Covid 7/​28/​22: Ruin­ing It For Everyone

ZviJul 28, 2022, 3:10 PM
32 points
8 comments12 min readLW link
(thezvi.wordpress.com)

Mon­key­pox Post #2

ZviJul 28, 2022, 1:20 PM
36 points
3 comments6 min readLW link
(thezvi.wordpress.com)

For Bet­ter Com­ment­ing, Stop Out Loud

DirectedEvolutionJul 28, 2022, 1:39 AM
18 points
30 comments1 min readLW link

Seek­ing beta read­ers who are ig­no­rant of biol­ogy but knowl­edge­able about AI safety

Holly_ElmoreJul 27, 2022, 11:02 PM
11 points
6 comments1 min readLW link

Prin­ci­ples of Pri­vacy for Align­ment Research

johnswentworthJul 27, 2022, 7:53 PM
73 points
31 comments7 min readLW link