[Question] Up­dates on FLI’s Value Alig­ment Map?

T431Sep 17, 2022, 10:27 PM
17 points
4 comments1 min readLW link

Most sen­si­ble ab­strac­tion & fea­ture set for a sys­tems lan­guage?

Jasen QinSep 17, 2022, 7:49 PM
0 points
5 comments10 min readLW link

Sparse tri­nary weighted RNNs as a path to bet­ter lan­guage model interpretability

Am8ryllisSep 17, 2022, 7:48 PM
19 points
13 comments3 min readLW link

Ap­ply for men­tor­ship in AI Safety field-building

Orpheus16Sep 17, 2022, 7:06 PM
9 points
0 comments1 min readLW link
(forum.effectivealtruism.org)

Refine’s Third Blog Post Day/​Week

adamShimiSep 17, 2022, 5:03 PM
18 points
0 comments1 min readLW link

[Closed] Prize and fast track to al­ign­ment re­search at ALTER

Vanessa KosoySep 17, 2022, 4:58 PM
63 points
8 comments3 min readLW link

Re­mote Lo­gin For Turnkey De­vices?

jefftkSep 17, 2022, 3:40 PM
9 points
2 comments2 min readLW link
(www.jefftk.com)

Many ther­apy schools work with in­ner mul­ti­plic­ity (not just IFS)

Sep 17, 2022, 10:27 AM
52 points
16 comments18 min readLW link

Should AI learn hu­man val­ues, hu­man norms or some­thing else?

Q HomeSep 17, 2022, 6:19 AM
5 points
1 comment4 min readLW link

Take­aways from our ro­bust in­jury clas­sifier pro­ject [Red­wood Re­search]

dmzSep 17, 2022, 3:55 AM
143 points
12 comments6 min readLW link1 review

[Question] Why doesn’t China (or didn’t any­one) en­courage/​man­date elas­tomeric res­pi­ra­tors to con­trol COVID?

Wei DaiSep 17, 2022, 3:07 AM
34 points
15 comments1 min readLW link

Emer­gency Res­i­den­tial So­lar Jury-Rigging

jefftkSep 17, 2022, 2:30 AM
34 points
0 comments3 min readLW link
(www.jefftk.com)

A Bite Sized In­tro­duc­tion to ELK

Luk27182Sep 17, 2022, 12:28 AM
5 points
0 comments6 min readLW link

D&D.Sci Septem­ber 2022: The Allo­ca­tion Helm

abstractapplicSep 16, 2022, 11:10 PM
34 points
34 comments1 min readLW link

Towards a philos­o­phy of safety

jasoncrawfordSep 16, 2022, 9:10 PM
12 points
2 comments8 min readLW link
(rootsofprogress.org)

Refine Blog­post Day #3: The short­forms I did write

Alexander Gietelink OldenzielSep 16, 2022, 9:03 PM
23 points
0 comments1 min readLW link

[Question] Why are we sure that AI will “want” some­thing?

ShmiSep 16, 2022, 8:35 PM
31 points
57 comments1 min readLW link

Katja Grace on Slow­ing Down AI, AI Ex­pert Sur­veys And Es­ti­mat­ing AI Risk

Michaël TrazziSep 16, 2022, 5:45 PM
40 points
2 comments3 min readLW link
(theinsideview.ai)

Levels of goals and alignment

zeshenSep 16, 2022, 4:44 PM
27 points
4 comments6 min readLW link

Rep­re­sen­ta­tional Tethers: Ty­ing AI La­tents To Hu­man Ones

Paul BricmanSep 16, 2022, 2:45 PM
30 points
0 comments16 min readLW link

I wrote a fan­tasy novel to pro­mote EA: More Chapters

Timothy UnderwoodSep 16, 2022, 9:47 AM
18 points
0 comments47 min readLW link

Guidelines for Mad Entrepreneurs

David UdellSep 16, 2022, 6:33 AM
31 points
0 comments11 min readLW link

Afford­able Hous­ing In­vest­ment Fund

jefftkSep 16, 2022, 2:30 AM
18 points
2 comments1 min readLW link
(www.jefftk.com)

In a world with­out AI, we need gene-edit­ing to pro­tect Na­ture. (Not how you think)

Erlja Jkdf.Sep 16, 2022, 1:24 AM
−11 points
2 comments1 min readLW link

As­tralCodexTen and Ra­tion­al­ity Meetup Or­ganisers’ Re­treat — Europe, Mid­dle East, and Africa 2023

Sam F. BrownSep 15, 2022, 10:38 PM
25 points
2 comments2 min readLW link
(www.rationalitymeetups.org)

A mar­ket is a neu­ral network

David Hugh-JonesSep 15, 2022, 9:53 PM
7 points
4 comments8 min readLW link

Un­der­stand­ing Con­jec­ture: Notes from Con­nor Leahy interview

Orpheus16Sep 15, 2022, 6:37 PM
107 points
23 comments15 min readLW link

How should Deep­Mind’s Chin­chilla re­vise our AI fore­casts?

Cleo NardoSep 15, 2022, 5:54 PM
35 points
12 comments13 min readLW link

Ra­tional An­i­ma­tions’ Script Writ­ing Contest

WriterSep 15, 2022, 4:56 PM
23 points
1 comment3 min readLW link

Covid 9/​15/​22: Per­ma­nent Normal

ZviSep 15, 2022, 4:00 PM
32 points
9 comments20 min readLW link
(thezvi.wordpress.com)

[Question] Are Hu­man Brains Univer­sal?

DragonGodSep 15, 2022, 3:15 PM
16 points
28 comments5 min readLW link

In­tel­li­gence failures and a the­ory of change for forecasting

NathanBarnardSep 15, 2022, 3:02 PM
5 points
0 comments10 min readLW link

Why de­cep­tive al­ign­ment mat­ters for AGI safety

Marius HobbhahnSep 15, 2022, 1:38 PM
68 points
13 comments13 min readLW link

FDT defects in a re­al­is­tic Twin Pri­son­ers’ Dilemma

SMKSep 15, 2022, 8:55 AM
38 points
1 comment26 min readLW link

[Question] What’s the longest a sen­tient ob­server could sur­vive in the Dark Era?

RaemonSep 15, 2022, 8:43 AM
33 points
15 comments1 min readLW link

The Value of Not Be­ing an Imposter

sudoSep 15, 2022, 8:32 AM
5 points
0 comments1 min readLW link

Ca­pa­bil­ity and Agency as Corner­stones of AI risk ­— My cur­rent model

wilmSep 15, 2022, 8:25 AM
10 points
4 comments12 min readLW link

Gen­eral ad­vice for tran­si­tion­ing into The­o­ret­i­cal AI Safety

Martín SotoSep 15, 2022, 5:23 AM
12 points
0 comments10 min readLW link

Se­quenc­ing In­tro II: Adapters

jefftkSep 15, 2022, 3:30 AM
12 points
0 comments2 min readLW link
(www.jefftk.com)

[Question] How do I find tu­tors for ob­scure skills/​sub­jects (i.e. fermi es­ti­ma­tion tu­tors)

joraineSep 15, 2022, 1:15 AM
11 points
2 comments1 min readLW link

[Question] Fore­cast­ing thread: How does AI risk level vary based on timelines?

eliflandSep 14, 2022, 11:56 PM
34 points
7 comments1 min readLW link

Co­or­di­nate-Free In­ter­pretabil­ity Theory

johnswentworthSep 14, 2022, 11:33 PM
52 points
16 comments5 min readLW link

Progress links and tweets, 2022-09-14

jasoncrawfordSep 14, 2022, 11:21 PM
9 points
2 comments1 min readLW link
(rootsofprogress.org)

Effec­tive al­tru­ism in the gar­den of ends

Tyler AltermanSep 14, 2022, 10:02 PM
24 points
1 comment27 min readLW link

The prob­lem with the me­dia pre­sen­ta­tion of “be­liev­ing in AI”

Roman LeventovSep 14, 2022, 9:05 PM
3 points
0 comments1 min readLW link

See­ing the Schema

vitaliyaSep 14, 2022, 8:45 PM
23 points
6 comments1 min readLW link

Re­spond­ing to ‘Beyond Hyper­an­thro­po­mor­phism’

ukc10014Sep 14, 2022, 8:37 PM
9 points
0 comments16 min readLW link

When is in­tent al­ign­ment suffi­cient or nec­es­sary to re­duce AGI con­flict?

Sep 14, 2022, 7:39 PM
40 points
0 comments9 min readLW link

When would AGIs en­gage in con­flict?

Sep 14, 2022, 7:38 PM
52 points
5 comments13 min readLW link

When does tech­ni­cal work to re­duce AGI con­flict make a differ­ence?: Introduction

Sep 14, 2022, 7:38 PM
52 points
3 comments6 min readLW link