[Question] Is there a good way to award a fixed prize in a pre­dic­tion con­test?

jchanNov 2, 2022, 9:37 PM
18 points
5 comments1 min readLW link

“Are Ex­per­i­ments Pos­si­ble?” Seeds of Science call for reviewers

rogersbaconNov 2, 2022, 8:05 PM
8 points
0 comments1 min readLW link

Hu­mans do acausal co­or­di­na­tion all the time

Adam JermynNov 2, 2022, 2:40 PM
57 points
35 comments3 min readLW link

Far-UVC Light Up­date: No, LEDs are not around the cor­ner (tweet­storm)

DavidmanheimNov 2, 2022, 12:57 PM
71 points
27 comments4 min readLW link
(twitter.com)

Hous­ing and Tran­sit Thoughts #1

ZviNov 2, 2022, 12:10 PM
35 points
5 comments16 min readLW link
(thezvi.wordpress.com)

Mind is uncountable

Filip SondejNov 2, 2022, 11:51 AM
18 points
22 comments1 min readLW link

AI Safety Needs Great Product Builders

goodgravyNov 2, 2022, 11:33 AM
14 points
2 comments1 min readLW link

Why is fiber good for you?

bracesNov 2, 2022, 2:04 AM
18 points
2 comments2 min readLW link

In­for­ma­tion Markets

eva_Nov 2, 2022, 1:24 AM
46 points
6 comments12 min readLW link

Se­quence Reread: Fake Beliefs [plus se­quence spotlight meta]

RaemonNov 2, 2022, 12:09 AM
27 points
3 comments1 min readLW link

Real-Time Re­search Record­ing: Can a Trans­former Re-Derive Po­si­tional Info?

Neel NandaNov 1, 2022, 11:56 PM
69 points
16 comments1 min readLW link
(youtu.be)

All AGI Safety ques­tions wel­come (es­pe­cially ba­sic ones) [~monthly thread]

Robert MilesNov 1, 2022, 11:23 PM
68 points
105 comments2 min readLW link

[Question] Which Is­sues in Con­cep­tual Align­ment have been For­mal­ised or Ob­served (or not)?

ojorgensenNov 1, 2022, 10:32 PM
4 points
0 comments1 min readLW link

AI as a Civ­i­liza­tional Risk Part 4/​6: Bioweapons and Philos­o­phy of Modification

PashaKamyshevNov 1, 2022, 8:50 PM
7 points
1 comment8 min readLW link

Open & Wel­come Thread—Novem­ber 2022

MondSemmelNov 1, 2022, 6:47 PM
14 points
46 comments1 min readLW link

Mildly Against Donor Lotteries

jefftkNov 1, 2022, 6:10 PM
10 points
9 comments3 min readLW link
(www.jefftk.com)

Progress links and tweets, 2022-11-01

jasoncrawfordNov 1, 2022, 5:48 PM
16 points
4 comments3 min readLW link
(rootsofprogress.org)

On the cor­re­spon­dence be­tween AI-mis­al­ign­ment and cog­ni­tive dis­so­nance us­ing a be­hav­ioral eco­nomics model

Stijn BruersNov 1, 2022, 5:39 PM
4 points
0 comments6 min readLW link

Threat Model Liter­a­ture Review

Nov 1, 2022, 11:03 AM
78 points
4 comments25 min readLW link

Clar­ify­ing AI X-risk

Nov 1, 2022, 11:03 AM
127 points
24 comments4 min readLW link1 review

Au­dit­ing games for high-level interpretability

Paul CologneseNov 1, 2022, 10:44 AM
33 points
1 comment7 min readLW link

Re­mem­ber to trans­late your thoughts back again

brookNov 1, 2022, 8:49 AM
25 points
11 comments3 min readLW link
(forum.effectivealtruism.org)

Con­ver­sa­tions on Al­co­hol Consumption

AnnapurnaNov 1, 2022, 5:09 AM
20 points
6 comments9 min readLW link

ML Safety Schol­ars Sum­mer 2022 Retrospective

TW123Nov 1, 2022, 3:09 AM
29 points
0 comments1 min readLW link

EA & LW Fo­rums Weekly Sum­mary (24 − 30th Oct 22′)

Zoe WilliamsNov 1, 2022, 2:58 AM
13 points
1 comment1 min readLW link

Cau­tion when in­ter­pret­ing Deep­mind’s In-con­text RL paper

Sam MarksNov 1, 2022, 2:42 AM
105 points
8 comments4 min readLW link

What sorts of sys­tems can be de­cep­tive?

Andrei AlexandruOct 31, 2022, 10:00 PM
16 points
0 comments7 min readLW link

“Cars and Elephants”: a hand­wavy ar­gu­ment/​anal­ogy against mechanis­tic interpretability

David Scott Krueger (formerly: capybaralet)Oct 31, 2022, 9:26 PM
48 points
25 comments2 min readLW link

Su­per­in­tel­li­gent AI is nec­es­sary for an amaz­ing fu­ture, but far from sufficient

So8resOct 31, 2022, 9:16 PM
132 points
48 comments34 min readLW link

San­ity-check­ing in an age of hyperbole

Ciprian Elliu IvanofOct 31, 2022, 8:04 PM
2 points
4 comments2 min readLW link

Why Aren’t There More Schel­ling Holi­days?

johnswentworthOct 31, 2022, 7:31 PM
63 points
21 comments1 min readLW link

The cir­cu­lar prob­lem of epistemic irresponsibility

Roman LeventovOct 31, 2022, 5:23 PM
5 points
2 comments8 min readLW link

AI as a Civ­i­liza­tional Risk Part 3/​6: Anti-econ­omy and Sig­nal Pollution

PashaKamyshevOct 31, 2022, 5:03 PM
7 points
4 comments14 min readLW link

Aver­age util­i­tar­i­anism is non-local

Yair HalberstadtOct 31, 2022, 4:36 PM
29 points
13 comments1 min readLW link

Marvel Snap: Phase 1

ZviOct 31, 2022, 3:20 PM
23 points
1 comment14 min readLW link
(thezvi.wordpress.com)

Boundaries vs Frames

Scott GarrabrantOct 31, 2022, 3:14 PM
58 points
10 comments7 min readLW link

Embed­ding safety in ML development

zeshenOct 31, 2022, 12:27 PM
24 points
1 comment18 min readLW link

[Book] In­ter­pretable Ma­chine Learn­ing: A Guide for Mak­ing Black Box Models Explainable

Esben KranOct 31, 2022, 11:38 AM
20 points
1 comment1 min readLW link
(christophm.github.io)

My (naive) take on Risks from Learned Optimization

Artyom KarpovOct 31, 2022, 10:59 AM
7 points
0 comments5 min readLW link

Tac­ti­cal Nu­clear Weapons Aren’t Cost-Effec­tive Com­pared to Pre­ci­sion Artillery

Lao MeinOct 31, 2022, 4:33 AM
28 points
7 comments3 min readLW link

Gan­dalf or Saru­man? A Soldier in Scout’s Clothing

DirectedEvolutionOct 31, 2022, 2:40 AM
41 points
1 comment4 min readLW link

Me (Steve Byrnes) on the “Brain In­spired” podcast

Steven ByrnesOct 30, 2022, 7:15 PM
26 points
1 comment1 min readLW link
(braininspired.co)

“Nor­mal” is the equil­ibrium state of past op­ti­miza­tion processes

Alex_AltairOct 30, 2022, 7:03 PM
81 points
5 comments5 min readLW link

AI as a Civ­i­liza­tional Risk Part 2/​6: Be­hav­ioral Modification

PashaKamyshevOct 30, 2022, 4:57 PM
9 points
0 comments10 min readLW link

In­stru­men­tal ig­nor­ing AI, Dumb but not use­less.

Donald HobsonOct 30, 2022, 4:55 PM
7 points
6 comments2 min readLW link

Weekly Roundup #3

ZviOct 30, 2022, 12:20 PM
23 points
5 comments15 min readLW link
(thezvi.wordpress.com)

Quickly re­fac­tor­ing the U.S. Constitution

lcOct 30, 2022, 7:17 AM
7 points
25 comments4 min readLW link

«Boundaries», Part 3a: Defin­ing bound­aries as di­rected Markov blankets

Andrew_CritchOct 30, 2022, 6:31 AM
90 points
20 comments15 min readLW link

Am I se­cretly ex­cited for AI get­ting weird?

porbyOct 29, 2022, 10:16 PM
116 points
4 comments4 min readLW link

AI as a Civ­i­liza­tional Risk Part 1/​6: His­tor­i­cal Priors

PashaKamyshevOct 29, 2022, 9:59 PM
2 points
2 comments7 min readLW link