South Bay ACX/​LW Meetup

IS8 May 2023 23:55 UTC
2 points
0 comments1 min readLW link

H-JEPA might be tech­ni­cally al­ignable in a mod­ified form

Roman Leventov8 May 2023 23:04 UTC
12 points
2 comments7 min readLW link

All AGI Safety ques­tions wel­come (es­pe­cially ba­sic ones) [May 2023]

steven04618 May 2023 22:30 UTC
33 points
44 comments2 min readLW link

Pre­dictable up­dat­ing about AI risk

Joe Carlsmith8 May 2023 21:53 UTC
292 points
25 comments36 min readLW link1 review

An­no­tated re­ply to Ben­gio’s “AI Scien­tists: Safe and Use­ful AI?”

Roman Leventov8 May 2023 21:26 UTC
18 points
2 comments7 min readLW link
(yoshuabengio.org)

Are healthy choices effec­tive for im­prov­ing live ex­pec­tancy any­more?

Christopher King8 May 2023 21:25 UTC
6 points
4 comments1 min readLW link

LeCun’s “A Path Towards Au­tonomous Ma­chine In­tel­li­gence” has an un­solved tech­ni­cal al­ign­ment problem

Steven Byrnes8 May 2023 19:35 UTC
137 points
37 comments15 min readLW link

Product En­dorse­ment: Apollo Neuro

Elizabeth8 May 2023 19:00 UTC
46 points
28 comments5 min readLW link
(acesounderglass.com)

Acausal trade nat­u­rally re­sults in the Nash bar­gain­ing solution

Christopher King8 May 2023 18:13 UTC
3 points
0 comments4 min readLW link

In­fer­ence Speed is Not Unbounded

OneManyNone8 May 2023 16:24 UTC
35 points
32 comments16 min readLW link

[Cross­post] Un­veiling the Amer­i­can Public Opinion on AI Mo­ra­to­rium and Govern­ment In­ter­ven­tion: The Im­pact of Me­dia Exposure

otto.barten8 May 2023 14:09 UTC
7 points
0 comments6 min readLW link
(forum.effectivealtruism.org)

Thriv­ing in the Weird Times: Prepar­ing for the 100X Economy

8 May 2023 13:44 UTC
23 points
16 comments2 min readLW link

Hous­ing and Tran­sit Roundup #4

Zvi8 May 2023 13:30 UTC
25 points
0 comments11 min readLW link
(thezvi.wordpress.com)

Dance Profit Sharing

jefftk8 May 2023 13:10 UTC
11 points
3 comments2 min readLW link
(www.jefftk.com)

How “AGI” could end up be­ing many differ­ent spe­cial­ized AI’s stitched together

titotal8 May 2023 12:32 UTC
9 points
2 comments1 min readLW link

What does it take to ban a thing?

qbolec8 May 2023 11:00 UTC
66 points
18 comments5 min readLW link

Solomonoff’s solip­sism

Mergimio H. Doefevmil8 May 2023 6:55 UTC
−13 points
9 comments1 min readLW link

A tech­ni­cal note on bil­in­ear lay­ers for interpretability

Lee Sharkey8 May 2023 6:06 UTC
58 points
0 comments1 min readLW link
(arxiv.org)

[Question] Is EDT cor­rect? Does “EDT” == “log­i­cal EDT” == “log­i­cal CDT”?

Vivek Hebbar8 May 2023 2:07 UTC
13 points
2 comments1 min readLW link

LLM cog­ni­tion is prob­a­bly not hu­man-like

Max H8 May 2023 1:22 UTC
26 points
15 comments7 min readLW link

[Question] If al­ign­ment prob­lem was un­solv­able, would that avoid doom?

Kinrany7 May 2023 22:13 UTC
3 points
3 comments1 min readLW link

An ar­tifi­cially struc­tured ar­gu­ment for ex­pect­ing AGI ruin

Rob Bensinger7 May 2023 21:52 UTC
91 points
26 comments19 min readLW link

Where “the Se­quences” Are Wrong

Thoth Hermes7 May 2023 20:21 UTC
−15 points
5 comments14 min readLW link
(thothhermes.substack.com)

What’s wrong with be­ing dumb?

Adam Zerner7 May 2023 18:31 UTC
14 points
17 comments2 min readLW link

Cat­e­gories of Ar­gu­ing Style : Why be­ing good among ra­tio­nal­ists isn’t enough to ar­gue with everyone

Camille Berger 7 May 2023 17:45 UTC
16 points
0 comments23 min readLW link

Self-Ad­ministered Gell-Mann Amnesia

krs7 May 2023 17:44 UTC
1 point
1 comment1 min readLW link

Un­der­stand­ing mesa-op­ti­miza­tion us­ing toy models

7 May 2023 17:00 UTC
43 points
2 comments10 min readLW link

How to have Poly­geni­cally Screened Children

GeneSmith7 May 2023 16:01 UTC
361 points
128 comments27 min readLW link1 review

Statis­ti­cal mod­els & the ir­rele­vance of rare exceptions

patrissimo7 May 2023 15:59 UTC
37 points
6 comments2 min readLW link

Let’s look for co­her­ence theorems

Valdes7 May 2023 14:45 UTC
25 points
18 comments6 min readLW link

Graph­i­cal Rep­re­sen­ta­tions of Paul Chris­ti­ano’s Doom Model

Nathan Young7 May 2023 13:03 UTC
7 points
0 comments1 min readLW link

An an­thro­po­mor­phic AI dilemma

TsviBT7 May 2023 12:44 UTC
26 points
0 comments7 min readLW link

Violin Supports

jefftk7 May 2023 12:10 UTC
12 points
1 comment1 min readLW link
(www.jefftk.com)

Prop­er­ties of Good Textbooks

niplav7 May 2023 8:38 UTC
50 points
11 comments1 min readLW link

Against sac­ri­fic­ing AI trans­parency for gen­er­al­ity gains

Ape in the coat7 May 2023 6:52 UTC
4 points
0 comments2 min readLW link

TED talk by Eliezer Yud­kowsky: Un­leash­ing the Power of Ar­tifi­cial Intelligence

bayesed7 May 2023 5:45 UTC
49 points
36 comments1 min readLW link
(www.youtube.com)

Think­ing of Con­ve­nience as an Eco­nomic Term

ozziegooen7 May 2023 1:21 UTC
6 points
0 comments12 min readLW link
(forum.effectivealtruism.org)

Cor­rigi­bil­ity, Much more de­tail than any­one wants to Read

Logan Zoellner7 May 2023 1:02 UTC
26 points
2 comments7 min readLW link

Resi­d­ual stream norms grow ex­po­nen­tially over the for­ward pass

7 May 2023 0:46 UTC
76 points
24 comments11 min readLW link

On the Loeb­ner Silver Prize (a Tur­ing test)

hold_my_fish7 May 2023 0:39 UTC
18 points
2 comments2 min readLW link

Time and En­ergy Costs to Erase a Bit

DaemonicSigil6 May 2023 23:29 UTC
24 points
32 comments7 min readLW link

How much do you be­lieve your re­sults?

Eric Neyman6 May 2023 20:31 UTC
477 points
17 comments15 min readLW link3 reviews
(ericneyman.wordpress.com)

Long Covid Risks: 2023 Update

Elizabeth6 May 2023 18:20 UTC
70 points
9 comments4 min readLW link
(acesounderglass.com)

Is “red” for GPT-4 the same as “red” for you?

Yusuke Hayashi6 May 2023 17:55 UTC
9 points
6 comments2 min readLW link

The Broader Fos­sil Fuel Community

Jeffrey Heninger6 May 2023 14:49 UTC
16 points
1 comment3 min readLW link

Es­ti­mat­ing Norovirus Prevalence

jefftk6 May 2023 11:40 UTC
16 points
0 comments2 min readLW link
(www.jefftk.com)

Align­ment as Func­tion Fitting

A.H.6 May 2023 11:38 UTC
7 points
0 comments12 min readLW link

My preferred fram­ings for re­ward mis­speci­fi­ca­tion and goal misgeneralisation

Yi-Yang6 May 2023 4:48 UTC
27 points
1 comment8 min readLW link

You don’t need to be a ge­nius to be in AI safety research

Claire Short6 May 2023 2:32 UTC
14 points
1 comment6 min readLW link

Nat­u­ral­ist Collection

LoganStrohl6 May 2023 0:37 UTC
66 points
7 comments15 min readLW link