Danger, AI Scien­tist, Danger

Zvi15 Aug 2024 22:40 UTC
106 points
9 comments7 min readLW link
(thezvi.wordpress.com)

[Linkpost] ‘The AI Scien­tist: Towards Fully Au­to­mated Open-Ended Scien­tific Dis­cov­ery’

Bogdan Ionut Cirstea15 Aug 2024 21:32 UTC
20 points
1 comment1 min readLW link
(arxiv.org)

On power and its amplification

Ted Sanders15 Aug 2024 20:13 UTC
−1 points
0 comments1 min readLW link

The Over­looked Ne­ces­sity of Com­plete Se­man­tic Rep­re­sen­ta­tion in AI Safety and Alignment

williamsae15 Aug 2024 19:42 UTC
−1 points
0 comments3 min readLW link

My ar­ti­cle in The Na­tion — Cal­ifor­nia’s AI Safety Bill Is a Mask-Off Mo­ment for the Industry

garrison15 Aug 2024 19:25 UTC
35 points
0 comments1 min readLW link
(www.thenation.com)

Cri­tique of ‘Many Peo­ple Fear A.I. They Shouldn’t’ by David Brooks.

Axel Ahlqvist15 Aug 2024 18:38 UTC
12 points
8 comments3 min readLW link

Ex­po­sure can’t rule out disasters

Chipmonk15 Aug 2024 17:03 UTC
5 points
19 comments1 min readLW link
(chrislakin.blog)

Pri­mary Per­cep­tive Systems

ChristianKl15 Aug 2024 11:26 UTC
14 points
2 comments3 min readLW link

Se­quence overview: Welfare and moral weights

MichaelStJules15 Aug 2024 4:22 UTC
7 points
0 comments1 min readLW link

Fund­ing for pro­grams and events on global catas­trophic risk, effec­tive al­tru­ism, and other topics

14 Aug 2024 23:59 UTC
7 points
0 comments2 min readLW link

Fund­ing for work that builds ca­pac­ity to ad­dress risks from trans­for­ma­tive AI

14 Aug 2024 23:52 UTC
14 points
0 comments5 min readLW link

GPT-2 Some­times Fails at IOI

Ronak_Mehta14 Aug 2024 23:24 UTC
13 points
0 comments2 min readLW link
(ronakrm.github.io)

Toward a Hu­man Hy­brid Lan­guage for En­hanced Hu­man-Ma­chine Com­mu­ni­ca­tion: Ad­dress­ing the AI Align­ment Problem

Andndn Dheudnd14 Aug 2024 22:19 UTC
−6 points
2 comments4 min readLW link

Ad­verse Selec­tion by Life-Sav­ing Char­i­ties

vaishnav9214 Aug 2024 20:46 UTC
41 points
16 comments5 min readLW link
(www.everythingisatrolley.com)

The great Enigma in the sky: The uni­verse as an en­cryp­tion machine

Alex_Shleizer14 Aug 2024 13:21 UTC
4 points
1 comment8 min readLW link

An anti-in­duc­tive sequence

Viliam14 Aug 2024 12:28 UTC
36 points
10 comments3 min readLW link

Rabin’s Paradox

Charlie Steiner14 Aug 2024 5:40 UTC
18 points
40 comments3 min readLW link

An­nounc­ing the $200k EA Com­mu­nity Choice

Austin Chen14 Aug 2024 0:39 UTC
58 points
8 comments1 min readLW link
(manifund.substack.com)

De­bate: Is it eth­i­cal to work at AI ca­pa­bil­ities com­pa­nies?

14 Aug 2024 0:18 UTC
36 points
21 comments11 min readLW link

Fields that I refer­ence when think­ing about AI takeover prevention

Buck13 Aug 2024 23:08 UTC
143 points
16 comments10 min readLW link
(redwoodresearch.substack.com)

Ten counter-ar­gu­ments that AI is (not) an ex­is­ten­tial risk (for now)

Ariel Kwiatkowski13 Aug 2024 22:35 UTC
19 points
5 comments8 min readLW link

Align­ment from equivariance

hamishtodd113 Aug 2024 21:09 UTC
3 points
0 comments5 min readLW link

[LDSL#6] When is quan­tifi­ca­tion needed, and when is it hard?

tailcalled13 Aug 2024 20:39 UTC
31 points
0 comments2 min readLW link

A com­pu­ta­tional com­plex­ity ar­gu­ment for many worlds

jessicata13 Aug 2024 19:35 UTC
32 points
15 comments5 min readLW link
(unstableontology.com)

The Con­scious­ness Co­nun­drum: Why We Can’t Dis­miss Ma­chine Sentience

SystematicApproach13 Aug 2024 18:01 UTC
−21 points
1 comment3 min readLW link

Ten ar­gu­ments that AI is an ex­is­ten­tial risk

13 Aug 2024 17:00 UTC
110 points
41 comments7 min readLW link
(blog.aiimpacts.org)

Eu­gen­ics And Re­pro­duc­tion Li­censes FAQs: For the Com­mon Good

Zero Contradictions13 Aug 2024 16:34 UTC
−8 points
14 comments4 min readLW link
(zerocontradictions.net)

Su­per­in­tel­li­gent AI is pos­si­ble in the 2020s

HunterJay13 Aug 2024 6:03 UTC
41 points
3 comments12 min readLW link

De­bate: Get a col­lege de­gree?

12 Aug 2024 22:23 UTC
42 points
14 comments21 min readLW link

SYSTEMA ROBOTICA

Ali Ahmed12 Aug 2024 20:34 UTC
12 points
2 comments30 min readLW link

Ex­tract­ing SAE task fea­tures for in-con­text learning

12 Aug 2024 20:34 UTC
31 points
1 comment9 min readLW link

Hyppotherapy

Marius Adrian Nicoară12 Aug 2024 20:07 UTC
−3 points
0 comments1 min readLW link

Cal­ifor­ni­ans, tell your reps to vote yes on SB 1047!

Holly_Elmore12 Aug 2024 19:50 UTC
40 points
24 comments1 min readLW link

[LDSL#5] Com­par­i­son and mag­ni­tude/​diminishment

tailcalled12 Aug 2024 18:47 UTC
21 points
0 comments2 min readLW link

In Defense of Open-Minded UDT

abramdemski12 Aug 2024 18:27 UTC
72 points
27 comments11 min readLW link

Hu­man­ity isn’t re­motely longter­mist, so ar­gu­ments for AGI x-risk should fo­cus on the near term

Seth Herd12 Aug 2024 18:10 UTC
46 points
10 comments1 min readLW link

Creat­ing a “Con­science Calcu­la­tor” to Guard-Rail an AGI

sweenesm12 Aug 2024 16:03 UTC
−2 points
0 comments13 min readLW link

Shift­ing Headspaces—Tran­si­tional Beast-Mode

Jonathan Moregård12 Aug 2024 13:02 UTC
36 points
9 comments2 min readLW link
(honestliving.substack.com)

Si­mul­ta­neous Foot­bass and Foot­drums II

jefftk11 Aug 2024 23:50 UTC
9 points
0 comments1 min readLW link
(www.jefftk.com)

CultFrisbee

Gauraventh11 Aug 2024 21:36 UTC
16 points
3 comments1 min readLW link
(y1d2.com)

Plea­sure and suffer­ing are not con­cep­tual opposites

MichaelStJules11 Aug 2024 18:32 UTC
10 points
0 comments1 min readLW link

Com­pu­ta­tional ir­re­ducibil­ity challenges the simu­la­tion hypothesis

Clément L11 Aug 2024 16:14 UTC
3 points
15 comments7 min readLW link

[LDSL#4] Root cause anal­y­sis ver­sus effect size estimation

tailcalled11 Aug 2024 16:12 UTC
29 points
0 comments2 min readLW link

Closed to Interpretation

Yeshua God11 Aug 2024 15:51 UTC
−18 points
0 comments2 min readLW link

The­o­ries of Knowledge

Zero Contradictions11 Aug 2024 8:55 UTC
−1 points
5 comments1 min readLW link
(thewaywardaxolotl.blogspot.com)

Un­nat­u­ral abstractions

Aprillion10 Aug 2024 22:31 UTC
3 points
3 comments4 min readLW link
(peter.hozak.info)

[LDSL#3] In­for­ma­tion-ori­en­ta­tion is in ten­sion with mag­ni­tude-orientation

tailcalled10 Aug 2024 21:58 UTC
22 points
2 comments3 min readLW link

The AI reg­u­la­tor’s toolbox: A list of con­crete AI gov­er­nance practices

Adam Jones10 Aug 2024 21:15 UTC
7 points
1 comment34 min readLW link
(adamjones.me)

Diffu­sion Guided NLP: bet­ter steer­ing, mostly a good thing

Nathan Helm-Burger10 Aug 2024 19:49 UTC
13 points
0 comments1 min readLW link
(arxiv.org)

Tall tales and long odds

Solenoid_Entity10 Aug 2024 15:22 UTC
11 points
0 comments5 min readLW link