Recom­men­da­tion: re­ports on the search for miss­ing hiker Bill Ewasko

eukaryote31 Jul 2024 22:15 UTC
168 points
28 comments14 min readLW link
(eukaryotewritesblog.com)

Eco­nomics101 pre­dicted the failure of spe­cial card pay­ments for re­fugees, 3 months later whole of Ger­many wants to adopt it

Yanling Guo31 Jul 2024 21:09 UTC
2 points
1 comment2 min readLW link

Am­bi­guity in Pre­dic­tion Mar­ket Re­s­olu­tion is Still Harmful

aphyer31 Jul 2024 20:32 UTC
43 points
17 comments3 min readLW link

AI labs can boost ex­ter­nal safety research

Zach Stein-Perlman31 Jul 2024 19:30 UTC
31 points
1 comment1 min readLW link

Women in AI Safety Lon­don Meetup

njg31 Jul 2024 18:13 UTC
1 point
0 comments1 min readLW link

Con­struct­ing Neu­ral Net­work Pa­ram­e­ters with Down­stream Trainability

ch271828n31 Jul 2024 18:13 UTC
1 point
0 comments1 min readLW link
(github.com)

Want to work on US emerg­ing tech policy? Con­sider the Hori­zon Fel­low­ship.

Elika31 Jul 2024 18:12 UTC
4 points
0 comments1 min readLW link

[Question] What are your cruxes for im­pre­cise prob­a­bil­ities /​ de­ci­sion rules?

Anthony DiGiovanni31 Jul 2024 15:42 UTC
36 points
29 comments1 min readLW link

The new UK gov­ern­ment’s stance on AI safety

Elliot Mckernon31 Jul 2024 15:23 UTC
17 points
0 comments4 min readLW link

Solu­tions to prob­lems with Bayesianism

B Jacobs31 Jul 2024 14:18 UTC
6 points
0 comments21 min readLW link
(bobjacobs.substack.com)

Cat Sus­te­nance Fortification

jefftk31 Jul 2024 2:30 UTC
14 points
7 comments1 min readLW link
(www.jefftk.com)

Twit­ter thread on open-source AI

Richard_Ngo31 Jul 2024 0:26 UTC
33 points
6 comments2 min readLW link
(x.com)

Twit­ter thread on AI takeover scenarios

Richard_Ngo31 Jul 2024 0:24 UTC
37 points
0 comments2 min readLW link
(x.com)

Twit­ter thread on AI safety evals

Richard_Ngo31 Jul 2024 0:18 UTC
62 points
3 comments2 min readLW link
(x.com)

Twit­ter thread on poli­tics of AI safety

Richard_Ngo31 Jul 2024 0:00 UTC
35 points
2 comments1 min readLW link
(x.com)

An ML pa­per on data steal­ing pro­vides a con­struc­tion for “gra­di­ent hack­ing”

David Scott Krueger (formerly: capybaralet)30 Jul 2024 21:44 UTC
21 points
1 comment1 min readLW link
(arxiv.org)

Open Source Au­to­mated In­ter­pretabil­ity for Sparse Au­toen­coder Features

30 Jul 2024 21:11 UTC
67 points
1 comment13 min readLW link
(blog.eleuther.ai)

Cater­pillars and Philosophy

Zero Contradictions30 Jul 2024 20:54 UTC
2 points
0 comments1 min readLW link
(thewaywardaxolotl.blogspot.com)

François Chol­let on the limi­ta­tions of LLMs in reasoning

2PuNCheeZ30 Jul 2024 20:04 UTC
1 point
1 comment2 min readLW link
(x.com)

Against AI As An Ex­is­ten­tial Risk

Noah Birnbaum30 Jul 2024 19:10 UTC
6 points
13 comments1 min readLW link
(irrationalitycommunity.substack.com)

[Question] Is ob­jec­tive moral­ity self-defeat­ing?

dialectica30 Jul 2024 18:23 UTC
−4 points
3 comments2 min readLW link

Limi­ta­tions on the In­ter­pretabil­ity of Learned Fea­tures from Sparse Dic­tionary Learning

Tom Angsten30 Jul 2024 16:36 UTC
6 points
0 comments9 min readLW link

Self-Other Over­lap: A Ne­glected Ap­proach to AI Alignment

30 Jul 2024 16:22 UTC
192 points
43 comments12 min readLW link

In­ves­ti­gat­ing the Abil­ity of LLMs to Rec­og­nize Their Own Writing

30 Jul 2024 15:41 UTC
32 points
0 comments15 min readLW link

Can Gen­er­al­ized Ad­ver­sar­ial Test­ing En­able More Ri­gor­ous LLM Safety Evals?

scasper30 Jul 2024 14:57 UTC
25 points
0 comments4 min readLW link

RTFB: Cal­ifor­nia’s AB 3211

Zvi30 Jul 2024 13:10 UTC
62 points
2 comments11 min readLW link
(thezvi.wordpress.com)

If You Can Climb Up, You Can Climb Down

jefftk30 Jul 2024 0:00 UTC
34 points
9 comments1 min readLW link
(www.jefftk.com)

What is Mo­ral­ity?

Zero Contradictions29 Jul 2024 19:19 UTC
−1 points
0 comments1 min readLW link
(thewaywardaxolotl.blogspot.com)

Arch-an­ar­chism and im­mor­tal­ity

Peter lawless 29 Jul 2024 18:10 UTC
−5 points
1 comment2 min readLW link

AI Safety Newslet­ter #39: Im­pli­ca­tions of a Trump Ad­minis­tra­tion for AI Policy Plus, Safety Engineering

29 Jul 2024 17:50 UTC
17 points
1 comment6 min readLW link
(newsletter.safe.ai)

New Blog Post Against AI Doom

Noah Birnbaum29 Jul 2024 17:21 UTC
1 point
5 comments1 min readLW link
(substack.com)

An In­ter­pretabil­ity Illu­sion from Pop­u­la­tion Statis­tics in Causal Analysis

Daniel Tan29 Jul 2024 14:50 UTC
9 points
3 comments1 min readLW link

[Question] How to­k­eniza­tion in­fluences prompt­ing?

Boris Kashirin29 Jul 2024 10:28 UTC
9 points
4 comments1 min readLW link

Un­der­stand­ing Po­si­tional Fea­tures in Layer 0 SAEs

29 Jul 2024 9:36 UTC
43 points
0 comments5 min readLW link

Pre­dic­tion Mar­kets Explained

Benjamin_Sturisky29 Jul 2024 8:02 UTC
1 point
0 comments9 min readLW link

San Fran­cisco ACX Meetup “First Satur­day”

Nate Sternberg29 Jul 2024 6:11 UTC
3 points
2 comments1 min readLW link

Rel­a­tivity The­ory for What the Fu­ture ‘You’ Is and Isn’t

FlorianH29 Jul 2024 2:01 UTC
7 points
48 comments4 min readLW link

Wittgen­stein and Word2vec: Cap­tur­ing Re­la­tional Mean­ing in Lan­guage and Thought

cleanwhiteroom28 Jul 2024 19:55 UTC
2 points
2 comments2 min readLW link

Mak­ing Beliefs Pay Rent

28 Jul 2024 17:59 UTC
7 points
2 comments1 min readLW link

This is already your sec­ond chance

Malmesbury28 Jul 2024 17:13 UTC
174 points
13 comments8 min readLW link

[Question] Has Eliezer pub­li­cly and satis­fac­to­rily re­sponded to at­tempted re­but­tals of the anal­ogy to evolu­tion?

kaler28 Jul 2024 12:23 UTC
10 points
14 comments1 min readLW link

Fam­ily and Society

Zero Contradictions28 Jul 2024 7:05 UTC
1 point
0 comments1 min readLW link
(thewaywardaxolotl.blogspot.com)

[Question] What is AI Safety’s line of re­treat?

Remmelt28 Jul 2024 5:43 UTC
12 points
12 comments1 min readLW link

AXRP Epi­sode 34 - AI Eval­u­a­tions with Beth Barnes

DanielFilan28 Jul 2024 3:30 UTC
23 points
0 comments69 min readLW link

Rats, Back a Candidate

Blake28 Jul 2024 3:19 UTC
−40 points
19 comments1 min readLW link

AI ex­is­ten­tial risk prob­a­bil­ities are too un­re­li­able to in­form policy

Oleg Trott28 Jul 2024 0:59 UTC
18 points
5 comments1 min readLW link
(www.aisnakeoil.com)

Idle Spec­u­la­tions on Pipeline Parallelism

DaemonicSigil27 Jul 2024 22:40 UTC
1 point
0 comments4 min readLW link
(pbement.com)

Re: An­thropic’s sug­gested SB-1047 amendments

RobertM27 Jul 2024 22:32 UTC
87 points
13 comments9 min readLW link
(www.documentcloud.org)

The prob­lem with psy­chol­ogy is that it has no the­ory.

Nicholas D.27 Jul 2024 19:36 UTC
2 points
7 comments4 min readLW link
(nicholasdecker.substack.com)

Bryan John­son and a search for healthy longevity

NancyLebovitz27 Jul 2024 15:28 UTC
18 points
17 comments1 min readLW link