The Prob­lem with Rea­son­ers by Ai­dan McLaughin

t14n25 Nov 2024 20:24 UTC
7 points
1 comment1 min readLW link
(aidanmclaughlin.notion.site)

Lo­cally op­ti­mal psychology

Chipmonk25 Nov 2024 18:35 UTC
37 points
7 comments2 min readLW link
(twitter.com)

a space habitat design

bhauth25 Nov 2024 17:28 UTC
53 points
13 comments9 min readLW link
(bhauth.com)

Arthro­pod (non) sentience

Arturo Macias25 Nov 2024 16:01 UTC
9 points
8 comments4 min readLW link

Cross­post: Devel­op­ing the mid­dle ground on po­larized topics

juliawise25 Nov 2024 14:39 UTC
13 points
16 comments3 min readLW link

Two fla­vors of com­pu­ta­tional functionalism

EuanMcLean25 Nov 2024 10:47 UTC
28 points
9 comments4 min readLW link

Align­ment is not intelligent

Donatas Lučiūnas25 Nov 2024 6:59 UTC
−17 points
18 comments5 min readLW link

Zaragoza ACX/​LW Meetup

Fernand025 Nov 2024 6:56 UTC
1 point
0 comments1 min readLW link

A bet­ter “State­ment on AI Risk?”

Knight Lee25 Nov 2024 4:50 UTC
4 points
4 comments3 min readLW link

Reflec­tions on ML4Good

james__p25 Nov 2024 2:40 UTC
12 points
0 comments1 min readLW link

AI Spe­cial­ized in ML Train­ing Could Create ASI: AGI Is Unnecessary

satopi25 Nov 2024 2:31 UTC
−5 points
1 comment1 min readLW link

I, Token

Ivan Vendrov25 Nov 2024 2:20 UTC
14 points
2 comments3 min readLW link
(nothinghuman.substack.com)

Pas­sages I High­lighted in The Let­ters of J.R.R.Tolkien

Ivan Vendrov25 Nov 2024 1:47 UTC
116 points
10 comments31 min readLW link

Dec­o­rated pedes­trian tunnels

dkl924 Nov 2024 22:16 UTC
0 points
3 comments1 min readLW link
(dkl9.net)

Gothen­burg LW/​ACX meetup

Stefan24 Nov 2024 19:40 UTC
2 points
0 comments1 min readLW link

[Question] Are You More Real If You’re Really For­get­ful?

Thane Ruthenis24 Nov 2024 19:30 UTC
39 points
25 comments5 min readLW link

Per­ils of Gen­er­al­iz­ing from One’s So­cial Group

localdeity24 Nov 2024 15:31 UTC
64 points
1 comment3 min readLW link

Disen­tan­gling Rep­re­sen­ta­tions through Multi-task Learning

Bogdan Ionut Cirstea24 Nov 2024 13:10 UTC
14 points
1 comment1 min readLW link
(arxiv.org)

The U.S. Na­tional Se­cu­rity State is Here to Make AI Even Less Trans­par­ent and Accountable

Matrice Jacobine24 Nov 2024 9:36 UTC
0 points
0 comments2 min readLW link
(www.eff.org)

Mechanis­tic In­ter­pretabil­ity of Llama 3.2 with Sparse Autoencoders

PaulPauls24 Nov 2024 5:45 UTC
20 points
3 comments1 min readLW link
(github.com)

SB-1047, ChatGPT and AI’s Game of Thrones

Rahul Chand24 Nov 2024 2:29 UTC
−3 points
1 comment13 min readLW link

Beyond Gaus­sian: Lan­guage Model Rep­re­sen­ta­tions and Distributions

Matt Levinson24 Nov 2024 1:53 UTC
5 points
1 comment5 min readLW link

How Univer­sal Ba­sic In­come Could Help Us Build a Brighter Future

Yanling Guo23 Nov 2024 22:03 UTC
−13 points
13 comments3 min readLW link

Com­pute and size limits on AI are the ac­tual danger

Shmi23 Nov 2024 21:29 UTC
31 points
5 comments2 min readLW link

Paradigm Shifts—change ev­ery­thing… ex­cept al­most everything

James Stephen Brown23 Nov 2024 18:34 UTC
1 point
0 comments3 min readLW link
(nonzerosum.games)

A Sober Look at Steer­ing Vec­tors for LLMs

23 Nov 2024 17:30 UTC
31 points
0 comments5 min readLW link

Text Posts from the Kids Group: 2018

jefftk23 Nov 2024 12:50 UTC
20 points
0 comments24 min readLW link
(www.jefftk.com)

Re­ward Bases: A sim­ple mechanism for adap­tive ac­qui­si­tion of mul­ti­ple re­ward type

Bogdan Ionut Cirstea23 Nov 2024 12:45 UTC
11 points
0 comments1 min readLW link

On The Ra­tion­al­ist Megameetup

Screwtape23 Nov 2024 9:08 UTC
25 points
3 comments10 min readLW link

[Question] Have we seen any “ReLU in­stead of sig­moid-type im­prove­ments” recently

KvmanThinking23 Nov 2024 3:51 UTC
2 points
4 comments1 min readLW link

A few ques­tions about re­cent de­vel­op­ments in EA

Peter Berggren23 Nov 2024 2:36 UTC
24 points
12 comments2 min readLW link

Parad­dic­tions: un­rea­son­ably com­pel­ling be­hav­iors and their uses

Michael Cohn22 Nov 2024 20:53 UTC
13 points
0 comments6 min readLW link

Liter­acy Rates Haven’t Fallen By 20% Since the Depart­ment of Ed­u­ca­tion Was Created

Maxwell Tabarrok22 Nov 2024 20:53 UTC
44 points
0 comments3 min readLW link
(www.maximum-progress.com)

Plau­si­bly Fac­tor­ing Con­jec­tures

22 Nov 2024 20:11 UTC
22 points
1 comment10 min readLW link

Op­ti­miz­ing Prob­lem-Solv­ing Strate­gies Through Pre­dic­tion Markets

patrik-cihal22 Nov 2024 19:58 UTC
1 point
0 comments2 min readLW link

Do­ing Re­search Part-Time is Great

casualphysicsenjoyer22 Nov 2024 19:01 UTC
37 points
7 comments5 min readLW link

Re­think­ing Laplace’s Rule of Succession

Cleo Nardo22 Nov 2024 18:46 UTC
9 points
5 comments2 min readLW link

(Salt) Water Gar­gling as an Antiviral

Elizabeth22 Nov 2024 18:00 UTC
88 points
6 comments5 min readLW link
(acesounderglass.com)

The Man­u­fac­tured Cri­sis: How So­ciety Is Willingly Ty­ing Its Own Noose

PROPHET22 Nov 2024 17:45 UTC
−2 points
2 comments8 min readLW link

Side­load­ing: cre­at­ing a model of a per­son via LLM with very large prompt

22 Nov 2024 16:41 UTC
12 points
4 comments35 min readLW link

Neu­ro­science of hu­man so­cial in­stincts: a sketch

Steven Byrnes22 Nov 2024 16:16 UTC
55 points
0 comments31 min readLW link

Re­but­ting Every Ob­jec­tion To Giv­ing To The Shrimp Welfare Pro­ject

omnizoid22 Nov 2024 16:12 UTC
−2 points
0 comments8 min readLW link

A very strange prob­a­bil­ity paradox

notfnofn22 Nov 2024 14:01 UTC
90 points
26 comments9 min readLW link

The boat

RomanS22 Nov 2024 12:56 UTC
3 points
0 comments2 min readLW link

[Question] Which things were you sur­prised to learn are metaphors?

Gordon Seidoh Worley22 Nov 2024 3:46 UTC
28 points
18 comments1 min readLW link

LLM chat­bots have ~half of the kinds of “con­scious­ness” that hu­mans be­lieve in. Hu­mans should avoid go­ing crazy about that.

Andrew_Critch22 Nov 2024 3:26 UTC
77 points
53 comments5 min readLW link

Read­ing RFK Jr so that you don’t have to

braces22 Nov 2024 0:59 UTC
56 points
1 comment8 min readLW link

Don’t want Good­hart? — Spec­ify the damn variables

Yan Lyutnev21 Nov 2024 22:45 UTC
−3 points
2 comments5 min readLW link

Don’t want Good­hart? — Spec­ify the vari­ables more

YanLyutnev21 Nov 2024 22:43 UTC
3 points
2 comments5 min readLW link

Align­ing AI Safety Pro­jects with a Repub­li­can Administration

Deric Cheng21 Nov 2024 22:12 UTC
29 points
1 comment8 min readLW link