Ra­tion­al­ism and so­cial rationalism

philosophybear10 Mar 2023 23:20 UTC
17 points
5 comments10 min readLW link
(philosophybear.substack.com)

Meetup Tip: Nametags

Screwtape10 Mar 2023 21:00 UTC
16 points
2 comments3 min readLW link

[Question] Is ChatGPT (or other LLMs) more ‘sen­tient’/​’con­scious/​etc. then a baby with­out a brain?

M. Y. Zuo10 Mar 2023 19:00 UTC
−4 points
2 comments1 min readLW link

The hu­man­ity’s biggest mistake

RomanS10 Mar 2023 16:30 UTC
0 points
1 comment2 min readLW link

Oper­a­tional­iz­ing timelines

Zach Stein-Perlman10 Mar 2023 16:30 UTC
7 points
1 comment3 min readLW link

[Question] What do you think is wrong with ra­tio­nal­ist cul­ture?

tailcalled10 Mar 2023 13:17 UTC
16 points
77 comments1 min readLW link

Dice De­ci­sion Making

Bart Bussmann10 Mar 2023 13:01 UTC
20 points
14 comments3 min readLW link

Stop call­ing it “jailbreak­ing” ChatGPT

Templarrr10 Mar 2023 11:41 UTC
7 points
9 comments2 min readLW link

Long-term mem­ory for LLM via self-repli­cat­ing prompt

avturchin10 Mar 2023 10:28 UTC
20 points
3 comments2 min readLW link

Thoughts on the OpenAI al­ign­ment plan: will AI re­search as­sis­tants be net-pos­i­tive for AI ex­is­ten­tial risk?

Jeffrey Ladish10 Mar 2023 8:21 UTC
58 points
3 comments9 min readLW link

Reflec­tions On The Fea­si­bil­ity Of Scal­able-Oversight

Felix Hofstätter10 Mar 2023 7:54 UTC
11 points
0 comments12 min readLW link

Ja­pan AI Align­ment Conference

10 Mar 2023 6:56 UTC
64 points
7 comments1 min readLW link
(www.conjecture.dev)

Every­thing’s nor­mal un­til it’s not

Eleni Angelou10 Mar 2023 2:02 UTC
7 points
0 comments3 min readLW link

Acolytes, re­form­ers, and atheists

lc10 Mar 2023 0:48 UTC
9 points
0 comments4 min readLW link

The hot mess the­ory of AI mis­al­ign­ment: More in­tel­li­gent agents be­have less coherently

Jonathan Yan10 Mar 2023 0:20 UTC
47 points
21 comments1 min readLW link
(sohl-dickstein.github.io)

Why Not Just Out­source Align­ment Re­search To An AI?

johnswentworth9 Mar 2023 21:49 UTC
139 points
49 comments9 min readLW link

What’s Not Our Problem

Jacob Falkovich9 Mar 2023 20:07 UTC
22 points
6 comments9 min readLW link

Ques­tions about Con­je­cure’s CoEm proposal

9 Mar 2023 19:32 UTC
51 points
4 comments2 min readLW link

What Ja­son has been read­ing, March 2023

jasoncrawford9 Mar 2023 18:46 UTC
12 points
0 comments6 min readLW link
(rootsofprogress.org)

[Question] “Provide C++ code for a func­tion that out­puts a Fibonacci se­quence of n terms, where n is pro­vided as a pa­ram­e­ter to the function

Thembeka999 Mar 2023 18:37 UTC
−21 points
2 comments1 min readLW link

An­thropic: Core Views on AI Safety: When, Why, What, and How

jonmenaster9 Mar 2023 17:34 UTC
17 points
1 comment22 min readLW link
(www.anthropic.com)

Why do we as­sume there is a “real” shog­goth be­hind the LLM? Why not masks all the way down?

Robert_AIZI9 Mar 2023 17:28 UTC
63 points
48 comments2 min readLW link

An­thropic’s Core Views on AI Safety

Zac Hatfield-Dodds9 Mar 2023 16:55 UTC
172 points
39 comments2 min readLW link
(www.anthropic.com)

Some ML-Re­lated Math I Now Un­der­stand Better

Fabien Roger9 Mar 2023 16:35 UTC
45 points
4 comments4 min readLW link

The Translu­cent Thoughts Hy­pothe­ses and Their Implications

Fabien Roger9 Mar 2023 16:30 UTC
142 points
7 comments19 min readLW link

IRL in Gen­eral Environments

michaelcohen9 Mar 2023 13:32 UTC
8 points
20 comments1 min readLW link

Utility un­cer­tainty vs. ex­pected in­for­ma­tion gain

michaelcohen9 Mar 2023 13:32 UTC
13 points
9 comments1 min readLW link

Value Learn­ing is only Asymp­tot­i­cally Safe

michaelcohen9 Mar 2023 13:32 UTC
5 points
19 comments1 min readLW link

Im­pact Mea­sure Test­ing with Honey Pots and Myopia

michaelcohen9 Mar 2023 13:32 UTC
13 points
9 comments1 min readLW link

Just Imi­tate Hu­mans?

michaelcohen9 Mar 2023 13:31 UTC
11 points
72 comments1 min readLW link

Build a Causal De­ci­sion Theorist

michaelcohen9 Mar 2023 13:31 UTC
−2 points
14 comments4 min readLW link

ChatGPT ex­plores the se­man­tic differential

Bill Benzon9 Mar 2023 13:09 UTC
7 points
2 comments7 min readLW link

AI #3

Zvi9 Mar 2023 12:20 UTC
55 points
12 comments62 min readLW link
(thezvi.wordpress.com)

The Scien­tific Ap­proach To Any­thing and Everything

Rami Rustom9 Mar 2023 11:27 UTC
5 points
5 comments16 min readLW link

Paper Sum­mary: The Effec­tive­ness of AI Ex­is­ten­tial Risk Com­mu­ni­ca­tion to the Amer­i­can and Dutch Public

otto.barten9 Mar 2023 10:47 UTC
14 points
6 comments4 min readLW link

Speed run­ning ev­ery­one through the bad al­ign­ment bingo. $5k bounty for a LW con­ver­sa­tional agent

ArthurB9 Mar 2023 9:26 UTC
140 points
33 comments2 min readLW link

Chom­sky on ChatGPT (link)

mukashi9 Mar 2023 7:00 UTC
2 points
6 comments1 min readLW link

How bad a fu­ture do ML re­searchers ex­pect?

KatjaGrace9 Mar 2023 4:50 UTC
122 points
8 comments2 min readLW link
(aiimpacts.org)

Challenge: con­struct a Gra­di­ent Hacker

9 Mar 2023 2:38 UTC
39 points
10 comments1 min readLW link

Ba­sic Facts Beanbag

Screwtape9 Mar 2023 0:05 UTC
6 points
0 comments4 min readLW link

A rank­ing scale for how se­vere the side effects of solu­tions to AI x-risk are

Christopher King8 Mar 2023 22:53 UTC
3 points
0 comments2 min readLW link

Progress links and tweets, 2023-03-08

jasoncrawford8 Mar 2023 20:37 UTC
16 points
0 comments1 min readLW link
(rootsofprogress.org)

Pro­ject “MIRI as a Ser­vice”

RomanS8 Mar 2023 19:22 UTC
42 points
4 comments1 min readLW link

2022 Sur­vey Results

Screwtape8 Mar 2023 19:16 UTC
48 points
8 comments20 min readLW link

Use the Nato Alphabet

Cedar8 Mar 2023 19:14 UTC
6 points
10 comments1 min readLW link

LessWrong needs a sage mechanic

lc8 Mar 2023 18:57 UTC
34 points
5 comments1 min readLW link

[Question] Math­e­mat­i­cal mod­els of Ethics

Victors8 Mar 2023 17:40 UTC
4 points
2 comments1 min readLW link

Against LLM Reductionism

Erich_Grunewald8 Mar 2023 15:52 UTC
140 points
17 comments18 min readLW link
(www.erichgrunewald.com)

Agency, LLMs and AI Safety—A First Pass

Giulio8 Mar 2023 15:42 UTC
2 points
0 comments4 min readLW link
(www.giuliostarace.com)

Why Un­con­trol­lable AI Looks More Likely Than Ever

8 Mar 2023 15:41 UTC
18 points
0 comments4 min readLW link
(time.com)