Telic in­tu­itions across the sciences

mrcbarbier22 Oct 2022 21:31 UTC
4 points
0 comments17 min readLW link

A ba­sic lex­i­con of telic concepts

mrcbarbier22 Oct 2022 21:28 UTC
2 points
0 comments3 min readLW link

Do we have the right kind of math for roles, goals and mean­ing?

mrcbarbier22 Oct 2022 21:28 UTC
13 points
5 comments7 min readLW link

[Question] The Last Year - is there an ex­ist­ing novel about the last year be­fore AI doom?

Luca Petrolati22 Oct 2022 20:44 UTC
4 points
4 comments1 min readLW link

The high­est-prob­a­bil­ity out­come can be out of distribution

tailcalled22 Oct 2022 20:00 UTC
14 points
5 comments1 min readLW link

Newslet­ter for Align­ment Re­search: The ML Safety Updates

Esben Kran22 Oct 2022 16:17 UTC
25 points
0 comments1 min readLW link

Crypto loves im­pact mar­kets: Notes from Schel­ling Point Bogotá

Rachel Shu22 Oct 2022 15:58 UTC
17 points
2 comments1 min readLW link

[Question] When try­ing to define gen­eral in­tel­li­gence is abil­ity to achieve goals the best met­ric?

jmh22 Oct 2022 3:09 UTC
5 points
0 comments1 min readLW link

[Question] Sim­ple ques­tion about cor­rigi­bil­ity and val­ues in AI.

jmh22 Oct 2022 2:59 UTC
6 points
1 comment1 min readLW link

Moorean Statements

David Udell22 Oct 2022 0:50 UTC
11 points
11 comments1 min readLW link

Wis­dom Can­not Be Unzipped

Sable22 Oct 2022 0:28 UTC
74 points
17 comments7 min readLW link1 review
(affablyevil.substack.com)

A frame­work and open ques­tions for game the­o­retic shard modeling

Garrett Baker21 Oct 2022 21:40 UTC
11 points
4 comments4 min readLW link

Co­op­er­a­tors are more pow­er­ful than agents

Ivan Vendrov21 Oct 2022 20:02 UTC
22 points
7 comments3 min readLW link

In­tel­li­gent be­havi­our across sys­tems, scales and substrates

Nora_Ammann21 Oct 2022 17:09 UTC
11 points
0 comments10 min readLW link

Deep­fake(?) Phishing

jefftk21 Oct 2022 14:30 UTC
37 points
9 comments1 min readLW link
(www.jefftk.com)

acronyms ftw

Emrik21 Oct 2022 13:36 UTC
−2 points
5 comments2 min readLW link

Cross­word puz­zle: LessWrong Hal­loween 2022

jchan21 Oct 2022 12:41 UTC
11 points
11 comments1 min readLW link

Weekly Roundup #2

Zvi21 Oct 2022 12:10 UTC
37 points
2 comments11 min readLW link
(thezvi.wordpress.com)

Im­proved Se­cu­rity to Prevent Hacker-AI and Digi­tal Ghosts

Erland Wittkotter21 Oct 2022 10:11 UTC
4 points
3 comments12 min readLW link

Two Guts

chanamessinger21 Oct 2022 10:01 UTC
21 points
0 comments1 min readLW link

The im­por­tance of study­ing sub­jec­tive experience

Q Home21 Oct 2022 8:43 UTC
10 points
3 comments7 min readLW link

Le­gal Brief: Plu­ral­ity Vot­ing is Unconstitutional

c.trout21 Oct 2022 4:55 UTC
6 points
20 comments11 min readLW link
(medium.com)

Learn­ing so­cietal val­ues from law as part of an AGI al­ign­ment strategy

John Nay21 Oct 2022 2:03 UTC
5 points
18 comments54 min readLW link

Covid 10/​20/​22: Wait, We Did WHAT?

Zvi20 Oct 2022 21:50 UTC
55 points
16 comments16 min readLW link
(thezvi.wordpress.com)

When ap­par­ently pos­i­tive ev­i­dence can be nega­tive evidence

cata20 Oct 2022 21:47 UTC
19 points
5 comments1 min readLW link
(www.ncbi.nlm.nih.gov)

Plans Are Pre­dic­tions, Not Op­ti­miza­tion Targets

johnswentworth20 Oct 2022 21:17 UTC
108 points
20 comments4 min readLW link1 review

In­tro­duc­tion to ab­stract entropy

Alex_Altair20 Oct 2022 21:03 UTC
234 points
78 comments18 min readLW link1 review

Tra­jec­to­ries to 2036

ukc1001420 Oct 2022 20:23 UTC
3 points
1 comment14 min readLW link

[Question] Rough Sketch for Product to En­hance Ci­ti­zen Par­ti­ci­pa­tion in Politics

Fer32dwt34r3dfsz20 Oct 2022 20:04 UTC
13 points
3 comments1 min readLW link

The her­i­ta­bil­ity of hu­man val­ues: A be­hav­ior ge­netic cri­tique of Shard Theory

geoffreymiller20 Oct 2022 15:51 UTC
80 points
59 comments21 min readLW link

A Longter­mist case against Veganism

Connor Tabarrok20 Oct 2022 14:30 UTC
−3 points
3 comments1 min readLW link

AI Re­search Pro­gram Pre­dic­tion Markets

tailcalled20 Oct 2022 13:42 UTC
38 points
10 comments1 min readLW link

[Question] Is the mean­ing of words cho­sen/​in­ter­preted to max­i­mize cor­re­la­tions with other rele­vant queries?

tailcalled20 Oct 2022 10:03 UTC
9 points
9 comments1 min readLW link

How to Write Read­able Posts

David Hartsough20 Oct 2022 7:48 UTC
7 points
0 comments1 min readLW link

Notes on “Can you con­trol the past”

So8res20 Oct 2022 3:41 UTC
57 points
41 comments21 min readLW link

Rhyth­mic Baby Toys

jefftk20 Oct 2022 1:50 UTC
15 points
1 comment1 min readLW link
(www.jefftk.com)

[Question] What Does AI Align­ment Suc­cess Look Like?

Shmi20 Oct 2022 0:32 UTC
23 points
7 comments1 min readLW link

Scal­ing Laws for Re­ward Model Overoptimization

20 Oct 2022 0:20 UTC
103 points
13 comments1 min readLW link
(arxiv.org)

What is Con­scious­ness?

belkarx19 Oct 2022 21:14 UTC
3 points
2 comments2 min readLW link

What to do if a nu­clear weapon is used in Ukraine?

Valentin202619 Oct 2022 18:43 UTC
8 points
4 comments3 min readLW link

[Question] If I asked for an ex­pla­na­tion of a perfect Utopia, could you give one?

Akkira19 Oct 2022 17:56 UTC
−4 points
2 comments1 min readLW link

[Question] Should we push for re­quiring AI train­ing data to be li­censed?

ChristianKl19 Oct 2022 17:49 UTC
37 points
32 comments1 min readLW link

Hacker-AI and Digi­tal Ghosts – Pre-AGI

Erland Wittkotter19 Oct 2022 15:33 UTC
9 points
7 comments8 min readLW link

The re­ward func­tion is already how well you ma­nipu­late humans

Kerry19 Oct 2022 1:52 UTC
20 points
9 comments2 min readLW link

Re­sponse to Katja Grace’s AI x-risk counterarguments

19 Oct 2022 1:17 UTC
77 points
18 comments15 min readLW link

(OLD) An Ex­tremely Opinionated An­no­tated List of My Favourite Mechanis­tic In­ter­pretabil­ity Papers

Neel Nanda18 Oct 2022 21:08 UTC
72 points
5 comments12 min readLW link
(www.neelnanda.io)

Distil­led Rep­re­sen­ta­tions Re­search Agenda

18 Oct 2022 20:59 UTC
15 points
2 comments8 min readLW link

Draft­ing a Covid Survey

jefftk18 Oct 2022 19:30 UTC
15 points
2 comments2 min readLW link
(www.jefftk.com)

How To Make Pre­dic­tion Mar­kets Use­ful For Align­ment Work

johnswentworth18 Oct 2022 19:01 UTC
97 points
18 comments2 min readLW link

A con­ver­sa­tion about Katja’s coun­ter­ar­gu­ments to AI risk

18 Oct 2022 18:40 UTC
43 points
9 comments33 min readLW link