[Si­mu­la­tors sem­i­nar se­quence] #1 Back­ground & shared assumptions

2 Jan 2023 23:48 UTC
50 points
4 comments3 min readLW link

Lin­ear Alge­bra Done Right, Axler

David Udell2 Jan 2023 22:54 UTC
56 points
6 comments9 min readLW link

MacArthur BART (Filk)

Gordon Seidoh Worley2 Jan 2023 22:50 UTC
10 points
1 comment1 min readLW link

Knottiness

abramdemski2 Jan 2023 22:13 UTC
43 points
4 comments2 min readLW link

[Question] De­fault Sort for Short­forms is Very Bad; How Do I Change It?

DragonGod2 Jan 2023 21:50 UTC
15 points
0 comments1 min readLW link

MAKE IT BETTER (a po­etic demon­stra­tion of the ba­nal­ity of GPT-3)

rogersbacon2 Jan 2023 20:47 UTC
7 points
2 comments5 min readLW link

Re­view of “Make Peo­ple Bet­ter”

Metacelsus2 Jan 2023 20:30 UTC
10 points
0 comments3 min readLW link
(denovo.substack.com)

Prepar­ing for Less Privacy

jefftk2 Jan 2023 20:30 UTC
23 points
1 comment2 min readLW link
(www.jefftk.com)

Large lan­guage mod­els can provide “nor­ma­tive as­sump­tions” for learn­ing hu­man preferences

Stuart_Armstrong2 Jan 2023 19:39 UTC
29 points
12 comments3 min readLW link

On the Im­por­tance of Open Sourc­ing Re­ward Models

elandgre2 Jan 2023 19:01 UTC
18 points
5 comments6 min readLW link

Pre­dic­tion Mar­kets for Science

Vaniver2 Jan 2023 17:55 UTC
27 points
7 comments5 min readLW link

Why don’t Ra­tion­al­ists use bidets?

Lakin2 Jan 2023 17:42 UTC
31 points
33 comments2 min readLW link

Soft op­ti­miza­tion makes the value tar­get bigger

Jeremy Gillen2 Jan 2023 16:06 UTC
117 points
20 comments12 min readLW link

Re­sults from the AI test­ing hackathon

Esben Kran2 Jan 2023 15:46 UTC
13 points
0 comments1 min readLW link

In­duc­tion heads—illustrated

CallumMcDougall2 Jan 2023 15:35 UTC
114 points
9 comments3 min readLW link

Op­por­tu­nity Cost Blackmail

adamShimi2 Jan 2023 13:48 UTC
70 points
11 comments2 min readLW link
(epistemologicalvigilance.substack.com)

The ul­ti­mate limits of al­ign­ment will de­ter­mine the shape of the long term future

beren2 Jan 2023 12:47 UTC
34 points
2 comments6 min readLW link

A ker­nel of Lie theory

Alok Singh2 Jan 2023 9:20 UTC
−1 points
8 comments1 min readLW link
(alok.github.io)

Belief Bias: Bias in Eval­u­at­ing AGI X-Risks

2 Jan 2023 8:59 UTC
−10 points
1 comment1 min readLW link

Pac­ing: in­ex­pli­ca­bly good

KatjaGrace2 Jan 2023 8:30 UTC
39 points
7 comments1 min readLW link
(worldspiritsockpuppet.com)

Align­ment, Anger, and Love: Prepar­ing for the Emer­gence of Su­per­in­tel­li­gent AI

tavurth2 Jan 2023 6:16 UTC
2 points
3 comments1 min readLW link

[Question] How can to­tal world in­dex fund growth out­pace money sup­ply growth over the long term?

pando2 Jan 2023 5:33 UTC
4 points
7 comments1 min readLW link

My first year in AI alignment

Alex_Altair2 Jan 2023 1:28 UTC
61 points
10 comments7 min readLW link

Sail Over Moun­tains of ICE...

AnthonyRepetto2 Jan 2023 0:27 UTC
26 points
51 comments7 min readLW link

Fun math facts about 2023

Adam Scherlis1 Jan 2023 23:38 UTC
9 points
6 comments1 min readLW link

The Thing­ness of Things

TsviBT1 Jan 2023 22:19 UTC
48 points
35 comments10 min readLW link

Thoughts On Ex­pand­ing the AI Safety Com­mu­nity: Benefits and Challenges of Outreach to Non-Tech­ni­cal Professionals

Yashvardhan Sharma1 Jan 2023 19:21 UTC
4 points
4 comments7 min readLW link

[Question] Would it be good or bad for the US mil­i­tary to get in­volved in AI risk?

Grant Demaree1 Jan 2023 19:02 UTC
50 points
12 comments1 min readLW link

Bet­ter New Year’s Goals through Align­ing the Elephant and the Rider

moridinamael1 Jan 2023 17:54 UTC
20 points
0 comments2 min readLW link
(guildoftherose.org)

A Löbian ar­gu­ment pat­tern for im­plicit rea­son­ing in nat­u­ral lan­guage: Löbian party invitations

Andrew_Critch1 Jan 2023 17:39 UTC
23 points
8 comments7 min readLW link

woke offline, anti-woke online

Yair Halberstadt1 Jan 2023 8:24 UTC
13 points
12 comments1 min readLW link

Sum­mary of 80k’s AI prob­lem profile

JakubK1 Jan 2023 7:30 UTC
7 points
0 comments5 min readLW link
(forum.effectivealtruism.org)

What per­cent of peo­ple work in moral mazes?

Raemon1 Jan 2023 4:33 UTC
21 points
9 comments4 min readLW link

Re­cur­sive Mid­dle Man­ager Hell

Raemon1 Jan 2023 4:33 UTC
221 points
46 comments11 min readLW link1 review

Challenge to the no­tion that any­thing is (maybe) pos­si­ble with AGI

1 Jan 2023 3:57 UTC
−27 points
4 comments1 min readLW link
(mflb.com)

The Roots of Progress’s 2022 in review

jasoncrawford1 Jan 2023 2:54 UTC
14 points
2 comments15 min readLW link
(rootsofprogress.org)

In­vest­ing for a World Trans­formed by AI

PeterMcCluskey1 Jan 2023 2:47 UTC
67 points
24 comments6 min readLW link1 review
(bayesianinvestor.com)

Why Free Will is NOT an illusion

Akira Pyinya1 Jan 2023 2:29 UTC
0 points
16 comments1 min readLW link

Lo­calhost Se­cu­rity Messaging

jefftk1 Jan 2023 2:20 UTC
7 points
3 comments1 min readLW link
(www.jefftk.com)

0 and 1 aren’t probabilities

Alok Singh1 Jan 2023 0:09 UTC
2 points
4 comments2 min readLW link
(en.wikipedia.org)

‘simu­la­tor’ fram­ing and con­fu­sions about LLMs

Beth Barnes31 Dec 2022 23:38 UTC
104 points
11 comments4 min readLW link

Mon­i­tor­ing de­vices I have loved

Elizabeth31 Dec 2022 22:51 UTC
62 points
13 comments3 min readLW link1 review

Slack mat­ters more than any outcome

Valentine31 Dec 2022 20:11 UTC
156 points
56 comments19 min readLW link1 review

To Be Par­tic­u­lar About Morality

AGO31 Dec 2022 19:58 UTC
6 points
2 comments7 min readLW link

200 COP in MI: In­ter­pret­ing Al­gorith­mic Problems

Neel Nanda31 Dec 2022 19:55 UTC
33 points
2 comments10 min readLW link

The Feel­ing of Idea Scarcity

johnswentworth31 Dec 2022 17:34 UTC
246 points
22 comments5 min readLW link1 review

Curse of knowl­edge and Naive re­al­ism: Bias in Eval­u­at­ing AGI X-Risks

31 Dec 2022 13:33 UTC
−7 points
1 comment1 min readLW link
(www.lesswrong.com)

[Question] What ca­reer ad­vice do you give to soft­ware en­g­ineers?

Antb31 Dec 2022 12:01 UTC
15 points
4 comments1 min readLW link

[Question] Are Mix­ture-of-Ex­perts Trans­form­ers More In­ter­pretable Than Dense Trans­form­ers?

simeon_c31 Dec 2022 11:34 UTC
8 points
5 comments1 min readLW link

[Question] In which cases can ChatGPT be used as an aid for the­sis or sci­en­tific pa­per writ­ing?

Bob Guran31 Dec 2022 10:50 UTC
1 point
1 comment1 min readLW link