South Bay Meetup

DavidFriedman30 Jan 2023 23:35 UTC
2 points
0 comments1 min readLW link

Peter Thiel’s speech at Oxford De­bat­ing Union on tech­nolog­i­cal stag­na­tion, Nu­clear weapons, COVID, En­vi­ron­ment, Align­ment, ‘anti-anti anti-anti-clas­si­cal liber­al­ism’, Bostrom, LW, etc.

M. Y. Zuo30 Jan 2023 23:31 UTC
8 points
33 comments1 min readLW link

Med­i­cal Image Regis­tra­tion: The ob­scure field where Deep Me­saop­ti­miz­ers are already at the top of the bench­marks. (post + co­lab note­book)

Hastings30 Jan 2023 22:46 UTC
34 points
1 comment3 min readLW link

Hu­mans Can Be Man­u­ally Strategic

Screwtape30 Jan 2023 22:35 UTC
13 points
0 comments3 min readLW link

Why I hate the “ac­ci­dent vs. mi­suse” AI x-risk di­chotomy (quick thoughts on “struc­tural risk”)

David Scott Krueger (formerly: capybaralet)30 Jan 2023 18:50 UTC
32 points
41 comments2 min readLW link

2022 Unoffi­cial LessWrong Gen­eral Cen­sus

Screwtape30 Jan 2023 18:36 UTC
97 points
33 comments2 min readLW link

Call for sub­mis­sions: “(In)hu­man Values and Ar­tifi­cial Agency”, ALIFE 2023

the gears to ascension30 Jan 2023 17:37 UTC
29 points
4 comments1 min readLW link
(humanvaluesandartificialagency.com)

What I mean by “al­ign­ment is in large part about mak­ing cog­ni­tion aimable at all”

So8res30 Jan 2023 15:22 UTC
168 points
25 comments2 min readLW link

The En­ergy Re­quire­ments and Fea­si­bil­ity of Off-World Mining

clans30 Jan 2023 15:07 UTC
31 points
1 comment8 min readLW link
(locationtbd.home.blog)

What­ever their ar­gu­ments, Covid vac­cine scep­tics will prob­a­bly never con­vince me

contrarianbrit30 Jan 2023 13:42 UTC
8 points
10 comments3 min readLW link
(thomasprosser.substack.com)

Si­mu­lacra Levels Summary

Zvi30 Jan 2023 13:40 UTC
73 points
14 comments7 min readLW link
(thezvi.wordpress.com)

A Few Prin­ci­ples of Suc­cess­ful AI Design

Vestozia30 Jan 2023 10:42 UTC
1 point
0 comments8 min readLW link

Against Boltz­mann mesaoptimizers

porby30 Jan 2023 2:55 UTC
76 points
6 comments4 min readLW link

How Likely is Los­ing a Google Ac­count?

jefftk30 Jan 2023 0:20 UTC
52 points
11 comments3 min readLW link
(www.jefftk.com)

Model-driven feed­back could am­plify al­ign­ment failures

aogara30 Jan 2023 0:00 UTC
21 points
1 comment2 min readLW link

Take­aways from cal­ibra­tion training

Olli Järviniemi29 Jan 2023 19:09 UTC
38 points
1 comment3 min readLW link

Struc­ture, cre­ativity, and nov­elty

TsviBT29 Jan 2023 14:30 UTC
18 points
4 comments7 min readLW link

What is the ground re­al­ity of coun­tries tak­ing steps to re­cal­ibrate AI de­vel­op­ment to­wards Align­ment first?

Nebuch29 Jan 2023 13:26 UTC
8 points
6 comments3 min readLW link

Com­pendium of prob­lems with RLHF

Charbel-Raphaël29 Jan 2023 11:40 UTC
120 points
16 comments10 min readLW link

for­mal al­ign­ment: what it is, and some proposals

Tamsin Leake29 Jan 2023 11:32 UTC
53 points
3 comments1 min readLW link
(carado.moe)

My biggest take­away from Red­wood Re­search REMIX

Alok Singh29 Jan 2023 11:00 UTC
0 points
0 comments1 min readLW link
(alok.github.io)

EA novel pub­lished on Amazon

Timothy Underwood29 Jan 2023 8:33 UTC
17 points
0 comments1 min readLW link

Re­v­erse RSS Stats

jefftk29 Jan 2023 3:40 UTC
12 points
2 comments1 min readLW link
(www.jefftk.com)

Why and How to Grad­u­ate Early [U.S.]

Tego29 Jan 2023 1:28 UTC
29 points
5 comments8 min readLW link

Stop-gra­di­ents lead to fixed point predictions

28 Jan 2023 22:47 UTC
37 points
2 comments24 min readLW link

Eli Dourado AMA on the Progress Forum

jasoncrawford28 Jan 2023 22:18 UTC
19 points
0 comments1 min readLW link
(rootsofprogress.org)

LW Filter Tags (Ra­tion­al­ity/​World Model­ing now pro­moted in Lat­est Posts)

28 Jan 2023 22:14 UTC
60 points
4 comments3 min readLW link

No Fire in the Equations

Carlos Ramirez28 Jan 2023 21:16 UTC
−16 points
4 comments3 min readLW link

Op­ti­mal­ity is the tiger, and an­noy­ing the user is its teeth

Christopher King28 Jan 2023 20:20 UTC
25 points
6 comments2 min readLW link

On not get­ting con­tam­i­nated by the wrong obe­sity ideas

Natália28 Jan 2023 20:18 UTC
311 points
68 comments30 min readLW link

Ad­vice I found helpful in 2022

Akash28 Jan 2023 19:48 UTC
36 points
5 comments2 min readLW link

The Knock­down Ar­gu­ment Paradox

Bryan Frances28 Jan 2023 19:23 UTC
−12 points
6 comments8 min readLW link

Less Wrong/​ACX Bu­dapest Feb 4th Meetup

28 Jan 2023 14:49 UTC
2 points
0 comments1 min readLW link

Reflec­tions on De­cep­tion & Gen­er­al­ity in Scal­able Over­sight (Another OpenAI Align­ment Re­view)

Shoshannah Tekofsky28 Jan 2023 5:26 UTC
53 points
7 comments7 min readLW link

A Sim­ple Align­ment Typology

Shoshannah Tekofsky28 Jan 2023 5:26 UTC
34 points
2 comments2 min readLW link

Spooky ac­tion at a dis­tance in the loss landscape

28 Jan 2023 0:22 UTC
61 points
4 comments7 min readLW link
(www.jessehoogland.com)

WaPo: “Big Tech was mov­ing cau­tiously on AI. Then came ChatGPT.”

Julian Bradshaw27 Jan 2023 22:54 UTC
26 points
5 comments1 min readLW link
(www.washingtonpost.com)

Liter­a­ture re­view of TAI timelines

27 Jan 2023 20:07 UTC
35 points
7 comments2 min readLW link
(epochai.org)

Scal­ing Laws Liter­a­ture Review

Pablo Villalobos27 Jan 2023 19:57 UTC
36 points
1 comment4 min readLW link
(epochai.org)

The role of Bayesian ML in AI safety—an overview

Marius Hobbhahn27 Jan 2023 19:40 UTC
31 points
6 comments10 min readLW link

to me, it’s in­stru­men­tal­ity that is alienating

Tamsin Leake27 Jan 2023 18:27 UTC
9 points
0 comments2 min readLW link
(carado.moe)

As­sign­ing Praise and Blame: De­cou­pling Episte­mol­ogy and De­ci­sion Theory

27 Jan 2023 18:16 UTC
59 points
5 comments3 min readLW link

[Question] How could hu­mans dom­i­nate over a su­per in­tel­li­gent AI?

Marco Discendenti27 Jan 2023 18:15 UTC
−5 points
8 comments1 min readLW link

ChatGPT un­der­stands language

philosophybear27 Jan 2023 7:14 UTC
27 points
4 comments6 min readLW link
(philosophybear.substack.com)

Jar of Chocolate

jefftk27 Jan 2023 3:40 UTC
10 points
0 comments1 min readLW link
(www.jefftk.com)

Ba­sics of Ra­tion­al­ist Discourse

Duncan Sabien (Deactivated)27 Jan 2023 2:40 UTC
266 points
181 comments36 min readLW link

The re­cent ba­nal­ity of ra­tio­nal­ity (and effec­tive al­tru­ism)

CraigMichael27 Jan 2023 1:19 UTC
−6 points
7 comments11 min readLW link

11 heuris­tics for choos­ing (al­ign­ment) re­search projects

27 Jan 2023 0:36 UTC
50 points
5 comments1 min readLW link

A differ­ent ob­ser­va­tion of Vav­ilov Day

Elizabeth26 Jan 2023 21:50 UTC
30 points
1 comment1 min readLW link
(acesounderglass.com)

All AGI Safety ques­tions wel­come (es­pe­cially ba­sic ones) [~monthly thread]

26 Jan 2023 21:01 UTC
39 points
81 comments2 min readLW link