Seat­tle Win­ter Solstice

a7xDec 20, 2023, 8:30 PM
6 points
1 comment1 min readLW link

How Would an Utopia-Max­i­mizer Look Like?

Thane RuthenisDec 20, 2023, 8:01 PM
31 points
23 comments10 min readLW link

Succession

Richard_NgoDec 20, 2023, 7:25 PM
159 points
48 comments11 min readLW link
(www.narrativeark.xyz)

Me­tac­u­lus In­tro­duces Mul­ti­ple Choice Questions

ChristianWilliamsDec 20, 2023, 7:00 PM
4 points
0 comments1 min readLW link
(www.metaculus.com)

Brighter Than To­day Versions

jefftkDec 20, 2023, 6:20 PM
16 points
2 comments2 min readLW link
(www.jefftk.com)

Gaia Net­work: a prac­ti­cal, in­cre­men­tal path­way to Open Agency Architecture

Dec 20, 2023, 5:11 PM
22 points
8 comments16 min readLW link

On the fu­ture of lan­guage models

owencbDec 20, 2023, 4:58 PM
105 points
17 comments1 min readLW link

[Valence se­ries] Ap­pendix A: He­donic tone /​ (dis)plea­sure /​ (dis)liking

Steven ByrnesDec 20, 2023, 3:54 PM
18 points
0 comments13 min readLW link

Ma­trix com­ple­tion prize results

paulfchristianoDec 20, 2023, 3:40 PM
41 points
0 comments2 min readLW link
(www.alignment.org)

[Question] What’s the min­i­mal ad­di­tive con­stant for Kol­mogorov Com­plex­ity that a pro­gram­ming lan­guage can achieve?

Noosphere89Dec 20, 2023, 3:36 PM
11 points
15 comments1 min readLW link

Le­gal­ize bu­tanol?

bhauthDec 20, 2023, 2:24 PM
39 points
20 comments5 min readLW link
(www.bhauth.com)

A short di­alogue on com­pa­ra­bil­ity of values

cousin_itDec 20, 2023, 2:08 PM
27 points
7 comments1 min readLW link

In­side View, Out­side View… And Op­pos­ing View

chaosmageDec 20, 2023, 12:35 PM
21 points
1 comment5 min readLW link

Heuris­tics for pre­vent­ing ma­jor life mistakes

SK2Dec 20, 2023, 8:01 AM
28 points
2 comments3 min readLW link

What should be reified?

herschelDec 20, 2023, 4:52 AM
4 points
2 comments2 min readLW link
(brothernin.substack.com)

(In)ap­pro­pri­ate (De)reification

herschelDec 20, 2023, 4:51 AM
10 points
1 comment4 min readLW link
(brothernin.substack.com)

Es­cap­ing Skeuomorphism

Stuart JohnsonDec 20, 2023, 3:51 AM
28 points
0 comments8 min readLW link

Ronny and Nate dis­cuss what sorts of minds hu­man­ity is likely to find by Ma­chine Learning

Dec 19, 2023, 11:39 PM
40 points
30 comments25 min readLW link

[Question] What are the best Siderea posts?

mike_hawkeDec 19, 2023, 11:07 PM
17 points
2 comments1 min readLW link

Mean­ing & Agency

abramdemskiDec 19, 2023, 10:27 PM
91 points
17 comments14 min readLW link

s/​acc: Safe Ac­cel­er­a­tionism Manifesto

lorepieriDec 19, 2023, 10:19 PM
−4 points
5 comments2 min readLW link
(lorenzopieri.com)

Don’t Share In­for­ma­tion Exfo­haz­ardous on Others’ AI-Risk Models

Thane RuthenisDec 19, 2023, 8:09 PM
66 points
11 comments1 min readLW link

Paper: Tell, Don’t Show- Declar­a­tive facts in­fluence how LLMs generalize

Dec 19, 2023, 7:14 PM
45 points
4 comments6 min readLW link
(arxiv.org)

In­ter­view: Ap­pli­ca­tions w/​ Alice Rigg

jacobhaimesDec 19, 2023, 7:03 PM
12 points
0 comments1 min readLW link
(into-ai-safety.github.io)

How does a toy 2 digit sub­trac­tion trans­former pre­dict the sign of the out­put?

Evan AndersDec 19, 2023, 6:56 PM
14 points
0 comments8 min readLW link
(evanhanders.blog)

In­cre­men­tal AI Risks from Proxy-Simulations

kmenouDec 19, 2023, 6:56 PM
2 points
0 comments1 min readLW link
(individual.utoronto.ca)

A propo­si­tion for the mod­ifi­ca­tion of our epistemology

JacobBowdenDec 19, 2023, 6:55 PM
−4 points
2 comments4 min readLW link

Goal-Com­plete­ness is like Tur­ing-Com­plete­ness for AGI

LironDec 19, 2023, 6:12 PM
50 points
26 comments3 min readLW link

So­ci­aLLM: pro­posal for a lan­guage model de­sign for per­son­al­ised apps, so­cial sci­ence, and AI safety research

Roman LeventovDec 19, 2023, 4:49 PM
17 points
5 comments3 min readLW link

Chord­ing “The Next Right Thing”

jefftkDec 19, 2023, 3:40 PM
11 points
0 comments2 min readLW link
(www.jefftk.com)

Monthly Roundup #13: De­cem­ber 2023

ZviDec 19, 2023, 3:10 PM
32 points
5 comments26 min readLW link
(thezvi.wordpress.com)

Effec­tive Asper­sions: How the Non­lin­ear In­ves­ti­ga­tion Went Wrong

TracingWoodgrainsDec 19, 2023, 12:00 PM
188 points
171 comments1 min readLW link1 review

A Univer­sal Emer­gent De­com­po­si­tion of Retrieval Tasks in Lan­guage Models

Dec 19, 2023, 11:52 AM
84 points
3 comments10 min readLW link
(arxiv.org)

Assess­ment of AI safety agen­das: think about the down­side risk

Roman LeventovDec 19, 2023, 9:00 AM
13 points
1 comment1 min readLW link

Con­stel­la­tions are Younger than Continents

Jeffrey HeningerDec 19, 2023, 6:12 AM
261 points
22 comments2 min readLW link

The Dark Arts

Dec 19, 2023, 4:41 AM
132 points
49 comments9 min readLW link

When sci­en­tists con­sider whether their re­search will end the world

HarlanDec 19, 2023, 3:47 AM
30 points
4 comments11 min readLW link
(blog.aiimpacts.org)

Is the far fu­ture in­evitably zero sum?

Srdjan MileticDec 19, 2023, 1:45 AM
8 points
2 comments2 min readLW link
(dissent.blog)

The ‘Ne­glected Ap­proaches’ Ap­proach: AE Stu­dio’s Align­ment Agenda

Dec 18, 2023, 8:35 PM
168 points
21 comments12 min readLW link

The Short­est Path Between Scylla and Charybdis

Thane RuthenisDec 18, 2023, 8:08 PM
50 points
8 comments5 min readLW link

OpenAI: Pre­pared­ness framework

Zach Stein-PerlmanDec 18, 2023, 6:30 PM
70 points
23 comments4 min readLW link
(openai.com)

[Valence se­ries] 5. “Valence Di­sor­ders” in Men­tal Health & Personality

Steven ByrnesDec 18, 2023, 3:26 PM
43 points
12 comments13 min readLW link

Dis­cus­sion: Challenges with Un­su­per­vised LLM Knowl­edge Discovery

Dec 18, 2023, 11:58 AM
147 points
21 comments10 min readLW link

In­ter­pret­ing the Learn­ing of Deceit

RogerDearnaleyDec 18, 2023, 8:12 AM
30 points
14 comments9 min readLW link

Talk: “AI Would Be A Lot Less Alarm­ing If We Un­der­stood Agents”

johnswentworthDec 17, 2023, 11:46 PM
58 points
3 comments1 min readLW link
(www.youtube.com)

∀: a story

Richard_NgoDec 17, 2023, 10:42 PM
37 points
1 comment8 min readLW link
(www.narrativeark.xyz)

Re­viv­ing a 2015 MacBook

jefftkDec 17, 2023, 9:00 PM
11 points
0 comments1 min readLW link
(www.jefftk.com)

A Com­mon-Sense Case For Mu­tu­ally-Misal­igned AGIs Ally­ing Against Humans

Thane RuthenisDec 17, 2023, 8:28 PM
29 points
7 comments11 min readLW link

The Limits of Ar­tifi­cial Con­scious­ness: A Biol­ogy-Based Cri­tique of Chalmers’ Fad­ing Qualia Argument

Štěpán LosDec 17, 2023, 7:11 PM
−6 points
9 comments17 min readLW link

What makes teach­ing math special

ViliamDec 17, 2023, 2:15 PM
41 points
27 comments11 min readLW link