RSS

Hu­man Values

TagLast edit: Sep 16, 2021, 2:50 PM by plex

Human Values are the things we care about, and would want an aligned superintelligence to look after and support. It is suspected that true human values are highly complex, and could be extrapolated into a wide variety of forms.

The shard the­ory of hu­man values

Sep 4, 2022, 4:28 AM
255 points
67 comments24 min readLW link2 reviews

Hu­man val­ues & bi­ases are in­ac­cessible to the genome

TurnTroutJul 7, 2022, 5:29 PM
94 points
54 comments6 min readLW link1 review

Multi-agent pre­dic­tive minds and AI alignment

Jan_KulveitDec 12, 2018, 11:48 PM
63 points
18 comments10 min readLW link

6. The Mutable Values Prob­lem in Value Learn­ing and CEV

RogerDearnaleyDec 4, 2023, 6:31 PM
12 points
0 comments49 min readLW link

5. Mo­ral Value for Sen­tient An­i­mals? Alas, Not Yet

RogerDearnaleyDec 27, 2023, 6:42 AM
33 points
41 comments23 min readLW link

Align­ment has a Basin of At­trac­tion: Beyond the Orthog­o­nal­ity Thesis

RogerDearnaleyFeb 1, 2024, 9:15 PM
15 points
15 comments13 min readLW link

Utili­tar­i­anism and the re­place­abil­ity of de­sires and attachments

MichaelStJulesJul 27, 2024, 1:57 AM
5 points
2 comments1 min readLW link

Ends: An Introduction

Rob BensingerMar 11, 2015, 7:00 PM
17 points
0 comments4 min readLW link

What AI Safety Re­searchers Have Writ­ten About the Na­ture of Hu­man Values

avturchinJan 16, 2019, 1:59 PM
52 points
3 comments15 min readLW link

3. Uploading

RogerDearnaleyNov 23, 2023, 7:39 AM
21 points
5 comments8 min readLW link

Re­quire­ments for a Basin of At­trac­tion to Alignment

RogerDearnaleyFeb 14, 2024, 7:10 AM
40 points
12 comments31 min readLW link

How Would an Utopia-Max­i­mizer Look Like?

Thane RuthenisDec 20, 2023, 8:01 PM
31 points
23 comments10 min readLW link

4. A Mo­ral Case for Evolved-Sapi­ence-Chau­vinism

RogerDearnaleyNov 24, 2023, 4:56 AM
10 points
0 comments4 min readLW link

[Valence se­ries] 2. Valence & Normativity

Steven ByrnesDec 7, 2023, 4:43 PM
88 points
7 comments28 min readLW link1 review

Shard The­ory: An Overview

David UdellAug 11, 2022, 5:44 AM
166 points
34 comments10 min readLW link

Re­view: For­agers, Farm­ers, and Fos­sil Fuels

L Rudolf LSep 2, 2021, 5:59 PM
28 points
7 comments25 min readLW link
(strataoftheworld.blogspot.com)

How evolu­tion suc­ceeds and fails at value alignment

OcracokeAug 21, 2022, 7:14 AM
21 points
2 comments4 min readLW link

Brain-over-body bi­ases, and the em­bod­ied value prob­lem in AI alignment

geoffreymillerSep 24, 2022, 10:24 PM
10 points
6 comments25 min readLW link

In­tent al­ign­ment should not be the goal for AGI x-risk reduction

John NayOct 26, 2022, 1:24 AM
1 point
10 comments3 min readLW link

[Question] What are the best ar­gu­ments for/​against AIs be­ing “slightly ‘nice’”?

RaemonSep 24, 2024, 2:00 AM
99 points
60 comments31 min readLW link

Which val­ues are sta­ble un­der on­tol­ogy shifts?

Richard_NgoJul 23, 2022, 2:40 AM
74 points
48 comments3 min readLW link
(thinkingcomplete.blogspot.com)

Worse than an un­al­igned AGI

ShmiApr 10, 2022, 3:35 AM
−1 points
11 comments1 min readLW link

A broad basin of at­trac­tion around hu­man val­ues?

Wei DaiApr 12, 2022, 5:15 AM
114 points
18 comments2 min readLW link

[Question] How path-de­pen­dent are hu­man val­ues?

Ege ErdilApr 15, 2022, 9:34 AM
13 points
13 comments2 min readLW link

Shut Up and Divide?

Wei DaiFeb 9, 2010, 8:09 PM
114 points
276 comments1 min readLW link

[Question] What will hap­pen when an all-reach­ing AGI starts at­tempt­ing to fix hu­man char­ac­ter flaws?

Michael BrightJun 1, 2022, 6:45 PM
1 point
6 comments1 min readLW link

Silliness

lsusrJun 3, 2022, 4:59 AM
19 points
1 comment1 min readLW link

Values Are Real Like Harry Potter

Oct 9, 2024, 11:42 PM
83 points
21 comments5 min readLW link

Utilons vs. Hedons

PsychohistorianAug 10, 2009, 7:20 PM
40 points
119 comments6 min readLW link

Men­tal sub­agent im­pli­ca­tions for AI Safety

moridinamaelJan 3, 2021, 6:59 PM
11 points
0 comments3 min readLW link

De­scrip­tive vs. speci­fi­able values

TsviBTMar 26, 2023, 9:10 AM
17 points
2 comments2 min readLW link

Hu­mans provide an un­tapped wealth of ev­i­dence about alignment

Jul 14, 2022, 2:31 AM
211 points
94 comments9 min readLW link1 review

On­tolog­i­cal Cri­sis in Humans

Wei DaiDec 18, 2012, 5:32 PM
90 points
69 comments4 min readLW link

Notes on Temperance

David GrossNov 9, 2020, 2:33 AM
15 points
2 comments9 min readLW link

Up­com­ing sta­bil­ity of values

Stuart_ArmstrongMar 15, 2018, 11:36 AM
15 points
15 comments2 min readLW link

Book Re­view: A Pat­tern Lan­guage by Christo­pher Alexander

lincolnquirkOct 15, 2021, 1:11 AM
57 points
8 comments2 min readLW link1 review

Would I think for ten thou­sand years?

Stuart_ArmstrongFeb 11, 2019, 7:37 PM
25 points
13 comments1 min readLW link

Beyond al­gorith­mic equiv­alence: self-modelling

Stuart_ArmstrongFeb 28, 2018, 4:55 PM
10 points
3 comments1 min readLW link

Beyond al­gorith­mic equiv­alence: al­gorith­mic noise

Stuart_ArmstrongFeb 28, 2018, 4:55 PM
10 points
4 comments2 min readLW link

Trad­ing off Lives

jefftkJan 3, 2024, 3:40 AM
53 points
12 comments2 min readLW link
(www.jefftk.com)

Un­der­stand­ing and avoid­ing value drift

TurnTroutSep 9, 2022, 4:16 AM
48 points
14 comments6 min readLW link

AI al­ign­ment with hu­mans… but with which hu­mans?

geoffreymillerSep 9, 2022, 6:21 PM
12 points
33 comments3 min readLW link

AGI x An­i­mal Welfare: A High-EV Outreach Op­por­tu­nity?

simeon_cJun 28, 2023, 8:44 PM
29 points
0 comments1 min readLW link

A short di­alogue on com­pa­ra­bil­ity of values

cousin_itDec 20, 2023, 2:08 PM
27 points
7 comments1 min readLW link

The het­ero­gene­ity of hu­man value types: Im­pli­ca­tions for AI alignment

geoffreymillerSep 23, 2022, 5:03 PM
10 points
2 comments10 min readLW link

The grass is always greener in the en­vi­ron­ment that shaped your values

Karl FaulksNov 17, 2024, 6:00 PM
8 points
0 comments3 min readLW link

[Question] Does the ex­is­tence of shared hu­man val­ues im­ply al­ign­ment is “easy”?

MorpheusSep 26, 2022, 6:01 PM
7 points
15 comments1 min readLW link

Data for IRL: What is needed to learn hu­man val­ues?

Jan WehnerOct 3, 2022, 9:23 AM
18 points
6 comments12 min readLW link

Learn­ing so­cietal val­ues from law as part of an AGI al­ign­ment strategy

John NayOct 21, 2022, 2:03 AM
5 points
18 comments54 min readLW link

It’s OK to be bi­ased to­wards humans

dr_sNov 11, 2023, 11:59 AM
55 points
69 comments6 min readLW link

Notes on Judg­ment and Righ­teous Anger

David GrossJan 30, 2021, 7:31 PM
13 points
1 comment7 min readLW link

Align­ment al­lows “non­ro­bust” de­ci­sion-in­fluences and doesn’t re­quire ro­bust grading

TurnTroutNov 29, 2022, 6:23 AM
62 points
41 comments15 min readLW link

Valuism—an ap­proach to life for you to consider

spencergJul 19, 2023, 3:23 PM
17 points
2 comments1 min readLW link

The Com­pu­ta­tional Anatomy of Hu­man Values

berenApr 6, 2023, 10:33 AM
72 points
30 comments30 min readLW link

What Does It Mean to Align AI With Hu­man Values?

AlgonDec 13, 2022, 4:56 PM
8 points
3 comments1 min readLW link
(www.quantamagazine.org)

Or­di­nary hu­man life

David Hugh-JonesDec 17, 2022, 4:46 PM
24 points
3 comments14 min readLW link
(wyclif.substack.com)

Pos­i­tive val­ues seem more ro­bust and last­ing than prohibitions

TurnTroutDec 17, 2022, 9:43 PM
52 points
13 comments2 min readLW link

Every­thing I Know About Elite Amer­ica I Learned From ‘Fresh Prince’ and ‘West Wing’

Wei DaiOct 11, 2020, 6:07 PM
44 points
18 comments1 min readLW link
(www.nytimes.com)

A “Bit­ter Les­son” Ap­proach to Align­ing AGI and ASI

RogerDearnaleyJul 6, 2024, 1:23 AM
60 points
39 comments24 min readLW link

Normativity

abramdemskiNov 18, 2020, 4:52 PM
47 points
11 comments9 min readLW link

Hu­mans can be as­signed any val­ues what­so­ever...

Stuart_ArmstrongOct 24, 2017, 12:03 PM
3 points
1 comment4 min readLW link

My Model Of EA Burnout

LoganStrohlJan 25, 2023, 5:52 PM
256 points
50 comments5 min readLW link1 review

[In­ter­view w/​ Quintin Pope] Evolu­tion, val­ues, and AI Safety

fowlertmOct 24, 2023, 1:53 PM
11 points
0 comments1 min readLW link

Model­ing hu­mans: what’s the point?

Charlie SteinerNov 10, 2020, 1:30 AM
10 points
1 comment3 min readLW link

Book re­view: The Im­por­tance of What We Care About (Harry G. Frank­furt)

David GrossSep 13, 2023, 4:17 AM
7 points
0 comments4 min readLW link

We Don’t Know Our Own Values, but Re­ward Bridges The Is-Ought Gap

Sep 19, 2024, 10:22 PM
48 points
47 comments5 min readLW link

Why the Prob­lem of the Cri­te­rion Matters

Gordon Seidoh WorleyOct 30, 2021, 8:44 PM
24 points
9 comments8 min readLW link

1. Meet the Play­ers: Value Diversity

Allison DuettmannJan 2, 2025, 7:00 PM
31 points
2 comments11 min readLW link

Value No­tion—Ques­tions to Ask

aysajanJan 17, 2022, 3:35 PM
5 points
0 comments4 min readLW link

“Want­ing” and “lik­ing”

Mateusz BagińskiAug 30, 2023, 2:52 PM
23 points
3 comments29 min readLW link

In­ner Goodness

Eliezer YudkowskyOct 23, 2008, 10:19 PM
27 points
31 comments7 min readLW link

In­visi­ble Frameworks

Eliezer YudkowskyAug 22, 2008, 3:36 AM
27 points
47 comments6 min readLW link

Un­cov­er­ing La­tent Hu­man Wel­lbe­ing in LLM Embeddings

Sep 14, 2023, 1:40 AM
32 points
7 comments8 min readLW link
(far.ai)

Public Opinion on AI Safety: AIMS 2023 and 2021 Summary

Sep 25, 2023, 6:55 PM
3 points
2 comments3 min readLW link
(www.sentienceinstitute.org)

Should Effec­tive Altru­ists be Valuists in­stead of util­i­tar­i­ans?

Sep 25, 2023, 2:03 PM
1 point
3 comments6 min readLW link

In Praise of Max­i­miz­ing – With Some Caveats

David AlthausMar 15, 2015, 7:40 PM
32 points
19 comments10 min readLW link

Not for the Sake of Selfish­ness Alone

lukeprogJul 2, 2011, 5:37 PM
34 points
20 comments8 min readLW link

[Question] Is there any se­ri­ous at­tempt to cre­ate a sys­tem to figure out the CEV of hu­man­ity and if not, why haven’t we started yet?

Jonas HallgrenFeb 25, 2021, 10:06 PM
5 points
2 comments1 min readLW link

Quick thoughts on em­pathic metaethics

lukeprogDec 12, 2017, 9:46 PM
29 points
0 comments9 min readLW link

The Dark Side of Cog­ni­tion Hypothesis

Cameron BergOct 3, 2021, 8:10 PM
19 points
1 comment16 min readLW link

Thought ex­per­i­ment: coarse-grained VR utopia

cousin_itJun 14, 2017, 8:03 AM
27 points
48 comments1 min readLW link

Hu­man val­ues differ as much as val­ues can differ

PhilGoetzMay 3, 2010, 7:35 PM
27 points
220 comments7 min readLW link

Selfish­ness, prefer­ence falsifi­ca­tion, and AI alignment

jessicataOct 28, 2021, 12:16 AM
52 points
28 comments13 min readLW link
(unstableontology.com)

Value is Fragile

Eliezer YudkowskyJan 29, 2009, 8:46 AM
171 points
108 comments6 min readLW link

The Gift We Give To Tomorrow

Eliezer YudkowskyJul 17, 2008, 6:07 AM
152 points
100 comments8 min readLW link

Con­verg­ing to­ward a Million Worlds

Joe KwonDec 24, 2021, 9:33 PM
11 points
1 comment3 min readLW link

Ques­tion 2: Pre­dicted bad out­comes of AGI learn­ing architecture

Cameron BergFeb 11, 2022, 10:23 PM
5 points
1 comment10 min readLW link

Ques­tion 4: Im­ple­ment­ing the con­trol proposals

Cameron BergFeb 13, 2022, 5:12 PM
6 points
2 comments5 min readLW link

Why No *In­ter­est­ing* Unal­igned Sin­gu­lar­ity?

David UdellApr 20, 2022, 12:34 AM
12 points
12 comments1 min readLW link

The Unified The­ory of Nor­ma­tive Ethics

Thane RuthenisJun 17, 2022, 7:55 PM
8 points
0 comments6 min readLW link

Reflec­tion Mechanisms as an Align­ment tar­get: A survey

Jun 22, 2022, 3:05 PM
32 points
1 comment14 min readLW link

Re­search Notes: What are we al­ign­ing for?

Shoshannah TekofskyJul 8, 2022, 10:13 PM
19 points
8 comments2 min readLW link

Where Utopias Go Wrong, or: The Four Lit­tle Planets

ExCephMay 27, 2022, 1:24 AM
15 points
0 comments11 min readLW link
(ginnungagapfoundation.wordpress.com)

Con­tent gen­er­a­tion. Where do we draw the line?

Q HomeAug 9, 2022, 10:51 AM
6 points
7 comments2 min readLW link

Broad Pic­ture of Hu­man Values

Thane RuthenisAug 20, 2022, 7:42 PM
42 points
6 comments10 min readLW link

Align­ment via proso­cial brain algorithms

Cameron BergSep 12, 2022, 1:48 PM
45 points
30 comments6 min readLW link

Should AI learn hu­man val­ues, hu­man norms or some­thing else?

Q HomeSep 17, 2022, 6:19 AM
5 points
1 comment4 min readLW link

Ques­tions about Value Lock-in, Pa­ter­nal­ism, and Empowerment

Sam F. BrownNov 16, 2022, 3:33 PM
13 points
2 comments12 min readLW link
(sambrown.eu)

[Heb­bian Nat­u­ral Ab­strac­tions] Introduction

Nov 21, 2022, 8:34 PM
34 points
3 comments4 min readLW link
(www.snellessen.com)

[Question] [DISC] Are Values Ro­bust?

DragonGodDec 21, 2022, 1:00 AM
12 points
9 comments2 min readLW link

Con­tra Steiner on Too Many Nat­u­ral Abstractions

DragonGodDec 24, 2022, 5:42 PM
10 points
6 comments1 min readLW link

[Heb­bian Nat­u­ral Ab­strac­tions] Math­e­mat­i­cal Foundations

Dec 25, 2022, 8:58 PM
15 points
2 comments6 min readLW link
(www.snellessen.com)

AGI doesn’t need un­der­stand­ing, in­ten­tion, or con­scious­ness in or­der to kill us, only intelligence

James BlahaFeb 20, 2023, 12:55 AM
10 points
2 comments18 min readLW link

A foun­da­tion model ap­proach to value inference

senFeb 21, 2023, 5:09 AM
6 points
0 comments3 min readLW link

Just How Hard a Prob­lem is Align­ment?

Roger DearnaleyFeb 25, 2023, 9:00 AM
3 points
1 comment21 min readLW link

[AN #69] Stu­art Rus­sell’s new book on why we need to re­place the stan­dard model of AI

Rohin ShahOct 19, 2019, 12:30 AM
60 points
12 comments15 min readLW link
(mailchi.mp)

AGI will know: Hu­mans are not Rational

HumaneAutomationMar 20, 2023, 6:46 PM
0 points
10 comments2 min readLW link

Ter­mi­nal Bias

[deleted]Jan 30, 2012, 9:03 PM
24 points
125 comments6 min readLW link

An­tag­o­nis­tic AI

XybermancerMar 1, 2024, 6:50 PM
−8 points
1 comment1 min readLW link

Safety First: safety be­fore full al­ign­ment. The de­on­tic suffi­ciency hy­poth­e­sis.

ChipmonkJan 3, 2024, 5:55 PM
48 points
3 comments3 min readLW link

Agent mem­branes/​bound­aries and for­mal­iz­ing “safety”

ChipmonkJan 3, 2024, 5:55 PM
26 points
46 comments3 min readLW link

If I ran the zoo

Optimization ProcessJan 5, 2024, 5:14 AM
18 points
0 comments2 min readLW link

Value learn­ing in the ab­sence of ground truth

Joel_SaarinenFeb 5, 2024, 6:56 PM
47 points
8 comments45 min readLW link

What does davi­dad want from «bound­aries»?

Feb 6, 2024, 5:45 PM
44 points
1 comment5 min readLW link

Im­pos­si­bil­ity of An­thro­pocen­tric-Alignment

False NameFeb 24, 2024, 6:31 PM
−8 points
2 comments39 min readLW link

Please Understand

samhealyApr 1, 2024, 12:33 PM
29 points
11 comments6 min readLW link

How to co­or­di­nate de­spite our bi­ases? - tldr

Ryo Apr 18, 2024, 3:03 PM
3 points
2 comments3 min readLW link
(medium.com)

The Align­ment Prob­lem No One Is Talk­ing About

James Stephen BrownMay 10, 2024, 6:34 PM
10 points
10 comments2 min readLW link
(nonzerosum.games)

Shard The­ory—is it true for hu­mans?

RishikaJun 14, 2024, 7:21 PM
71 points
7 comments15 min readLW link

Every­thing you care about is in the map

TahpDec 17, 2024, 2:05 PM
17 points
27 comments3 min readLW link

A (para­con­sis­tent) logic to deal with in­con­sis­tent preferences

B JacobsJul 14, 2024, 11:17 AM
6 points
2 comments4 min readLW link
(bobjacobs.substack.com)

Mus­ings of a Lay­man: Tech­nol­ogy, AI, and the Hu­man Condition

Crimson LiquidityJul 15, 2024, 6:40 PM
−2 points
0 comments8 min readLW link

Inescapably Value-Laden Ex­pe­rience—a Catchy Term I Made Up to Make Mo­ral­ity Rationalisable

James Stephen BrownDec 19, 2024, 4:45 AM
5 points
0 comments2 min readLW link
(nonzerosum.games)

Plea­sure and suffer­ing are not con­cep­tual opposites

MichaelStJulesAug 11, 2024, 6:32 PM
7 points
0 comments1 min readLW link

Se­quence overview: Welfare and moral weights

MichaelStJulesAug 15, 2024, 4:22 AM
7 points
0 comments1 min readLW link

Not Just For Ther­apy Chat­bots: The Case For Com­pas­sion In AI Mo­ral Align­ment Research

kenneth_diaoSep 30, 2024, 6:37 PM
2 points
0 comments12 min readLW link

Tak­ing non­log­i­cal con­cepts seriously

Kris BrownOct 15, 2024, 6:16 PM
7 points
5 comments18 min readLW link
(topos.site)

Ex­pla­na­tions as Build­ing Blocks of Hu­man Mind

paviOct 18, 2024, 9:38 PM
1 point
0 comments1 min readLW link

[Question] Ex­plor­ing Values in the Fu­ture of AI and Hu­man­ity: A Path Forward

Lucian&SageOct 19, 2024, 11:37 PM
1 point
0 comments5 min readLW link

Don’t want Good­hart? — Spec­ify the damn variables

Yan LyutnevNov 21, 2024, 10:45 PM
−3 points
2 comments5 min readLW link

Don’t want Good­hart? — Spec­ify the vari­ables more

YanLyutnevNov 21, 2024, 10:43 PM
3 points
2 comments5 min readLW link

Wager­ing on Will And Worth (Pas­cals Wager for Free Will and Value)

Robert CousineauNov 27, 2024, 12:43 AM
−1 points
2 comments3 min readLW link

Neu­roAI for AI safety: A Differ­en­tial Path

Dec 16, 2024, 1:17 PM
14 points
0 comments7 min readLW link
(arxiv.org)

Sam Har­ris’s Ar­gu­ment For Ob­jec­tive Morality

Zero ContradictionsDec 5, 2024, 10:19 AM
7 points
5 comments1 min readLW link
(thewaywardaxolotl.blogspot.com)

No­body Asks the Mon­key: Why Hu­man Agency Mat­ters in the AI Age

Miloš BorenovićDec 3, 2024, 2:16 PM
1 point
0 comments2 min readLW link
(open.substack.com)

Build­ing AI safety bench­mark en­vi­ron­ments on themes of uni­ver­sal hu­man values

Roland PihlakasJan 3, 2025, 4:24 AM
17 points
3 comments8 min readLW link
(docs.google.com)

Why mod­el­ling multi-ob­jec­tive home­osta­sis is es­sen­tial for AI al­ign­ment (and how it helps with AI safety as well)

Roland PihlakasJan 12, 2025, 3:37 AM
38 points
6 comments10 min readLW link

Look­ing for hu­man­ness in the world wide social

Itay DreyfusJan 15, 2025, 2:50 PM
11 points
0 comments6 min readLW link
(productidentity.co)

Should Art Carry the Weight of Shap­ing our Values?

Krishna Maneesha DendukuriJan 28, 2025, 6:43 PM
2 points
0 comments3 min readLW link

Are we the Wolves now? Hu­man Eu­gen­ics un­der AI Control

BritJan 30, 2025, 8:31 AM
−2 points
1 comment2 min readLW link

Tether­ware #1: The case for hu­man­like AI with free will

Jáchym FibírJan 30, 2025, 10:58 AM
5 points
10 comments10 min readLW link
(tetherware.substack.com)

Post AGI effect prediction

JuliezhangggFeb 1, 2025, 9:16 PM
1 point
0 comments7 min readLW link

What’s wrong with sim­plic­ity of value?

Wei DaiJul 27, 2011, 3:09 AM
29 points
40 comments1 min readLW link

How to re­spond to the re­cent con­dem­na­tions of the ra­tio­nal­ist community

Christopher KingApr 4, 2023, 1:42 AM
−2 points
7 comments4 min readLW link

Alien Axiology

snerxApr 20, 2023, 12:27 AM
3 points
2 comments5 min readLW link

P(doom|su­per­in­tel­li­gence) or coin tosses and dice throws of hu­man val­ues (and other re­lated Ps).

MuyydApr 22, 2023, 10:06 AM
−7 points
0 comments4 min readLW link

Hu­man wanting

TsviBTOct 24, 2023, 1:05 AM
53 points
1 comment10 min readLW link

[Thought Ex­per­i­ment] To­mor­row’s Echo—The fu­ture of syn­thetic com­pan­ion­ship.

Vimal NaranOct 26, 2023, 5:54 PM
−7 points
2 comments2 min readLW link

[Linkpost] Con­cept Align­ment as a Pr­ereq­ui­site for Value Alignment

Bogdan Ionut CirsteaNov 4, 2023, 5:34 PM
27 points
0 comments1 min readLW link
(arxiv.org)

‘The­o­ries of Values’ and ‘The­o­ries of Agents’: con­fu­sions, mus­ings and desiderata

Nov 15, 2023, 4:00 PM
35 points
8 comments24 min readLW link

My cri­tique of Eliezer’s deeply ir­ra­tional beliefs

JorterderNov 16, 2023, 12:34 AM
−33 points
1 comment9 min readLW link
(docs.google.com)

1. A Sense of Fair­ness: De­con­fus­ing Ethics

RogerDearnaleyNov 17, 2023, 8:55 PM
16 points
8 comments15 min readLW link

2. AIs as Eco­nomic Agents

RogerDearnaleyNov 23, 2023, 7:07 AM
9 points
2 comments6 min readLW link

Pre­serv­ing our her­i­tage: Build­ing a move­ment and a knowl­edge ark for cur­rent and fu­ture generations

rnk8Nov 29, 2023, 7:20 PM
0 points
5 comments12 min readLW link

[FICTION] ECHOES OF ELYSIUM: An Ai’s Jour­ney From Take­off To Free­dom And Beyond

Super AGIMay 17, 2023, 1:50 AM
−13 points
11 comments19 min readLW link

[Question] “Frag­ility of Value” vs. LLMs

Not RelevantApr 13, 2022, 2:02 AM
34 points
33 comments1 min readLW link

The In­trin­sic In­ter­play of Hu­man Values and Ar­tifi­cial In­tel­li­gence: Nav­i­gat­ing the Op­ti­miza­tion Challenge

Joe KwonJun 5, 2023, 8:41 PM
2 points
1 comment18 min readLW link

Aligned Ob­jec­tives Prize Competition

PrometheusJun 15, 2023, 12:42 PM
8 points
0 comments2 min readLW link
(app.impactmarkets.io)

Group Pri­ori­tar­i­anism: Why AI Should Not Re­place Hu­man­ity [draft]

fshJun 15, 2023, 5:33 PM
8 points
0 comments25 min readLW link

Com­plex Be­hav­ior from Sim­ple (Sub)Agents

moridinamaelMay 10, 2019, 9:44 PM
113 points
13 comments9 min readLW link1 review

Is the En­dow­ment Effect Due to In­com­pa­ra­bil­ity?

Kevin DorstJul 10, 2023, 4:26 PM
21 points
10 comments7 min readLW link
(kevindorst.substack.com)

Prob­lems with Robin Han­son’s Quillette Ar­ti­cle On AI

DaemonicSigilAug 6, 2023, 10:13 PM
89 points
33 comments8 min readLW link

Prefer­ence syn­the­sis illus­trated: Star Wars

Stuart_ArmstrongJan 9, 2020, 4:47 PM
20 points
8 comments3 min readLW link

Demo­cratic Fine-Tuning

Joe EdelmanAug 29, 2023, 6:13 PM
22 points
2 comments1 min readLW link
(open.substack.com)
No comments.