Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
A basic lexicon of telic concepts
mrcbarbier
Oct 22, 2022, 9:28 PM
2
points
0
comments
3
min read
LW
link
Do we have the right kind of math for roles, goals and meaning?
mrcbarbier
Oct 22, 2022, 9:28 PM
13
points
5
comments
7
min read
LW
link
[Question]
The Last Year - is there an existing novel about the last year before AI doom?
Luca Petrolati
Oct 22, 2022, 8:44 PM
4
points
4
comments
1
min read
LW
link
The highest-probability outcome can be out of distribution
tailcalled
Oct 22, 2022, 8:00 PM
14
points
5
comments
1
min read
LW
link
Newsletter for Alignment Research: The ML Safety Updates
Esben Kran
Oct 22, 2022, 4:17 PM
25
points
0
comments
LW
link
Crypto loves impact markets: Notes from Schelling Point Bogotá
Rachel Shu
Oct 22, 2022, 3:58 PM
17
points
2
comments
LW
link
[Question]
When trying to define general intelligence is ability to achieve goals the best metric?
jmh
Oct 22, 2022, 3:09 AM
5
points
0
comments
1
min read
LW
link
[Question]
Simple question about corrigibility and values in AI.
jmh
Oct 22, 2022, 2:59 AM
6
points
1
comment
1
min read
LW
link
Moorean Statements
David Udell
Oct 22, 2022, 12:50 AM
11
points
11
comments
1
min read
LW
link
Wisdom Cannot Be Unzipped
Sable
Oct 22, 2022, 12:28 AM
74
points
17
comments
7
min read
LW
link
1
review
(affablyevil.substack.com)
A framework and open questions for game theoretic shard modeling
Garrett Baker
Oct 21, 2022, 9:40 PM
11
points
4
comments
4
min read
LW
link
Cooperators are more powerful than agents
Ivan Vendrov
Oct 21, 2022, 8:02 PM
29
points
7
comments
3
min read
LW
link
Intelligent behaviour across systems, scales and substrates
Nora_Ammann
Oct 21, 2022, 5:09 PM
11
points
0
comments
10
min read
LW
link
Deepfake(?) Phishing
jefftk
Oct 21, 2022, 2:30 PM
37
points
9
comments
1
min read
LW
link
(www.jefftk.com)
acronyms ftw
Emrik
Oct 21, 2022, 1:36 PM
−2
points
5
comments
2
min read
LW
link
Crossword puzzle: LessWrong Halloween 2022
jchan
Oct 21, 2022, 12:41 PM
11
points
11
comments
1
min read
LW
link
Weekly Roundup #2
Zvi
Oct 21, 2022, 12:10 PM
37
points
2
comments
11
min read
LW
link
(thezvi.wordpress.com)
Improved Security to Prevent Hacker-AI and Digital Ghosts
Erland Wittkotter
Oct 21, 2022, 10:11 AM
4
points
3
comments
12
min read
LW
link
Two Guts
chanamessinger
Oct 21, 2022, 10:01 AM
21
points
0
comments
LW
link
The importance of studying subjective experience
Q Home
Oct 21, 2022, 8:43 AM
10
points
3
comments
7
min read
LW
link
Legal Brief: Plurality Voting is Unconstitutional
c.trout
Oct 21, 2022, 4:55 AM
6
points
20
comments
11
min read
LW
link
(medium.com)
Learning societal values from law as part of an AGI alignment strategy
John Nay
Oct 21, 2022, 2:03 AM
5
points
18
comments
54
min read
LW
link
Covid 10/20/22: Wait, We Did WHAT?
Zvi
Oct 20, 2022, 9:50 PM
55
points
16
comments
16
min read
LW
link
(thezvi.wordpress.com)
When apparently positive evidence can be negative evidence
cata
Oct 20, 2022, 9:47 PM
19
points
5
comments
1
min read
LW
link
(www.ncbi.nlm.nih.gov)
Plans Are Predictions, Not Optimization Targets
johnswentworth
Oct 20, 2022, 9:17 PM
108
points
20
comments
4
min read
LW
link
1
review
Introduction to abstract entropy
Alex_Altair
Oct 20, 2022, 9:03 PM
238
points
78
comments
18
min read
LW
link
1
review
Trajectories to 2036
ukc10014
Oct 20, 2022, 8:23 PM
3
points
1
comment
14
min read
LW
link
[Question]
Rough Sketch for Product to Enhance Citizen Participation in Politics
T431
Oct 20, 2022, 8:04 PM
13
points
5
comments
1
min read
LW
link
The heritability of human values: A behavior genetic critique of Shard Theory
geoffreymiller
Oct 20, 2022, 3:51 PM
82
points
63
comments
21
min read
LW
link
A Longtermist case against Veganism
Connor Tabarrok
Oct 20, 2022, 2:30 PM
−3
points
3
comments
LW
link
AI Research Program Prediction Markets
tailcalled
Oct 20, 2022, 1:42 PM
38
points
10
comments
1
min read
LW
link
[Question]
Is the meaning of words chosen/interpreted to maximize correlations with other relevant queries?
tailcalled
Oct 20, 2022, 10:03 AM
9
points
9
comments
1
min read
LW
link
How to Write Readable Posts
David Hartsough
Oct 20, 2022, 7:48 AM
7
points
0
comments
LW
link
Notes on “Can you control the past”
So8res
Oct 20, 2022, 3:41 AM
64
points
41
comments
21
min read
LW
link
Rhythmic Baby Toys
jefftk
Oct 20, 2022, 1:50 AM
15
points
1
comment
1
min read
LW
link
(www.jefftk.com)
[Question]
What Does AI Alignment Success Look Like?
Shmi
Oct 20, 2022, 12:32 AM
23
points
7
comments
1
min read
LW
link
Scaling Laws for Reward Model Overoptimization
leogao
,
John Schulman
and
Jacob_Hilton
Oct 20, 2022, 12:20 AM
103
points
13
comments
1
min read
LW
link
(arxiv.org)
What is Consciousness?
belkarx
Oct 19, 2022, 9:14 PM
3
points
2
comments
2
min read
LW
link
What to do if a nuclear weapon is used in Ukraine?
Valentin2026
Oct 19, 2022, 6:43 PM
13
points
9
comments
3
min read
LW
link
[Question]
If I asked for an explanation of a perfect Utopia, could you give one?
Akkira
Oct 19, 2022, 5:56 PM
−4
points
2
comments
1
min read
LW
link
[Question]
Should we push for requiring AI training data to be licensed?
ChristianKl
Oct 19, 2022, 5:49 PM
37
points
32
comments
1
min read
LW
link
Hacker-AI and Digital Ghosts – Pre-AGI
Erland Wittkotter
Oct 19, 2022, 3:33 PM
9
points
7
comments
8
min read
LW
link
The reward function is already how well you manipulate humans
Kerry
Oct 19, 2022, 1:52 AM
20
points
9
comments
2
min read
LW
link
Response to Katja Grace’s AI x-risk counterarguments
Erik Jenner
and
Johannes Treutlein
Oct 19, 2022, 1:17 AM
77
points
18
comments
15
min read
LW
link
(OLD) An Extremely Opinionated Annotated List of My Favourite Mechanistic Interpretability Papers
Neel Nanda
Oct 18, 2022, 9:08 PM
72
points
5
comments
12
min read
LW
link
(www.neelnanda.io)
Distilled Representations Research Agenda
Hoagy
and
mishajw
Oct 18, 2022, 8:59 PM
15
points
2
comments
8
min read
LW
link
Drafting a Covid Survey
jefftk
Oct 18, 2022, 7:30 PM
15
points
2
comments
2
min read
LW
link
(www.jefftk.com)
How To Make Prediction Markets Useful For Alignment Work
johnswentworth
Oct 18, 2022, 7:01 PM
97
points
18
comments
2
min read
LW
link
A conversation about Katja’s counterarguments to AI risk
Matthew Barnett
and
Ege Erdil
Oct 18, 2022, 6:40 PM
43
points
9
comments
33
min read
LW
link
ACX Zurich October Meetup
MB
Oct 18, 2022, 6:24 PM
1
point
1
comment
1
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel