Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
[Question]
Should you write under a blog or your own name?
Dalton Mabery
Jul 6, 2022, 3:26 PM
2
points
2
comments
1
min read
LW
link
Carrying the Torch: A Response to Anna Salamon by the Guild of the Rose
moridinamael
Jul 6, 2022, 2:20 PM
136
points
16
comments
6
min read
LW
link
Predicting Parental Emotional Changes?
jefftk
Jul 6, 2022, 1:50 PM
39
points
11
comments
2
min read
LW
link
(www.jefftk.com)
Berlin AI Safety Open Meetup July 2022
pranomostro
Jul 6, 2022, 12:41 PM
6
points
0
comments
1
min read
LW
link
Forecasting Through Fiction
Yitz
Jul 6, 2022, 5:03 AM
5
points
2
comments
8
min read
LW
link
Introducing the Fund for Alignment Research (We’re Hiring!)
AdamGleave
,
Scott Emmons
,
Ethan Perez
and
Claudia Shi
Jul 6, 2022, 2:07 AM
62
points
0
comments
4
min read
LW
link
My vision of a good future, part I
Jeffrey Ladish
Jul 6, 2022, 1:23 AM
66
points
18
comments
9
min read
LW
link
Imperial Russia was doing fine without the Soviets
Davis Kedrosky
Jul 5, 2022, 10:24 PM
6
points
3
comments
14
min read
LW
link
(daviskedrosky.substack.com)
A Pattern Language For Rationality
Vaniver
Jul 5, 2022, 7:08 PM
75
points
14
comments
15
min read
LW
link
How to destroy the universe with a hypercomputer
Trevor Cappallo
Jul 5, 2022, 7:05 PM
2
points
3
comments
1
min read
LW
link
The curious case of Pretty Good human inner/outer alignment
PavleMiha
Jul 5, 2022, 7:04 PM
41
points
45
comments
4
min read
LW
link
When is it appropriate to use statistical models and probabilities for decision making ?
Younes Kamel
Jul 5, 2022, 12:34 PM
10
points
7
comments
4
min read
LW
link
(youneskamel.substack.com)
Goal Factoring
CFAR!Duncan
Jul 5, 2022, 7:10 AM
92
points
2
comments
8
min read
LW
link
Assorted thoughts about abstraction
Adam Zerner
Jul 5, 2022, 6:40 AM
16
points
9
comments
7
min read
LW
link
[AN #172] Sorry for the long hiatus!
Rohin Shah
Jul 5, 2022, 6:20 AM
54
points
0
comments
3
min read
LW
link
(mailchi.mp)
Outline: The Rectifying of Maps
hamnox
Jul 5, 2022, 5:14 AM
7
points
0
comments
2
min read
LW
link
[Question]
Seeking opinions on the current and forward state of cryptocurrencies.
jmh
Jul 5, 2022, 5:01 AM
6
points
6
comments
1
min read
LW
link
ITT-passing and civility are good; “charity” is bad; steelmanning is niche
Rob Bensinger
Jul 5, 2022, 12:15 AM
163
points
36
comments
6
min read
LW
link
1
review
Please help us communicate AI xrisk. It could save the world.
otto.barten
Jul 4, 2022, 9:47 PM
4
points
7
comments
2
min read
LW
link
Benchmark for successful concept extrapolation/avoiding goal misgeneralization
Stuart_Armstrong
Jul 4, 2022, 8:48 PM
83
points
12
comments
4
min read
LW
link
Procedural Executive Function, Part 1
DaystarEld
Jul 4, 2022, 6:51 PM
52
points
8
comments
14
min read
LW
link
(daystareld.com)
Anthropic’s SoLU (Softmax Linear Unit)
Joel Burget
Jul 4, 2022, 6:38 PM
21
points
1
comment
4
min read
LW
link
(transformer-circuits.pub)
Book Review: The Righteous Mind
ErnestScribbler
Jul 4, 2022, 5:45 PM
34
points
8
comments
35
min read
LW
link
My Most Likely Reason to Die Young is AI X-Risk
AISafetyIsNotLongtermist
Jul 4, 2022, 5:08 PM
61
points
24
comments
4
min read
LW
link
(forum.effectivealtruism.org)
Is General Intelligence “Compact”?
DragonGod
Jul 4, 2022, 1:27 PM
27
points
6
comments
22
min read
LW
link
Remaking EfficientZero (as best I can)
Hoagy
Jul 4, 2022, 11:03 AM
36
points
9
comments
22
min read
LW
link
We Need a Consolidated List of Bad AI Alignment Solutions
Double
Jul 4, 2022, 6:54 AM
9
points
14
comments
1
min read
LW
link
AI Forecasting: One Year In
jsteinhardt
Jul 4, 2022, 5:10 AM
132
points
12
comments
6
min read
LW
link
(bounded-regret.ghost.io)
A compressed take on recent disagreements
kman
Jul 4, 2022, 4:39 AM
33
points
9
comments
1
min read
LW
link
New US Senate Bill on X-Risk Mitigation [Linkpost]
Evan R. Murphy
Jul 4, 2022, 1:25 AM
35
points
12
comments
1
min read
LW
link
(www.hsgac.senate.gov)
Monthly Shorts 6/22
Celer
Jul 3, 2022, 11:40 PM
5
points
2
comments
5
min read
LW
link
(keller.substack.com)
Decision theory and dynamic inconsistency
paulfchristiano
Jul 3, 2022, 10:20 PM
80
points
33
comments
10
min read
LW
link
(sideways-view.com)
Five routes of access to scientific literature
DirectedEvolution
Jul 3, 2022, 8:53 PM
13
points
4
comments
6
min read
LW
link
Toni Kurz and the Insanity of Climbing Mountains
GeneSmith
Jul 3, 2022, 8:51 PM
271
points
67
comments
11
min read
LW
link
2
reviews
Wonder and The Golden AI Rule
JeffreyK
Jul 3, 2022, 6:21 PM
0
points
4
comments
6
min read
LW
link
Nature abhors an immutable replicator… usually
MSRayne
Jul 3, 2022, 3:08 PM
28
points
10
comments
3
min read
LW
link
Post hoc justifications as Compression Algorithm
Johannes C. Mayer
Jul 3, 2022, 5:02 AM
8
points
0
comments
1
min read
LW
link
SOMA—A story about Consciousness
Johannes C. Mayer
Jul 3, 2022, 4:46 AM
10
points
0
comments
1
min read
LW
link
(www.youtube.com)
Sexual self-acceptance
Johannes C. Mayer
Jul 3, 2022, 4:26 AM
11
points
6
comments
1
min read
LW
link
Donohue, Levitt, Roe, and Wade: T-minus 20 years to a massive crime wave?
Paul Logan
Jul 3, 2022, 3:03 AM
−24
points
6
comments
3
min read
LW
link
(laulpogan.substack.com)
Can we achieve AGI Alignment by balancing multiple human objectives?
Ben Smith
Jul 3, 2022, 2:51 AM
11
points
1
comment
4
min read
LW
link
Trigger-Action Planning
CFAR!Duncan
Jul 3, 2022, 1:42 AM
90
points
14
comments
13
min read
LW
link
2
reviews
[Question]
Which one of these two academic routes should I take to end up in AI Safety?
Martín Soto
Jul 3, 2022, 1:05 AM
5
points
2
comments
1
min read
LW
link
Naive Hypotheses on AI Alignment
Shoshannah Tekofsky
Jul 2, 2022, 7:03 PM
98
points
29
comments
5
min read
LW
link
The Tree of Life: Stanford AI Alignment Theory of Change
Gabe M
Jul 2, 2022, 6:36 PM
25
points
0
comments
14
min read
LW
link
Follow along with Columbia EA’s Advanced AI Safety Fellowship!
RohanS
Jul 2, 2022, 5:45 PM
3
points
0
comments
2
min read
LW
link
(forum.effectivealtruism.org)
Welcome to Analogia! (Chapter 7)
Justin Bullock
Jul 2, 2022, 5:04 PM
5
points
0
comments
11
min read
LW
link
[Question]
What about transhumans and beyond?
AlignmentMirror
Jul 2, 2022, 1:58 PM
7
points
6
comments
1
min read
LW
link
Goal-directedness: tackling complexity
Morgan_Rogers
Jul 2, 2022, 1:51 PM
8
points
0
comments
38
min read
LW
link
Literature recommendations July 2022
ChristianKl
Jul 2, 2022, 9:14 AM
17
points
9
comments
1
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel