Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
Anthropic’s SoLU (Softmax Linear Unit)
Joel Burget
Jul 4, 2022, 6:38 PM
21
points
1
comment
4
min read
LW
link
(transformer-circuits.pub)
Book Review: The Righteous Mind
ErnestScribbler
Jul 4, 2022, 5:45 PM
34
points
8
comments
35
min read
LW
link
My Most Likely Reason to Die Young is AI X-Risk
AISafetyIsNotLongtermist
Jul 4, 2022, 5:08 PM
61
points
24
comments
4
min read
LW
link
(forum.effectivealtruism.org)
Is General Intelligence “Compact”?
DragonGod
Jul 4, 2022, 1:27 PM
27
points
6
comments
22
min read
LW
link
Remaking EfficientZero (as best I can)
Hoagy
Jul 4, 2022, 11:03 AM
36
points
9
comments
22
min read
LW
link
We Need a Consolidated List of Bad AI Alignment Solutions
Double
Jul 4, 2022, 6:54 AM
9
points
14
comments
1
min read
LW
link
AI Forecasting: One Year In
jsteinhardt
Jul 4, 2022, 5:10 AM
132
points
12
comments
6
min read
LW
link
(bounded-regret.ghost.io)
A compressed take on recent disagreements
kman
Jul 4, 2022, 4:39 AM
33
points
9
comments
1
min read
LW
link
New US Senate Bill on X-Risk Mitigation [Linkpost]
Evan R. Murphy
Jul 4, 2022, 1:25 AM
35
points
12
comments
1
min read
LW
link
(www.hsgac.senate.gov)
Monthly Shorts 6/22
Celer
Jul 3, 2022, 11:40 PM
5
points
2
comments
5
min read
LW
link
(keller.substack.com)
Decision theory and dynamic inconsistency
paulfchristiano
Jul 3, 2022, 10:20 PM
80
points
33
comments
10
min read
LW
link
(sideways-view.com)
Five routes of access to scientific literature
DirectedEvolution
Jul 3, 2022, 8:53 PM
13
points
4
comments
6
min read
LW
link
Toni Kurz and the Insanity of Climbing Mountains
GeneSmith
Jul 3, 2022, 8:51 PM
271
points
67
comments
11
min read
LW
link
2
reviews
Wonder and The Golden AI Rule
JeffreyK
Jul 3, 2022, 6:21 PM
0
points
4
comments
6
min read
LW
link
Nature abhors an immutable replicator… usually
MSRayne
Jul 3, 2022, 3:08 PM
28
points
10
comments
3
min read
LW
link
Post hoc justifications as Compression Algorithm
Johannes C. Mayer
Jul 3, 2022, 5:02 AM
8
points
0
comments
1
min read
LW
link
SOMA—A story about Consciousness
Johannes C. Mayer
Jul 3, 2022, 4:46 AM
10
points
0
comments
1
min read
LW
link
(www.youtube.com)
Sexual self-acceptance
Johannes C. Mayer
Jul 3, 2022, 4:26 AM
11
points
6
comments
1
min read
LW
link
Donohue, Levitt, Roe, and Wade: T-minus 20 years to a massive crime wave?
Paul Logan
Jul 3, 2022, 3:03 AM
−24
points
6
comments
3
min read
LW
link
(laulpogan.substack.com)
Can we achieve AGI Alignment by balancing multiple human objectives?
Ben Smith
Jul 3, 2022, 2:51 AM
11
points
1
comment
4
min read
LW
link
Trigger-Action Planning
CFAR!Duncan
Jul 3, 2022, 1:42 AM
90
points
14
comments
13
min read
LW
link
2
reviews
[Question]
Which one of these two academic routes should I take to end up in AI Safety?
Martín Soto
Jul 3, 2022, 1:05 AM
5
points
2
comments
1
min read
LW
link
Naive Hypotheses on AI Alignment
Shoshannah Tekofsky
Jul 2, 2022, 7:03 PM
98
points
29
comments
5
min read
LW
link
The Tree of Life: Stanford AI Alignment Theory of Change
Gabe M
Jul 2, 2022, 6:36 PM
25
points
0
comments
14
min read
LW
link
Follow along with Columbia EA’s Advanced AI Safety Fellowship!
RohanS
Jul 2, 2022, 5:45 PM
3
points
0
comments
2
min read
LW
link
(forum.effectivealtruism.org)
Welcome to Analogia! (Chapter 7)
Justin Bullock
Jul 2, 2022, 5:04 PM
5
points
0
comments
11
min read
LW
link
[Question]
What about transhumans and beyond?
AlignmentMirror
Jul 2, 2022, 1:58 PM
7
points
6
comments
1
min read
LW
link
Goal-directedness: tackling complexity
Morgan_Rogers
Jul 2, 2022, 1:51 PM
8
points
0
comments
38
min read
LW
link
Literature recommendations July 2022
ChristianKl
Jul 2, 2022, 9:14 AM
17
points
9
comments
1
min read
LW
link
Deontological Evil
lsusr
Jul 2, 2022, 6:57 AM
45
points
4
comments
2
min read
LW
link
Could an AI Alignment Sandbox be useful?
Michael Soareverix
Jul 2, 2022, 5:06 AM
2
points
1
comment
1
min read
LW
link
Five views of Bayes’ Theorem
Adam Scherlis
Jul 2, 2022, 2:25 AM
38
points
4
comments
1
min read
LW
link
[Linkpost] Existential Risk Analysis in Empirical Research Papers
Dan H
Jul 2, 2022, 12:09 AM
40
points
0
comments
1
min read
LW
link
(arxiv.org)
Agenty AGI – How Tempting?
PeterMcCluskey
Jul 1, 2022, 11:40 PM
22
points
3
comments
5
min read
LW
link
(www.bayesianinvestor.com)
AXRP Episode 16 - Preparing for Debate AI with Geoffrey Irving
DanielFilan
Jul 1, 2022, 10:20 PM
20
points
0
comments
37
min read
LW
link
[Question]
Examples of practical implications of Judea Pearl’s Causality work
ChristianKl
Jul 1, 2022, 8:58 PM
23
points
6
comments
1
min read
LW
link
Minerva
Algon
Jul 1, 2022, 8:06 PM
36
points
6
comments
2
min read
LW
link
(ai.googleblog.com)
Disarming status
sano
Jul 1, 2022, 8:00 PM
−4
points
1
comment
6
min read
LW
link
Paper: Forecasting world events with neural nets
Owain_Evans
,
Dan H
and
Joe Kwon
Jul 1, 2022, 7:40 PM
39
points
3
comments
4
min read
LW
link
Reframing the AI Risk
Thane Ruthenis
Jul 1, 2022, 6:44 PM
26
points
7
comments
6
min read
LW
link
Who is this MSRayne person anyway?
MSRayne
Jul 1, 2022, 5:32 PM
32
points
30
comments
11
min read
LW
link
Limerence Messes Up Your Rationality Real Bad, Yo
Raemon
Jul 1, 2022, 4:53 PM
128
points
41
comments
3
min read
LW
link
2
reviews
[Link] On the paradox of tolerance in relation to fascism and online content moderation – Unstable Ontology
Kenny
Jul 1, 2022, 4:43 PM
5
points
0
comments
1
min read
LW
link
Trends in GPU price-performance
Marius Hobbhahn
and
Tamay
Jul 1, 2022, 3:51 PM
85
points
13
comments
1
min read
LW
link
1
review
(epochai.org)
[Question]
How to deal with non-schedulable one-off stimulus-response-pair-like situations when planning/organising projects?
mikbp
Jul 1, 2022, 3:22 PM
2
points
3
comments
1
min read
LW
link
What Is The True Name of Modularity?
CallumMcDougall
,
Lucius Bushnaq
and
Avery
Jul 1, 2022, 2:55 PM
39
points
10
comments
12
min read
LW
link
Defining Optimization in a Deeper Way Part 1
J Bostock
Jul 1, 2022, 2:03 PM
7
points
0
comments
2
min read
LW
link
Safetywashing
Adam Scholl
Jul 1, 2022, 11:56 AM
261
points
20
comments
1
min read
LW
link
2
reviews
[Question]
AGI alignment with what?
AlignmentMirror
Jul 1, 2022, 10:22 AM
6
points
10
comments
1
min read
LW
link
Open & Welcome Thread—July 2022
Kaj_Sotala
Jul 1, 2022, 7:47 AM
20
points
61
comments
1
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel