Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
Page
1
Automated Nomic Game 2
jefftk
5 Feb 2019 22:11 UTC
19
points
2
comments
2
min read
LW
link
Should we bait criminals using clones ?
Aël Chappuit
5 Feb 2019 21:13 UTC
−23
points
3
comments
1
min read
LW
link
Describing things: parsimony, fruitfulness, and adaptability
Mary Chernyshenko
5 Feb 2019 20:59 UTC
1
point
0
comments
1
min read
LW
link
Philosophy as low-energy approximation
Charlie Steiner
5 Feb 2019 19:34 UTC
40
points
20
comments
3
min read
LW
link
When to use quantilization
RyanCarey
5 Feb 2019 17:17 UTC
65
points
5
comments
4
min read
LW
link
(notes on) Policy Desiderata for Superintelligent AI: A Vector Field Approach
Ben Pace
4 Feb 2019 22:08 UTC
43
points
5
comments
7
min read
LW
link
SSC Paris Meetup, 09/02/18
fbreton
4 Feb 2019 19:54 UTC
1
point
0
comments
1
min read
LW
link
January 2019 gwern.net newsletter
gwern
4 Feb 2019 15:53 UTC
15
points
0
comments
1
min read
LW
link
(www.gwern.net)
My atheism story
Pausecafe
4 Feb 2019 14:33 UTC
26
points
3
comments
7
min read
LW
link
(Why) Does the Basilisk Argument fail?
Lookingforyourlogic
3 Feb 2019 23:50 UTC
0
points
11
comments
2
min read
LW
link
Constructing Goodhart
johnswentworth
3 Feb 2019 21:59 UTC
29
points
10
comments
3
min read
LW
link
Conclusion to the sequence on value learning
Rohin Shah
3 Feb 2019 21:05 UTC
51
points
20
comments
5
min read
LW
link
AI Safety Prerequisites Course: Revamp and New Lessons
philip_b
3 Feb 2019 21:04 UTC
24
points
5
comments
1
min read
LW
link
[Question]
What are some of bizarre theories based on anthropic reasoning?
Dr. Jamchie
3 Feb 2019 18:48 UTC
21
points
13
comments
1
min read
LW
link
Rationality: What’s the point?
Hazard
3 Feb 2019 16:34 UTC
12
points
11
comments
1
min read
LW
link
Quantifying Human Suffering and “Everyday Suffering”
willfranks
3 Feb 2019 13:07 UTC
7
points
3
comments
1
min read
LW
link
[Question]
How to stay concentrated for a long period of time?
infinickel
3 Feb 2019 5:24 UTC
6
points
15
comments
1
min read
LW
link
How to notice being mind-hacked
Shmi
2 Feb 2019 23:13 UTC
18
points
22
comments
2
min read
LW
link
Depression philosophizing
aaq
2 Feb 2019 22:54 UTC
6
points
2
comments
1
min read
LW
link
LessWrong DC: Metameetup
rusalkii
2 Feb 2019 18:50 UTC
1
point
0
comments
1
min read
LW
link
SSC Atlanta Meetup
Steve French
2 Feb 2019 3:11 UTC
2
points
0
comments
1
min read
LW
link
[Question]
How does Gradient Descent Interact with Goodhart?
Scott Garrabrant
2 Feb 2019 0:14 UTC
68
points
19
comments
4
min read
LW
link
Philadelphia SSC Meetup
Majuscule
1 Feb 2019 23:51 UTC
1
point
0
comments
1
min read
LW
link
STRUCTURE: Reality and rational best practice
Hazard
1 Feb 2019 23:51 UTC
5
points
2
comments
1
min read
LW
link
An Attempt To Explain No-Self In Simple Terms
Justin Vriend
1 Feb 2019 23:50 UTC
1
point
0
comments
3
min read
LW
link
STRUCTURE: How the Social Affects your rationality
Hazard
1 Feb 2019 23:35 UTC
0
points
0
comments
1
min read
LW
link
STRUCTURE: A Crash Course in Your Brain
Hazard
1 Feb 2019 23:17 UTC
6
points
4
comments
1
min read
LW
link
February Nashville SSC Meetup
Dude McDude
1 Feb 2019 22:36 UTC
1
point
0
comments
1
min read
LW
link
[Question]
What kind of information would serve as the best evidence for resolving the debate of whether a centrist or leftist Democratic nominee is likelier to take the White House in 2020?
Evan_Gaensbauer
1 Feb 2019 18:40 UTC
10
points
10
comments
3
min read
LW
link
Urgent & important: How (not) to do your to-do list
bfinn
1 Feb 2019 17:44 UTC
51
points
20
comments
13
min read
LW
link
Who wants to be a Millionaire?
Bucky
1 Feb 2019 14:02 UTC
29
points
1
comment
11
min read
LW
link
What is Wrong?
Inyuki
1 Feb 2019 12:02 UTC
1
point
2
comments
2
min read
LW
link
Drexler on AI Risk
PeterMcCluskey
1 Feb 2019 5:11 UTC
35
points
10
comments
9
min read
LW
link
(www.bayesianinvestor.com)
Boundaries—A map and territory experiment. [post-rationality]
Elo
1 Feb 2019 2:08 UTC
−18
points
14
comments
2
min read
LW
link
[Question]
Why is this utilitarian calculus wrong? Or is it?
EconomicModel
31 Jan 2019 23:57 UTC
15
points
21
comments
1
min read
LW
link
Small hope for less bias and more practability
ArthurLidia
31 Jan 2019 22:09 UTC
0
points
0
comments
1
min read
LW
link
Reliability amplification
paulfchristiano
31 Jan 2019 21:12 UTC
24
points
3
comments
7
min read
LW
link
Cambridge (UK) SSC meetup
thisheavenlyconjugation
31 Jan 2019 11:45 UTC
1
point
0
comments
1
min read
LW
link
The role of epistemic vs. aleatory uncertainty in quantifying AI-Xrisk
David Scott Krueger (formerly: capybaralet)
31 Jan 2019 6:13 UTC
15
points
6
comments
2
min read
LW
link
[Question]
Applied Rationality podcast—feedback?
Bae's Theorem
31 Jan 2019 1:46 UTC
11
points
12
comments
1
min read
LW
link
Wireheading is in the eye of the beholder
Stuart_Armstrong
30 Jan 2019 18:23 UTC
26
points
10
comments
1
min read
LW
link
Masculine Virtues
Jacob Falkovich
30 Jan 2019 16:03 UTC
52
points
32
comments
13
min read
LW
link
Deconfusing Logical Counterfactuals
Chris_Leong
30 Jan 2019 15:13 UTC
27
points
16
comments
11
min read
LW
link
Book Trilogy Review: Remembrance of Earth’s Past (The Three Body Problem)
Zvi
30 Jan 2019 1:10 UTC
48
points
15
comments
40
min read
LW
link
(thezvi.wordpress.com)
Alignment Newsletter #43
Rohin Shah
29 Jan 2019 21:10 UTC
14
points
2
comments
13
min read
LW
link
(mailchi.mp)
The Question Of Perception
The Arkon
29 Jan 2019 20:59 UTC
0
points
18
comments
5
min read
LW
link
[Question]
Which textbook would you recommend to learn decision theory?
supermartingale
29 Jan 2019 20:48 UTC
27
points
6
comments
1
min read
LW
link
Towards equilibria-breaking methods
ryan_b
29 Jan 2019 16:19 UTC
22
points
3
comments
2
min read
LW
link
Can there be an indescribable hellworld?
Stuart_Armstrong
29 Jan 2019 15:00 UTC
39
points
19
comments
2
min read
LW
link
How much can value learning be disentangled?
Stuart_Armstrong
29 Jan 2019 14:17 UTC
22
points
30
comments
2
min read
LW
link
Back to top
Next