Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
1
Clarifying the confusion around inner alignment
Rauno Arike
May 13, 2022, 11:05 PM
31
points
0
comments
11
min read
LW
link
Costs and benefits of amniocentesis for normal pregnancies
braces
May 13, 2022, 10:47 PM
13
points
4
comments
3
min read
LW
link
Frame for Take-Off Speeds to inform compute governance & scaling alignment
Logan Riggs
May 13, 2022, 10:23 PM
15
points
2
comments
2
min read
LW
link
Alignment as Constraints
Logan Riggs
May 13, 2022, 10:07 PM
10
points
0
comments
2
min read
LW
link
How close to nuclear war did we get over Cuba?
NathanBarnard
May 13, 2022, 7:58 PM
13
points
0
comments
10
min read
LW
link
Against Time in Agent Models
johnswentworth
May 13, 2022, 7:55 PM
62
points
13
comments
3
min read
LW
link
Agency As a Natural Abstraction
Thane Ruthenis
May 13, 2022, 6:02 PM
55
points
9
comments
13
min read
LW
link
Fermi estimation of the impact you might have working on AI safety
Fabien Roger
May 13, 2022, 5:49 PM
10
points
0
comments
1
min read
LW
link
“Tech company singularities”, and steering them to reduce x-risk
Andrew_Critch
May 13, 2022, 5:24 PM
75
points
11
comments
4
min read
LW
link
An observation about Hubinger et al.’s framework for learned optimization
carboniferous_umbraculum
May 13, 2022, 4:20 PM
34
points
9
comments
8
min read
LW
link
[Question]
The Economics of a New Energy Source
casualphysicsenjoyer
May 13, 2022, 2:08 PM
2
points
13
comments
1
min read
LW
link
[Question]
Still possible to change username?
gabrielrecc
May 13, 2022, 1:41 PM
7
points
4
comments
1
min read
LW
link
[Rough notes, BAIS] Human values and cyclical preferences
pranomostro
,
Jayjay
and
Lucie Philippon
May 13, 2022, 1:28 PM
5
points
0
comments
4
min read
LW
link
[Question]
Can moderators fix old sequences posts?
EniScien
May 13, 2022, 12:30 PM
10
points
1
comment
1
min read
LW
link
DeepMind is hiring for the Scalable Alignment and Alignment Teams
Rohin Shah
and
Geoffrey Irving
May 13, 2022, 12:17 PM
150
points
34
comments
9
min read
LW
link
Thoughts on AI Safety Camp
Charlie Steiner
May 13, 2022, 7:16 AM
33
points
8
comments
7
min read
LW
link
Deferring
owencb
May 12, 2022, 11:56 PM
18
points
2
comments
11
min read
LW
link
RLHF
Ansh Radhakrishnan
May 12, 2022, 9:18 PM
18
points
5
comments
5
min read
LW
link
[Question]
What to do when starting a business in an imminent-AGI world?
ryan_b
May 12, 2022, 9:07 PM
25
points
7
comments
1
min read
LW
link
Interpretability’s Alignment-Solving Potential: Analysis of 7 Scenarios
Evan R. Murphy
May 12, 2022, 8:01 PM
58
points
0
comments
59
min read
LW
link
Introduction to the sequence: Interpretability Research for the Most Important Century
Evan R. Murphy
May 12, 2022, 7:59 PM
16
points
0
comments
8
min read
LW
link
A tentative dialogue with a Friendly-boxed-super-AGI on brain uploads
Ramiro P.
May 12, 2022, 7:40 PM
1
point
12
comments
4
min read
LW
link
The Last Paperclip
Logan Zoellner
May 12, 2022, 7:25 PM
63
points
15
comments
18
min read
LW
link
Deepmind’s Gato: Generalist Agent
Daniel Kokotajlo
May 12, 2022, 4:01 PM
165
points
62
comments
1
min read
LW
link
“A Generalist Agent”: New DeepMind Publication
1a3orn
May 12, 2022, 3:30 PM
79
points
43
comments
1
min read
LW
link
Covid 5/12/22: Other Priorities
Zvi
May 12, 2022, 1:30 PM
31
points
4
comments
15
min read
LW
link
(thezvi.wordpress.com)
[Question]
How would public media outlets need to be governed to cover all political views?
ChristianKl
May 12, 2022, 12:55 PM
13
points
14
comments
1
min read
LW
link
[Question]
What’s keeping concerned capabilities gain researchers from leaving the field?
sovran
May 12, 2022, 12:16 PM
19
points
4
comments
1
min read
LW
link
Positive outcomes under an unaligned AGI takeover
Yitz
May 12, 2022, 7:45 AM
19
points
10
comments
3
min read
LW
link
[Question]
What are your recommendations for technical AI alignment podcasts?
Evan_Gaensbauer
May 11, 2022, 9:52 PM
5
points
4
comments
1
min read
LW
link
Gracefully correcting uncalibrated shame
AF2022
May 11, 2022, 7:51 PM
−31
points
34
comments
4
min read
LW
link
[Intro to brain-like-AGI safety] 14. Controlled AGI
Steven Byrnes
May 11, 2022, 1:17 PM
45
points
25
comments
20
min read
LW
link
ProjectLawful.com: Eliezer’s latest story, past 1M words
Eliezer Yudkowsky
May 11, 2022, 6:18 AM
234
points
112
comments
1
min read
LW
link
4
reviews
An Inside View of AI Alignment
Ansh Radhakrishnan
May 11, 2022, 2:16 AM
32
points
2
comments
2
min read
LW
link
Fighting in various places for a really long time
KatjaGrace
May 11, 2022, 1:50 AM
36
points
12
comments
4
min read
LW
link
(worldspiritsockpuppet.com)
Stuff I might do if I had covid
KatjaGrace
May 11, 2022, 12:00 AM
39
points
9
comments
1
min read
LW
link
(worldspiritsockpuppet.com)
Crises Don’t Need Your Software
GabrielExists
May 10, 2022, 9:06 PM
59
points
18
comments
6
min read
LW
link
Ceiling Fan Air Filter
jefftk
May 10, 2022, 2:20 PM
18
points
9
comments
1
min read
LW
link
(www.jefftk.com)
The limits of AI safety via debate
Marius Hobbhahn
May 10, 2022, 1:33 PM
35
points
8
comments
10
min read
LW
link
Examining Armstrong’s category of generalized models
Morgan_Rogers
May 10, 2022, 9:07 AM
14
points
0
comments
7
min read
LW
link
Dath Ilani Rule of Law
David Udell
May 10, 2022, 6:17 AM
24
points
25
comments
4
min read
LW
link
AI safety should be made more accessible using non text-based media
Massimog
May 10, 2022, 3:14 AM
2
points
4
comments
4
min read
LW
link
LessWrong Now Has Dark Mode
jimrandomh
May 10, 2022, 1:21 AM
135
points
31
comments
1
min read
LW
link
Conditions for mathematical equivalence of Stochastic Gradient Descent and Natural Selection
Oliver Sourbut
May 9, 2022, 9:38 PM
70
points
19
comments
8
min read
LW
link
1
review
(www.oliversourbut.net)
AI Alignment YouTube Playlists
jacquesthibs
and
remember
May 9, 2022, 9:33 PM
30
points
4
comments
1
min read
LW
link
When is AI safety research harmful?
NathanBarnard
May 9, 2022, 6:19 PM
2
points
0
comments
8
min read
LW
link
A Bird’s Eye View of the ML Field [Pragmatic AI Safety #2]
Dan H
and
TW123
May 9, 2022, 5:18 PM
163
points
8
comments
35
min read
LW
link
Introduction to Pragmatic AI Safety [Pragmatic AI Safety #1]
Dan H
and
TW123
May 9, 2022, 5:06 PM
80
points
3
comments
6
min read
LW
link
Jobs: Help scale up LM alignment research at NYU
Sam Bowman
May 9, 2022, 2:12 PM
60
points
1
comment
1
min read
LW
link
Microphone on Electric Mandolin
jefftk
May 9, 2022, 2:00 PM
16
points
0
comments
1
min read
LW
link
(www.jefftk.com)
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel