Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
Metaculus and medians
rossry
Aug 6, 2022, 3:34 AM
18
points
4
comments
4
min read
LW
link
Announcing the Introduction to ML Safety course
Dan H
,
TW123
and
ozhang
Aug 6, 2022, 2:46 AM
73
points
6
comments
7
min read
LW
link
«Boundaries», Part 2: trends in EA’s handling of boundaries
Andrew_Critch
Aug 6, 2022, 12:42 AM
81
points
15
comments
7
min read
LW
link
“Just hiring people” is sometimes still actually possible
lc
Aug 5, 2022, 9:44 PM
38
points
11
comments
5
min read
LW
link
The need for certainty
Thomas McMurtry
Aug 5, 2022, 8:18 PM
2
points
0
comments
4
min read
LW
link
Rant on Problem Factorization for Alignment
johnswentworth
Aug 5, 2022, 7:23 PM
104
points
53
comments
6
min read
LW
link
Counterfactuals are Confusing because of an Ontological Shift
Chris_Leong
Aug 5, 2022, 7:03 PM
17
points
35
comments
2
min read
LW
link
Orange county ACX/Less-Wrong discussion group and hang-out. (orange county)
Michael Michalchik
Aug 5, 2022, 6:25 PM
2
points
0
comments
1
min read
LW
link
Gears-Level Understanding, Deliberate Performance, The Strategic Level
CFAR!Duncan
Aug 5, 2022, 5:11 PM
30
points
3
comments
5
min read
LW
link
[Question]
COVID-19 Group Testing Post-mortem?
gwern
Aug 5, 2022, 4:32 PM
72
points
6
comments
2
min read
LW
link
Where are the red lines for AI?
Karl von Wendt
Aug 5, 2022, 9:34 AM
26
points
10
comments
6
min read
LW
link
Bridging Expected Utility Maximization and Optimization
Daniel Herrmann
Aug 5, 2022, 8:18 AM
25
points
5
comments
14
min read
LW
link
Deontology and Tool AI
Nathan1123
Aug 5, 2022, 5:20 AM
4
points
5
comments
6
min read
LW
link
An attempt to understand the Complexity of Values
Dalton Mabery
Aug 5, 2022, 4:43 AM
3
points
0
comments
5
min read
LW
link
$20K In Bounties for AI Safety Public Materials
Dan H
,
TW123
and
ozhang
Aug 5, 2022, 2:52 AM
71
points
9
comments
6
min read
LW
link
Two Kids Crosswise
jefftk
Aug 5, 2022, 2:40 AM
16
points
3
comments
1
min read
LW
link
(www.jefftk.com)
The Falling Drill
Screwtape
Aug 5, 2022, 12:08 AM
46
points
3
comments
2
min read
LW
link
Convergence Towards World-Models: A Gears-Level Model
Thane Ruthenis
Aug 4, 2022, 11:31 PM
38
points
1
comment
13
min read
LW
link
Cambist Booking
Screwtape
Aug 4, 2022, 10:40 PM
20
points
3
comments
4
min read
LW
link
Calibration Trivia
Screwtape
Aug 4, 2022, 10:31 PM
12
points
9
comments
4
min read
LW
link
Monthly Shorts 7/22
Celer
Aug 4, 2022, 10:30 PM
5
points
0
comments
3
min read
LW
link
(keller.substack.com)
The Pragmascope Idea
johnswentworth
Aug 4, 2022, 9:52 PM
59
points
20
comments
3
min read
LW
link
Running a Basic Meetup
Screwtape
Aug 4, 2022, 9:49 PM
21
points
1
comment
2
min read
LW
link
Fiber arts, mysterious dodecahedrons, and waiting on “Eureka!”
eukaryote
Aug 4, 2022, 8:37 PM
124
points
15
comments
9
min read
LW
link
1
review
(eukaryotewritesblog.com)
[Question]
Would “Manhattan Project” style be beneficial or deleterious for AI Alignment?
Valentin2026
Aug 4, 2022, 7:12 PM
5
points
1
comment
1
min read
LW
link
[Question]
AI alignment: Would a lazy self-preservation instinct be sufficient?
BrainFrog
Aug 4, 2022, 5:53 PM
−1
points
4
comments
1
min read
LW
link
Socratic Ducking, OODA Loops, Frame-by-Frame Debugging
CFAR!Duncan
Aug 4, 2022, 5:44 PM
26
points
1
comment
5
min read
LW
link
What do ML researchers think about AI in 2022?
KatjaGrace
Aug 4, 2022, 3:40 PM
221
points
33
comments
3
min read
LW
link
(aiimpacts.org)
Interpretability isn’t Free
Joel Burget
Aug 4, 2022, 3:02 PM
12
points
1
comment
2
min read
LW
link
Covid 8/4/22: Rebound
Zvi
Aug 4, 2022, 11:20 AM
36
points
0
comments
11
min read
LW
link
(thezvi.wordpress.com)
High Reliability Orgs, and AI Companies
Raemon
Aug 4, 2022, 5:45 AM
86
points
7
comments
12
min read
LW
link
1
review
Surprised by ELK report’s counterexample to Debate, IDA
Evan R. Murphy
Aug 4, 2022, 2:12 AM
18
points
0
comments
5
min read
LW
link
Clapping Lower
jefftk
Aug 4, 2022, 2:10 AM
38
points
7
comments
1
min read
LW
link
(www.jefftk.com)
[Question]
How do I know if my first post should be a post, or a question?
Nathan1123
Aug 4, 2022, 1:46 AM
3
points
4
comments
1
min read
LW
link
Three pillars for avoiding AGI catastrophe: Technical alignment, deployment decisions, and coordination
LintzA
Aug 3, 2022, 11:15 PM
24
points
0
comments
11
min read
LW
link
Precursor checking for deceptive alignment
evhub
Aug 3, 2022, 10:56 PM
24
points
0
comments
14
min read
LW
link
Transformer language models are doing something more general
Numendil
Aug 3, 2022, 9:13 PM
53
points
6
comments
2
min read
LW
link
[Question]
Some doubts about Non Superintelligent AIs
aditya malik
Aug 3, 2022, 7:55 PM
0
points
4
comments
1
min read
LW
link
Announcing Squiggle: Early Access
ozziegooen
Aug 3, 2022, 7:48 PM
51
points
7
comments
7
min read
LW
link
(forum.effectivealtruism.org)
Survey: What (de)motivates you about AI risk?
Daniel_Friedrich
Aug 3, 2022, 7:17 PM
1
point
0
comments
1
min read
LW
link
(forms.gle)
Externalized reasoning oversight: a research direction for language model alignment
tamera
Aug 3, 2022, 12:03 PM
136
points
23
comments
6
min read
LW
link
Open & Welcome Thread—Aug/Sep 2022
Thomas
Aug 3, 2022, 10:22 AM
9
points
32
comments
1
min read
LW
link
[Question]
How does one recognize information and differentiate it from noise?
M. Y. Zuo
Aug 3, 2022, 3:57 AM
4
points
29
comments
1
min read
LW
link
Law-Following AI 4: Don’t Rely on Vicarious Liability
Cullen
Aug 2, 2022, 11:26 PM
5
points
2
comments
3
min read
LW
link
Two-year update on my personal AI timelines
Ajeya Cotra
2 Aug 2022 23:07 UTC
293
points
60
comments
16
min read
LW
link
What are the Red Flags for Neural Network Suffering? - Seeds of Science call for reviewers
rogersbacon
2 Aug 2022 22:37 UTC
24
points
6
comments
1
min read
LW
link
Againstness
CFAR!Duncan
2 Aug 2022 19:29 UTC
50
points
8
comments
9
min read
LW
link
(Summary) Sequence Highlights—Thinking Better on Purpose
qazzquimby
2 Aug 2022 17:45 UTC
33
points
3
comments
11
min read
LW
link
Progress links and tweets, 2022-08-02
jasoncrawford
2 Aug 2022 17:03 UTC
9
points
0
comments
1
min read
LW
link
(rootsofprogress.org)
[Question]
I want to donate some money (not much, just what I can afford) to AGI Alignment research, to whatever organization has the best chance of making sure that AGI goes well and doesn’t kill us all. What are my best options, where can I make the most difference per dollar?
lumenwrites
2 Aug 2022 12:08 UTC
15
points
9
comments
1
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel