Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
The Falling Drill
Screwtape
Aug 5, 2022, 12:08 AM
46
points
3
comments
2
min read
LW
link
Convergence Towards World-Models: A Gears-Level Model
Thane Ruthenis
Aug 4, 2022, 11:31 PM
38
points
1
comment
13
min read
LW
link
Cambist Booking
Screwtape
Aug 4, 2022, 10:40 PM
20
points
3
comments
4
min read
LW
link
Calibration Trivia
Screwtape
Aug 4, 2022, 10:31 PM
12
points
9
comments
4
min read
LW
link
Monthly Shorts 7/22
Celer
Aug 4, 2022, 10:30 PM
5
points
0
comments
3
min read
LW
link
(keller.substack.com)
The Pragmascope Idea
johnswentworth
Aug 4, 2022, 9:52 PM
59
points
20
comments
3
min read
LW
link
Running a Basic Meetup
Screwtape
Aug 4, 2022, 9:49 PM
21
points
1
comment
2
min read
LW
link
Fiber arts, mysterious dodecahedrons, and waiting on “Eureka!”
eukaryote
Aug 4, 2022, 8:37 PM
124
points
15
comments
9
min read
LW
link
1
review
(eukaryotewritesblog.com)
[Question]
Would “Manhattan Project” style be beneficial or deleterious for AI Alignment?
Valentin2026
Aug 4, 2022, 7:12 PM
5
points
1
comment
1
min read
LW
link
[Question]
AI alignment: Would a lazy self-preservation instinct be sufficient?
BrainFrog
Aug 4, 2022, 5:53 PM
−1
points
4
comments
1
min read
LW
link
Socratic Ducking, OODA Loops, Frame-by-Frame Debugging
CFAR!Duncan
Aug 4, 2022, 5:44 PM
26
points
1
comment
5
min read
LW
link
What do ML researchers think about AI in 2022?
KatjaGrace
Aug 4, 2022, 3:40 PM
221
points
33
comments
3
min read
LW
link
(aiimpacts.org)
Interpretability isn’t Free
Joel Burget
Aug 4, 2022, 3:02 PM
12
points
1
comment
2
min read
LW
link
Covid 8/4/22: Rebound
Zvi
Aug 4, 2022, 11:20 AM
36
points
0
comments
11
min read
LW
link
(thezvi.wordpress.com)
High Reliability Orgs, and AI Companies
Raemon
Aug 4, 2022, 5:45 AM
86
points
7
comments
12
min read
LW
link
1
review
Surprised by ELK report’s counterexample to Debate, IDA
Evan R. Murphy
Aug 4, 2022, 2:12 AM
18
points
0
comments
5
min read
LW
link
Clapping Lower
jefftk
Aug 4, 2022, 2:10 AM
38
points
7
comments
1
min read
LW
link
(www.jefftk.com)
[Question]
How do I know if my first post should be a post, or a question?
Nathan1123
Aug 4, 2022, 1:46 AM
3
points
4
comments
1
min read
LW
link
Three pillars for avoiding AGI catastrophe: Technical alignment, deployment decisions, and coordination
LintzA
Aug 3, 2022, 11:15 PM
24
points
0
comments
11
min read
LW
link
Precursor checking for deceptive alignment
evhub
Aug 3, 2022, 10:56 PM
24
points
0
comments
14
min read
LW
link
Transformer language models are doing something more general
Numendil
Aug 3, 2022, 9:13 PM
53
points
6
comments
2
min read
LW
link
[Question]
Some doubts about Non Superintelligent AIs
aditya malik
Aug 3, 2022, 7:55 PM
0
points
4
comments
1
min read
LW
link
Announcing Squiggle: Early Access
ozziegooen
Aug 3, 2022, 7:48 PM
51
points
7
comments
7
min read
LW
link
(forum.effectivealtruism.org)
Survey: What (de)motivates you about AI risk?
Daniel_Friedrich
Aug 3, 2022, 7:17 PM
1
point
0
comments
1
min read
LW
link
(forms.gle)
Externalized reasoning oversight: a research direction for language model alignment
tamera
Aug 3, 2022, 12:03 PM
136
points
23
comments
6
min read
LW
link
Open & Welcome Thread—Aug/Sep 2022
Thomas
Aug 3, 2022, 10:22 AM
9
points
32
comments
1
min read
LW
link
[Question]
How does one recognize information and differentiate it from noise?
M. Y. Zuo
Aug 3, 2022, 3:57 AM
4
points
29
comments
1
min read
LW
link
Law-Following AI 4: Don’t Rely on Vicarious Liability
Cullen
Aug 2, 2022, 11:26 PM
5
points
2
comments
3
min read
LW
link
Two-year update on my personal AI timelines
Ajeya Cotra
Aug 2, 2022, 11:07 PM
293
points
60
comments
16
min read
LW
link
What are the Red Flags for Neural Network Suffering? - Seeds of Science call for reviewers
rogersbacon
Aug 2, 2022, 10:37 PM
24
points
6
comments
1
min read
LW
link
Againstness
CFAR!Duncan
Aug 2, 2022, 7:29 PM
50
points
8
comments
9
min read
LW
link
(Summary) Sequence Highlights—Thinking Better on Purpose
qazzquimby
Aug 2, 2022, 5:45 PM
33
points
3
comments
11
min read
LW
link
Progress links and tweets, 2022-08-02
jasoncrawford
Aug 2, 2022, 5:03 PM
9
points
0
comments
1
min read
LW
link
(rootsofprogress.org)
[Question]
I want to donate some money (not much, just what I can afford) to AGI Alignment research, to whatever organization has the best chance of making sure that AGI goes well and doesn’t kill us all. What are my best options, where can I make the most difference per dollar?
lumenwrites
Aug 2, 2022, 12:08 PM
15
points
9
comments
1
min read
LW
link
Thinking without priors?
Q Home
Aug 2, 2022, 9:17 AM
7
points
0
comments
9
min read
LW
link
[Question]
Would quantum immortality mean subjective immortality?
n0ah
Aug 2, 2022, 4:54 AM
2
points
10
comments
1
min read
LW
link
Turbocharging
CFAR!Duncan
2 Aug 2022 0:01 UTC
52
points
5
comments
9
min read
LW
link
Letter from leading Soviet Academicians to party and government leaders of the Soviet Union regarding signs of decline and structural problems of the economic-political system (1970)
M. Y. Zuo
1 Aug 2022 22:35 UTC
20
points
10
comments
16
min read
LW
link
Technical AI Alignment Study Group
Eric K
1 Aug 2022 18:33 UTC
5
points
0
comments
1
min read
LW
link
[Question]
Is there any writing about prompt engineering for humans?
Alex Hollow
1 Aug 2022 12:52 UTC
18
points
8
comments
1
min read
LW
link
Meditation course claims 65% enlightenment rate: my review
KatWoods
1 Aug 2022 11:25 UTC
111
points
35
comments
14
min read
LW
link
[Question]
Which intro-to-AI-risk text would you recommend to...
Sherrinford
1 Aug 2022 9:36 UTC
12
points
1
comment
1
min read
LW
link
Polaris, Five-Second Versions, and Thought Lengths
CFAR!Duncan
1 Aug 2022 7:14 UTC
50
points
12
comments
8
min read
LW
link
A Word is Worth 1,000 Pictures
Kully
1 Aug 2022 4:08 UTC
1
point
0
comments
2
min read
LW
link
On akrasia: starting at the bottom
seecrow
1 Aug 2022 4:08 UTC
37
points
2
comments
3
min read
LW
link
[Question]
How likely do you think worse-than-extinction type fates to be?
span1
1 Aug 2022 4:08 UTC
3
points
3
comments
1
min read
LW
link
Abstraction sacrifices causal clarity
Marv K
31 Jul 2022 19:24 UTC
2
points
0
comments
3
min read
LW
link
Time-logging programs and/or spreadsheets (2022)
mikbp
31 Jul 2022 18:18 UTC
3
points
3
comments
1
min read
LW
link
Conservatism is a rational response to epistemic uncertainty
contrarianbrit
31 Jul 2022 18:04 UTC
2
points
11
comments
9
min read
LW
link
(thomasprosser.substack.com)
South Bay ACX/LW Meetup
IS
31 Jul 2022 15:30 UTC
2
points
0
comments
1
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel