Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
Interpretability isn’t Free
Joel Burget
Aug 4, 2022, 3:02 PM
12
points
1
comment
2
min read
LW
link
Covid 8/4/22: Rebound
Zvi
Aug 4, 2022, 11:20 AM
36
points
0
comments
11
min read
LW
link
(thezvi.wordpress.com)
High Reliability Orgs, and AI Companies
Raemon
Aug 4, 2022, 5:45 AM
86
points
7
comments
12
min read
LW
link
1
review
Surprised by ELK report’s counterexample to Debate, IDA
Evan R. Murphy
Aug 4, 2022, 2:12 AM
18
points
0
comments
5
min read
LW
link
Clapping Lower
jefftk
Aug 4, 2022, 2:10 AM
38
points
7
comments
1
min read
LW
link
(www.jefftk.com)
[Question]
How do I know if my first post should be a post, or a question?
Nathan1123
Aug 4, 2022, 1:46 AM
3
points
4
comments
1
min read
LW
link
Three pillars for avoiding AGI catastrophe: Technical alignment, deployment decisions, and coordination
LintzA
Aug 3, 2022, 11:15 PM
24
points
0
comments
11
min read
LW
link
Precursor checking for deceptive alignment
evhub
Aug 3, 2022, 10:56 PM
24
points
0
comments
14
min read
LW
link
Transformer language models are doing something more general
Numendil
Aug 3, 2022, 9:13 PM
53
points
6
comments
2
min read
LW
link
[Question]
Some doubts about Non Superintelligent AIs
aditya malik
Aug 3, 2022, 7:55 PM
0
points
4
comments
1
min read
LW
link
Announcing Squiggle: Early Access
ozziegooen
Aug 3, 2022, 7:48 PM
51
points
7
comments
7
min read
LW
link
(forum.effectivealtruism.org)
Survey: What (de)motivates you about AI risk?
Daniel_Friedrich
Aug 3, 2022, 7:17 PM
1
point
0
comments
1
min read
LW
link
(forms.gle)
Externalized reasoning oversight: a research direction for language model alignment
tamera
Aug 3, 2022, 12:03 PM
136
points
23
comments
6
min read
LW
link
Open & Welcome Thread—Aug/Sep 2022
Thomas
Aug 3, 2022, 10:22 AM
9
points
32
comments
1
min read
LW
link
[Question]
How does one recognize information and differentiate it from noise?
M. Y. Zuo
Aug 3, 2022, 3:57 AM
4
points
29
comments
1
min read
LW
link
Law-Following AI 4: Don’t Rely on Vicarious Liability
Cullen
Aug 2, 2022, 11:26 PM
5
points
2
comments
3
min read
LW
link
Two-year update on my personal AI timelines
Ajeya Cotra
Aug 2, 2022, 11:07 PM
293
points
60
comments
16
min read
LW
link
What are the Red Flags for Neural Network Suffering? - Seeds of Science call for reviewers
rogersbacon
Aug 2, 2022, 10:37 PM
24
points
6
comments
1
min read
LW
link
Againstness
CFAR!Duncan
Aug 2, 2022, 7:29 PM
50
points
8
comments
9
min read
LW
link
(Summary) Sequence Highlights—Thinking Better on Purpose
qazzquimby
Aug 2, 2022, 5:45 PM
33
points
3
comments
11
min read
LW
link
Progress links and tweets, 2022-08-02
jasoncrawford
Aug 2, 2022, 5:03 PM
9
points
0
comments
1
min read
LW
link
(rootsofprogress.org)
[Question]
I want to donate some money (not much, just what I can afford) to AGI Alignment research, to whatever organization has the best chance of making sure that AGI goes well and doesn’t kill us all. What are my best options, where can I make the most difference per dollar?
lumenwrites
Aug 2, 2022, 12:08 PM
15
points
9
comments
1
min read
LW
link
Thinking without priors?
Q Home
Aug 2, 2022, 9:17 AM
7
points
0
comments
9
min read
LW
link
[Question]
Would quantum immortality mean subjective immortality?
n0ah
Aug 2, 2022, 4:54 AM
2
points
10
comments
1
min read
LW
link
Turbocharging
CFAR!Duncan
Aug 2, 2022, 12:01 AM
52
points
5
comments
9
min read
LW
link
Letter from leading Soviet Academicians to party and government leaders of the Soviet Union regarding signs of decline and structural problems of the economic-political system (1970)
M. Y. Zuo
Aug 1, 2022, 10:35 PM
20
points
10
comments
16
min read
LW
link
Technical AI Alignment Study Group
Eric K
Aug 1, 2022, 6:33 PM
5
points
0
comments
1
min read
LW
link
[Question]
Is there any writing about prompt engineering for humans?
Alex Hollow
Aug 1, 2022, 12:52 PM
18
points
8
comments
1
min read
LW
link
Meditation course claims 65% enlightenment rate: my review
KatWoods
Aug 1, 2022, 11:25 AM
111
points
35
comments
14
min read
LW
link
[Question]
Which intro-to-AI-risk text would you recommend to...
Sherrinford
Aug 1, 2022, 9:36 AM
12
points
1
comment
1
min read
LW
link
Polaris, Five-Second Versions, and Thought Lengths
CFAR!Duncan
Aug 1, 2022, 7:14 AM
50
points
12
comments
8
min read
LW
link
A Word is Worth 1,000 Pictures
Kully
Aug 1, 2022, 4:08 AM
1
point
0
comments
2
min read
LW
link
On akrasia: starting at the bottom
seecrow
Aug 1, 2022, 4:08 AM
37
points
2
comments
3
min read
LW
link
[Question]
How likely do you think worse-than-extinction type fates to be?
span1
Aug 1, 2022, 4:08 AM
3
points
3
comments
1
min read
LW
link
Abstraction sacrifices causal clarity
Marv K
Jul 31, 2022, 7:24 PM
2
points
0
comments
3
min read
LW
link
Time-logging programs and/or spreadsheets (2022)
mikbp
Jul 31, 2022, 6:18 PM
3
points
3
comments
1
min read
LW
link
Conservatism is a rational response to epistemic uncertainty
contrarianbrit
Jul 31, 2022, 6:04 PM
2
points
11
comments
9
min read
LW
link
(thomasprosser.substack.com)
South Bay ACX/LW Meetup
IS
Jul 31, 2022, 3:30 PM
2
points
0
comments
1
min read
LW
link
Perverse Independence Incentives
jefftk
Jul 31, 2022, 2:40 PM
61
points
3
comments
1
min read
LW
link
(www.jefftk.com)
Wolfram Research v Cook
Kenny
Jul 31, 2022, 1:35 PM
7
points
3
comments
8
min read
LW
link
Wanted: Notation for credal resilience
PeterH
Jul 31, 2022, 7:35 AM
21
points
12
comments
1
min read
LW
link
Anatomy of a Dating Document
squidious
Jul 31, 2022, 2:40 AM
29
points
24
comments
4
min read
LW
link
(opalsandbonobos.blogspot.com)
chinchilla’s wild implications
nostalgebraist
Jul 31, 2022, 1:18 AM
424
points
128
comments
10
min read
LW
link
1
review
AGI-level reasoner will appear sooner than an agent; what the humanity will do with this reasoner is critical
Roman Leventov
30 Jul 2022 20:56 UTC
24
points
10
comments
1
min read
LW
link
[Question]
What job should I do?
Tom Paine
30 Jul 2022 9:15 UTC
2
points
8
comments
1
min read
LW
link
How transparency changed over time
ViktoriaMalyasova
30 Jul 2022 4:36 UTC
21
points
0
comments
6
min read
LW
link
Translating between Latent Spaces
JamesH
,
Jeremy Gillen
and
NickyP
30 Jul 2022 3:25 UTC
27
points
2
comments
8
min read
LW
link
Drexler’s Nanotech Forecast
PeterMcCluskey
30 Jul 2022 0:45 UTC
25
points
28
comments
3
min read
LW
link
(www.bayesianinvestor.com)
Humans Reflecting on HRH
leogao
29 Jul 2022 21:56 UTC
27
points
4
comments
2
min read
LW
link
Comparing Four Approaches to Inner Alignment
Lucas Teixeira
29 Jul 2022 21:06 UTC
38
points
1
comment
9
min read
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel