Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
1
And the word was “God”
pchvykov
Aug 30, 2022, 9:13 PM
−22
points
4
comments
3
min read
LW
link
Worlds Where Iterative Design Fails
johnswentworth
Aug 30, 2022, 8:48 PM
208
points
30
comments
10
min read
LW
link
1
review
Inner Alignment via Superpowers
JamesH
,
Thomas Larsen
and
Jeremy Gillen
Aug 30, 2022, 8:01 PM
37
points
13
comments
4
min read
LW
link
ML Model Attribution Challenge [Linkpost]
aog
Aug 30, 2022, 7:34 PM
11
points
0
comments
1
min read
LW
link
(mlmac.io)
How likely is deceptive alignment?
evhub
Aug 30, 2022, 7:34 PM
104
points
28
comments
60
min read
LW
link
Built-In Bundling For Faster Loading
jefftk
Aug 30, 2022, 7:20 PM
15
points
0
comments
2
min read
LW
link
(www.jefftk.com)
[Question]
A bayesian updating on expert opinions
amarai
Aug 30, 2022, 11:56 AM
1
point
1
comment
1
min read
LW
link
Any Utilitarianism Makes Sense As Policy
George3d6
Aug 30, 2022, 9:55 AM
6
points
6
comments
7
min read
LW
link
(www.epistem.ink)
A gentle primer on caring, including in strange senses, with applications
Kaarel
Aug 30, 2022, 8:05 AM
10
points
4
comments
18
min read
LW
link
Modified Guess Culture
konstell
Aug 30, 2022, 2:30 AM
5
points
5
comments
1
min read
LW
link
(konstell.com)
[Question]
What is the best critique of AI existential risk arguments?
joshc
Aug 30, 2022, 2:18 AM
6
points
11
comments
1
min read
LW
link
How to plan for a radically uncertain future?
Kerry
Aug 30, 2022, 2:14 AM
57
points
35
comments
1
min read
LW
link
EA & LW Forums Weekly Summary (21 Aug − 27 Aug 22′)
Zoe Williams
Aug 30, 2022, 1:42 AM
57
points
4
comments
12
min read
LW
link
Can We Align a Self-Improving AGI?
Peter S. Park
Aug 30, 2022, 12:14 AM
8
points
5
comments
11
min read
LW
link
On the nature of help—a framework for helping
Faustify
Aug 29, 2022, 8:42 PM
3
points
2
comments
13
min read
LW
link
Fundamental Uncertainty: Chapter 4 - Why don’t we do what we think we should?
Gordon Seidoh Worley
Aug 29, 2022, 7:25 PM
15
points
6
comments
13
min read
LW
link
[Question]
How can I reconcile the two most likely requirements for humanities near-term survival.
Erlja Jkdf.
Aug 29, 2022, 6:46 PM
1
point
6
comments
1
min read
LW
link
*New* Canada AI Safety & Governance community
Wyatt Tessari L'Allié
Aug 29, 2022, 6:45 PM
21
points
0
comments
1
min read
LW
link
Are Generative World Models a Mesa-Optimization Risk?
Thane Ruthenis
Aug 29, 2022, 6:37 PM
14
points
2
comments
3
min read
LW
link
Sequencing Intro
jefftk
Aug 29, 2022, 5:50 PM
39
points
3
comments
5
min read
LW
link
(www.jefftk.com)
How Do AI Timelines Affect Existential Risk?
Stephen McAleese
Aug 29, 2022, 4:57 PM
7
points
9
comments
23
min read
LW
link
How might we align transformative AI if it’s developed very soon?
HoldenKarnofsky
Aug 29, 2022, 3:42 PM
140
points
55
comments
45
min read
LW
link
1
review
An Audio Introduction to Nick Bostrom
PeterH
Aug 29, 2022, 8:50 AM
12
points
0
comments
1
min read
LW
link
(forum.effectivealtruism.org)
Please Do Fight the Hypothetical
Lone Pine
Aug 29, 2022, 8:35 AM
18
points
6
comments
3
min read
LW
link
Have you considered getting rid of death?
Willa
Aug 29, 2022, 1:31 AM
20
points
19
comments
1
min read
LW
link
(immortalityisgreat.substack.com)
(My understanding of) What Everyone in Technical Alignment is Doing and Why
Thomas Larsen
and
elifland
Aug 29, 2022, 1:23 AM
413
points
90
comments
37
min read
LW
link
1
review
Breaking down the training/deployment dichotomy
Erik Jenner
Aug 28, 2022, 9:45 PM
30
points
3
comments
3
min read
LW
link
More Clothes Over Time?
jefftk
Aug 28, 2022, 8:30 PM
30
points
1
comment
1
min read
LW
link
(www.jefftk.com)
The Expanding Moral Cinematic Universe
Raemon
Aug 28, 2022, 6:42 PM
67
points
9
comments
14
min read
LW
link
An Introduction to Current Theories of Consciousness
hohenheim
Aug 28, 2022, 5:55 PM
60
points
43
comments
49
min read
LW
link
[Linkpost] Can lab-grown brains become conscious?
Jack R
Aug 28, 2022, 5:45 PM
14
points
3
comments
1
min read
LW
link
Robert Long On Why Artificial Sentience Might Matter
Michaël Trazzi
Aug 28, 2022, 5:30 PM
29
points
5
comments
5
min read
LW
link
(theinsideview.ai)
Artificial Moral Advisors: A New Perspective from Moral Psychology
David Gross
Aug 28, 2022, 4:37 PM
25
points
1
comment
1
min read
LW
link
(dl.acm.org)
Pronunciations
Solenoid_Entity
Aug 28, 2022, 11:43 AM
15
points
7
comments
2
min read
LW
link
First thing AI will do when it takes over is get fission going
visiax
Aug 28, 2022, 5:56 AM
−2
points
0
comments
1
min read
LW
link
Who ordered alignment’s apple?
Eleni Angelou
Aug 28, 2022, 4:05 AM
6
points
3
comments
3
min read
LW
link
Sufficiently many Godzillas as an alignment strategy
142857
Aug 28, 2022, 12:08 AM
8
points
3
comments
1
min read
LW
link
[Question]
What would you expect a massive multimodal online federated learner to be capable of?
Aryeh Englander
Aug 27, 2022, 5:31 PM
13
points
4
comments
1
min read
LW
link
Basin broadness depends on the size and number of orthogonal features
CallumMcDougall
,
Avery
and
Lucius Bushnaq
Aug 27, 2022, 5:29 PM
36
points
21
comments
6
min read
LW
link
Informal semantics and Orders
Q Home
Aug 27, 2022, 4:17 AM
14
points
10
comments
26
min read
LW
link
Help Understanding Preferences And Evil
Netcentrica
Aug 27, 2022, 3:42 AM
6
points
7
comments
2
min read
LW
link
Contra Dance Contact Tracing
jefftk
Aug 27, 2022, 1:50 AM
9
points
0
comments
1
min read
LW
link
(www.jefftk.com)
Annual AGI Benchmarking Event
Lawrence Phillips
Aug 27, 2022, 12:06 AM
24
points
3
comments
2
min read
LW
link
(www.metaculus.com)
Is there a benefit in low capability AI Alignment research?
Letti
Aug 26, 2022, 11:51 PM
1
point
1
comment
2
min read
LW
link
AI Risk in Terms of Unstable Nuclear Software
Thane Ruthenis
Aug 26, 2022, 6:49 PM
30
points
1
comment
6
min read
LW
link
Taking the parameters which seem to matter and rotating them until they don’t
Garrett Baker
Aug 26, 2022, 6:26 PM
120
points
48
comments
1
min read
LW
link
ACX Meetups Everywhere List
Scott Alexander
Aug 26, 2022, 6:12 PM
63
points
1
comment
41
min read
LW
link
What’s the Most Impressive Thing That GPT-4 Could Plausibly Do?
bayesed
Aug 26, 2022, 3:34 PM
24
points
22
comments
1
min read
LW
link
[Question]
Is population collapse due to low birth rates a problem?
mukashi
Aug 26, 2022, 3:28 PM
6
points
36
comments
1
min read
LW
link
[Question]
Could you please share a tool to help with reasoning or make better decisions?
hodovani
Aug 26, 2022, 10:36 AM
1
point
0
comments
1
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel