Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
1
Positive values seem more robust and lasting than prohibitions
TurnTrout
Dec 17, 2022, 9:43 PM
52
points
13
comments
2
min read
LW
link
What we owe the microbiome
weverka
Dec 17, 2022, 7:40 PM
2
points
0
comments
1
min read
LW
link
(forum.effectivealtruism.org)
Why write more: improve your epistemics, self-care, & 28 other reasons
KatWoods
Dec 17, 2022, 7:25 PM
24
points
1
comment
6
min read
LW
link
Looking for an alignment tutor
JanB
Dec 17, 2022, 7:08 PM
15
points
2
comments
1
min read
LW
link
[Question]
How to Convince my Son that Drugs are Bad
concerned_dad
Dec 17, 2022, 6:47 PM
140
points
84
comments
2
min read
LW
link
Ordinary human life
David Hugh-Jones
Dec 17, 2022, 4:46 PM
24
points
3
comments
14
min read
LW
link
(wyclif.substack.com)
Predictive Processing, Heterosexuality and Delusions of Grandeur
lsusr
Dec 17, 2022, 7:37 AM
37
points
13
comments
5
min read
LW
link
[Link] Escape the Echo Chamber (2018)
CronoDAS
Dec 17, 2022, 6:14 AM
13
points
0
comments
2
min read
LW
link
(aeon.co)
“Starry Night” Solstice Cookies
maia
Dec 17, 2022, 5:31 AM
26
points
7
comments
1
min read
LW
link
There have been 3 planes (billionaire donors) and 2 have crashed
trevor
Dec 17, 2022, 3:58 AM
16
points
10
comments
2
min read
LW
link
[Question]
What about non-degree seeking?
Lao Mein
Dec 17, 2022, 2:22 AM
5
points
5
comments
1
min read
LW
link
Using Information Theory to tackle AI Alignment: A Practical Approach
Daniel Salami
Dec 17, 2022, 1:37 AM
10
points
4
comments
7
min read
LW
link
Paper: Constitutional AI: Harmlessness from AI Feedback (Anthropic)
LawrenceC
Dec 16, 2022, 10:12 PM
68
points
11
comments
1
min read
LW
link
(www.anthropic.com)
Vaguely interested in Effective Altruism? Please Take the Official 2022 EA Survey
Peter Wildeford
Dec 16, 2022, 9:07 PM
22
points
4
comments
1
min read
LW
link
(rethinkpriorities.qualtrics.com)
Abstract concepts and metalingual definition: Does ChatGPT understand justice and charity?
Bill Benzon
Dec 16, 2022, 9:01 PM
2
points
0
comments
13
min read
LW
link
Beyond the moment of invention
jasoncrawford
Dec 16, 2022, 8:18 PM
35
points
0
comments
2
min read
LW
link
(rootsofprogress.org)
[Question]
What’s the best time-efficient alternative to the Sequences?
trevor
Dec 16, 2022, 8:17 PM
7
points
7
comments
1
min read
LW
link
Can we efficiently explain model behaviors?
paulfchristiano
Dec 16, 2022, 7:40 PM
64
points
3
comments
9
min read
LW
link
(ai-alignment.com)
Proper scoring rules don’t guarantee predicting fixed points
Johannes Treutlein
,
Rubi J. Hudson
and
Caspar Oesterheld
Dec 16, 2022, 6:22 PM
79
points
8
comments
21
min read
LW
link
A learned agent is not the same as a learning agent
Ben Amitay
Dec 16, 2022, 5:27 PM
4
points
5
comments
4
min read
LW
link
[Question]
College Selection Advice for Technical Alignment
TempCollegeAsk
Dec 16, 2022, 5:11 PM
11
points
8
comments
1
min read
LW
link
How important are accurate AI timelines for the optimal spending schedule on AI risk interventions?
Tristan Cook
Dec 16, 2022, 4:05 PM
27
points
2
comments
LW
link
Introducing Shrubgrazer
jefftk
Dec 16, 2022, 2:50 PM
22
points
0
comments
2
min read
LW
link
(www.jefftk.com)
Paper: Transformers learn in-context by gradient descent
LawrenceC
Dec 16, 2022, 11:10 AM
28
points
11
comments
2
min read
LW
link
(arxiv.org)
Will Machines Ever Rule the World? MLAISU W50
Esben Kran
Dec 16, 2022, 11:03 AM
12
points
7
comments
4
min read
LW
link
(newsletter.apartresearch.com)
AI overhangs depend on whether algorithms, compute and data are substitutes or complements
NathanBarnard
Dec 16, 2022, 2:23 AM
2
points
0
comments
3
min read
LW
link
AI Safety Movement Builders should help the community to optimise three factors: contributors, contributions and coordination
peterslattery
Dec 15, 2022, 10:50 PM
4
points
0
comments
6
min read
LW
link
Masking to Avoid Missing Things
jefftk
Dec 15, 2022, 9:00 PM
17
points
2
comments
1
min read
LW
link
(www.jefftk.com)
Consider working more hours and taking more stimulants
Arjun Panickssery
Dec 15, 2022, 8:38 PM
36
points
11
comments
LW
link
We’ve stepped over the threshold into the Fourth Arena, but don’t recognize it
Bill Benzon
Dec 15, 2022, 8:22 PM
2
points
0
comments
7
min read
LW
link
[Question]
How is ARC planning to use ELK?
jacquesthibs
Dec 15, 2022, 8:11 PM
24
points
5
comments
1
min read
LW
link
How “Discovering Latent Knowledge in Language Models Without Supervision” Fits Into a Broader Alignment Scheme
Collin
Dec 15, 2022, 6:22 PM
244
points
39
comments
16
min read
LW
link
1
review
High-level hopes for AI alignment
HoldenKarnofsky
Dec 15, 2022, 6:00 PM
58
points
3
comments
19
min read
LW
link
(www.cold-takes.com)
Two Dogmas of LessWrong
omnizoid
Dec 15, 2022, 5:56 PM
−7
points
155
comments
69
min read
LW
link
Covid 12/15/22: China’s Wave Begins
Zvi
Dec 15, 2022, 4:20 PM
32
points
7
comments
10
min read
LW
link
(thezvi.wordpress.com)
The next decades might be wild
Marius Hobbhahn
Dec 15, 2022, 4:10 PM
175
points
42
comments
41
min read
LW
link
1
review
Basic building blocks of dependent type theory
Thomas Kehrenberg
Dec 15, 2022, 2:54 PM
49
points
9
comments
13
min read
LW
link
AI Neorealism: a threat model & success criterion for existential safety
davidad
Dec 15, 2022, 1:42 PM
67
points
1
comment
3
min read
LW
link
Who should write the definitive post on Ziz?
Nicholas / Heather Kross
Dec 15, 2022, 6:37 AM
4
points
45
comments
3
min read
LW
link
[Question]
Is Paul Christiano still as optimistic about Approval-Directed Agents as he was in 2018?
Chris_Leong
Dec 14, 2022, 11:28 PM
8
points
0
comments
1
min read
LW
link
«Boundaries», Part 3b: Alignment problems in terms of boundaries
Andrew_Critch
Dec 14, 2022, 10:34 PM
72
points
7
comments
13
min read
LW
link
Aligning alignment with performance
Marv K
Dec 14, 2022, 10:19 PM
2
points
0
comments
2
min read
LW
link
Contrary to List of Lethality’s point 22, alignment’s door number 2
False Name
Dec 14, 2022, 10:01 PM
−2
points
5
comments
22
min read
LW
link
Kolmogorov Complexity and Simulation Hypothesis
False Name
Dec 14, 2022, 10:01 PM
−3
points
0
comments
7
min read
LW
link
[Question]
Stanley Meyer’s water fuel cell
mikbp
Dec 14, 2022, 9:19 PM
2
points
6
comments
1
min read
LW
link
[Question]
Is the AI timeline too short to have children?
Yoreth
Dec 14, 2022, 6:32 PM
38
points
20
comments
1
min read
LW
link
Predicting GPU performance
Marius Hobbhahn
and
Tamay
Dec 14, 2022, 4:27 PM
60
points
26
comments
1
min read
LW
link
(epochai.org)
[Incomplete] What is Computation Anyway?
DragonGod
Dec 14, 2022, 4:17 PM
16
points
1
comment
13
min read
LW
link
(arxiv.org)
Chair Hanging Peg
jefftk
Dec 14, 2022, 3:30 PM
11
points
0
comments
1
min read
LW
link
(www.jefftk.com)
My AGI safety research—2022 review, ’23 plans
Steven Byrnes
Dec 14, 2022, 3:15 PM
51
points
10
comments
7
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel