Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
1
AGIs may value intrinsic rewards more than extrinsic ones
catubc
Nov 17, 2022, 9:49 PM
8
points
6
comments
4
min read
LW
link
LLMs may capture key components of human agency
catubc
Nov 17, 2022, 8:14 PM
27
points
0
comments
4
min read
LW
link
Mastodon Replies as Comments
jefftk
Nov 17, 2022, 8:10 PM
20
points
0
comments
1
min read
LW
link
(www.jefftk.com)
Announcing the Progress Forum
jasoncrawford
Nov 17, 2022, 7:26 PM
83
points
9
comments
1
min read
LW
link
[Question]
What kind of bias is this?
Daniel Samuel
Nov 17, 2022, 6:44 PM
3
points
2
comments
1
min read
LW
link
AI Forecasting Research Ideas
Jsevillamol
Nov 17, 2022, 5:37 PM
21
points
2
comments
LW
link
Results from the interpretability hackathon
Esben Kran
and
Neel Nanda
Nov 17, 2022, 2:51 PM
81
points
0
comments
6
min read
LW
link
(alignmentjam.com)
Covid 11/17/22: Slow Recovery
Zvi
Nov 17, 2022, 2:50 PM
33
points
3
comments
4
min read
LW
link
(thezvi.wordpress.com)
Sadly, FTX
Zvi
Nov 17, 2022, 2:30 PM
133
points
18
comments
47
min read
LW
link
(thezvi.wordpress.com)
Deontology and virtue ethics as “effective theories” of consequentialist ethics
Jan_Kulveit
Nov 17, 2022, 2:11 PM
68
points
9
comments
LW
link
1
review
The Ground Truth Problem (Or, Why Evaluating Interpretability Methods Is Hard)
Jessica Rumbelow
Nov 17, 2022, 11:06 AM
27
points
2
comments
2
min read
LW
link
[Question]
[Personal Question] Can anyone help me navigate this potentially painful interpersonal dynamic rationally?
SlainLadyMondegreen
Nov 17, 2022, 8:53 AM
9
points
3
comments
4
min read
LW
link
Massive Scaling Should be Frowned Upon
harsimony
Nov 17, 2022, 8:43 AM
4
points
6
comments
5
min read
LW
link
[Question]
Why are profitable companies laying off staff?
Yair Halberstadt
Nov 17, 2022, 6:19 AM
15
points
10
comments
1
min read
LW
link
Discussion: Was SBF a naive utilitarian, or a sociopath?
Nicholas / Heather Kross
Nov 17, 2022, 2:52 AM
0
points
4
comments
LW
link
Kelsey Piper’s recent interview of SBF
agucova
Nov 16, 2022, 8:30 PM
51
points
29
comments
LW
link
The Echo Principle
Jonathan Moregård
Nov 16, 2022, 8:09 PM
4
points
0
comments
3
min read
LW
link
(honestliving.substack.com)
[Question]
Is there some reason LLMs haven’t seen broader use?
tailcalled
Nov 16, 2022, 8:04 PM
25
points
27
comments
1
min read
LW
link
When should we be surprised that an invention took “so long”?
jasoncrawford
Nov 16, 2022, 8:04 PM
32
points
11
comments
4
min read
LW
link
(rootsofprogress.org)
Questions about Value Lock-in, Paternalism, and Empowerment
Sam F. Brown
Nov 16, 2022, 3:33 PM
13
points
2
comments
12
min read
LW
link
(sambrown.eu)
If Professional Investors Missed This...
jefftk
Nov 16, 2022, 3:00 PM
37
points
18
comments
3
min read
LW
link
(www.jefftk.com)
Disagreement with bio anchors that lead to shorter timelines
Marius Hobbhahn
Nov 16, 2022, 2:40 PM
75
points
17
comments
7
min read
LW
link
1
review
Current themes in mechanistic interpretability research
Lee Sharkey
,
Sid Black
and
beren
Nov 16, 2022, 2:14 PM
89
points
2
comments
12
min read
LW
link
Unpacking “Shard Theory” as Hunch, Question, Theory, and Insight
Jacy Reese Anthis
Nov 16, 2022, 1:54 PM
31
points
9
comments
2
min read
LW
link
Miracles and why not to believe them
mruwnik
Nov 16, 2022, 12:07 PM
4
points
0
comments
2
min read
LW
link
[Question]
How do people do remote research collaborations effectively?
Krieger
Nov 16, 2022, 11:51 AM
8
points
0
comments
1
min read
LW
link
Method of statements: an alternative to taboo
Q Home
Nov 16, 2022, 10:57 AM
7
points
0
comments
41
min read
LW
link
The two conceptions of Active Inference: an intelligence architecture and a theory of agency
Roman Leventov
Nov 16, 2022, 9:30 AM
17
points
0
comments
4
min read
LW
link
Developer experience for the motivation
Adam Zerner
Nov 16, 2022, 7:12 AM
49
points
7
comments
4
min read
LW
link
Progress links and tweets, 2022-11-15
jasoncrawford
Nov 16, 2022, 3:21 AM
9
points
0
comments
2
min read
LW
link
(rootsofprogress.org)
EA & LW Forums Weekly Summary (7th Nov − 13th Nov 22′)
Zoe Williams
Nov 16, 2022, 3:04 AM
19
points
0
comments
LW
link
The FTX Saga—Simplified
Annapurna
Nov 16, 2022, 2:42 AM
44
points
10
comments
7
min read
LW
link
(jorgevelez.substack.com)
Utilitarianism and the idea of a “rational agent” are fundamentally inconsistent with reality
banev
Nov 16, 2022, 12:19 AM
−4
points
1
comment
1
min read
LW
link
[Question]
Is the speed of training large models going to increase significantly in the near future due to Cerebras Andromeda?
Amal
Nov 15, 2022, 10:50 PM
13
points
11
comments
1
min read
LW
link
[Question]
What is our current best infohazard policy for AGI (safety) research?
Roman Leventov
Nov 15, 2022, 10:33 PM
12
points
2
comments
1
min read
LW
link
ACX/SSC Meetup 1 pm Sunday Nov 20
svfritz
Nov 15, 2022, 8:39 PM
2
points
0
comments
1
min read
LW
link
SBF x LoL
Nicholas / Heather Kross
Nov 15, 2022, 8:24 PM
17
points
6
comments
LW
link
Some research ideas in forecasting
Jsevillamol
Nov 15, 2022, 7:47 PM
35
points
2
comments
LW
link
Strategy of Inner Conflict
Jonathan Moregård
Nov 15, 2022, 7:38 PM
9
points
4
comments
6
min read
LW
link
(honestliving.substack.com)
The limited upside of interpretability
Peter S. Park
Nov 15, 2022, 6:46 PM
13
points
11
comments
LW
link
Why bet Kelly?
AlexMennen
Nov 15, 2022, 6:12 PM
32
points
14
comments
5
min read
LW
link
Entropy Scaling And Intrinsic Memory
Alexander Gietelink Oldenziel
and
Adam Shai
Nov 15, 2022, 6:11 PM
20
points
5
comments
5
min read
LW
link
[Question]
Will nanotech/biotech be what leads to AI doom?
tailcalled
Nov 15, 2022, 5:38 PM
4
points
9
comments
2
min read
LW
link
Value Formation: An Overarching Model
Thane Ruthenis
15 Nov 2022 17:16 UTC
34
points
20
comments
34
min read
LW
link
Internal communication framework
rosehadshar
and
Nora_Ammann
15 Nov 2022 12:41 UTC
38
points
13
comments
12
min read
LW
link
Better Mastodon Aliases
jefftk
15 Nov 2022 12:10 UTC
14
points
3
comments
1
min read
LW
link
(www.jefftk.com)
The economy as an analogy for advanced AI systems
rosehadshar
and
particlemania
15 Nov 2022 11:16 UTC
28
points
0
comments
5
min read
LW
link
We need better prediction markets
eigen
15 Nov 2022 4:54 UTC
9
points
8
comments
1
min read
LW
link
Preventing, reversing, and addressing data leakage: some thoughts
VipulNaik
15 Nov 2022 2:09 UTC
14
points
4
comments
25
min read
LW
link
Winners of the AI Safety Nudge Competition
Marc Carauleanu
15 Nov 2022 1:06 UTC
4
points
0
comments
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel