Archive
Sequences
About
Search
Log In
Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
RSS
New
Hot
Active
Old
Page
1
Solving Slop
Noam Makavy
18 Mar 2026 5:18 UTC
0
points
0
comments
1
min read
LW
link
Sycophancy Towards Researchers Drives Performative Misalignment
Taywon Min
,
rustem17
and
David Vella Zarb
18 Mar 2026 4:59 UTC
15
points
0
comments
21
min read
LW
link
The Psychopathy Spectrum
Dawn Drescher
17 Mar 2026 21:36 UTC
21
points
0
comments
1
min read
LW
link
(impartial-priorities.org)
LLMs as Giant Lookup-Tables of Shallow Circuits
niplav
and
Claude+
17 Mar 2026 21:35 UTC
53
points
8
comments
7
min read
LW
link
Requiem for a Transhuman Timeline
Ihor Kendiukhov
17 Mar 2026 21:27 UTC
53
points
1
comment
5
min read
LW
link
There is No One There: A simple experiment to convince yourself that LLMs probably are not conscious
Peter Kuhn
17 Mar 2026 17:26 UTC
5
points
8
comments
5
min read
LW
link
Research note on window shifting training
Kei Nishimura-Gasparian
and
np_x
17 Mar 2026 15:58 UTC
18
points
0
comments
15
min read
LW
link
How to not do decision theory backwards
Anthony DiGiovanni
17 Mar 2026 7:22 UTC
14
points
0
comments
16
min read
LW
link
Monday AI Radar #17
Against Moloch
17 Mar 2026 4:42 UTC
5
points
0
comments
6
min read
LW
link
The bitter lesson for software
zef
,
rohuang
and
kaivu
16 Mar 2026 23:38 UTC
14
points
2
comments
2
min read
LW
link
(fulcruminc.substack.com)
Types of Handoff to AIs
Daniel Kokotajlo
16 Mar 2026 22:24 UTC
54
points
7
comments
8
min read
LW
link
You can’t imitation-learn how to continual-learn
Steven Byrnes
16 Mar 2026 21:20 UTC
105
points
23
comments
6
min read
LW
link
PSA: Predictions markets often have very low liquidity; be careful citing them.
Eye You
16 Mar 2026 21:07 UTC
105
points
9
comments
3
min read
LW
link
The Plan
Commander Zander
16 Mar 2026 20:58 UTC
5
points
0
comments
1
min read
LW
link
What Are My Values?
Corm
16 Mar 2026 20:43 UTC
5
points
0
comments
8
min read
LW
link
Three Properties for Alignment (and Why We’re Not Training Them)
Quentin FEUILLADE--MONTIXI
16 Mar 2026 20:26 UTC
8
points
4
comments
3
min read
LW
link
Do LLMs Have Stable Preferences?
Robert Gambee
16 Mar 2026 20:09 UTC
6
points
0
comments
7
min read
LW
link
(github.com)
The Fermi Paradox Implies Domination
Noam Makavy
16 Mar 2026 20:04 UTC
1
point
3
comments
2
min read
LW
link
Adding Typos Made Haiku’s Accuracy Go Up
bira
16 Mar 2026 18:31 UTC
31
points
3
comments
3
min read
LW
link
What are the best ways to publish rational fiction nowadays?
Ihor Kendiukhov
16 Mar 2026 18:13 UTC
14
points
11
comments
3
min read
LW
link
Back to top
Next