Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
2
London Rationalish Meetup 2022-09-11
calmiguana
Sep 9, 2022, 6:39 PM
1
point
0
comments
1
min read
LW
link
AI alignment with humans… but with which humans?
geoffreymiller
Sep 9, 2022, 6:21 PM
12
points
33
comments
3
min read
LW
link
[Question]
Should you refrain from having children because of the risk posed by artificial intelligence?
Mientras
Sep 9, 2022, 5:39 PM
17
points
31
comments
1
min read
LW
link
Notes on Resolve
David Gross
Sep 9, 2022, 4:42 PM
10
points
3
comments
31
min read
LW
link
Oversight Leagues: The Training Game as a Feature
Paul Bricman
Sep 9, 2022, 10:08 AM
20
points
6
comments
10
min read
LW
link
Understanding and avoiding value drift
TurnTrout
Sep 9, 2022, 4:16 AM
48
points
14
comments
6
min read
LW
link
Samotsvety’s AI risk forecasts
elifland
Sep 9, 2022, 4:01 AM
44
points
0
comments
4
min read
LW
link
Most People Start With The Same Few Bad Ideas
johnswentworth
Sep 9, 2022, 12:29 AM
165
points
30
comments
3
min read
LW
link
Monitoring for deceptive alignment
evhub
Sep 8, 2022, 11:07 PM
135
points
8
comments
9
min read
LW
link
[An email with a bunch of links I sent an experienced ML researcher interested in learning about Alignment / x-safety.]
David Scott Krueger (formerly: capybaralet)
Sep 8, 2022, 10:28 PM
47
points
1
comment
5
min read
LW
link
Progress links & tweets, 2022-09-08
jasoncrawford
Sep 8, 2022, 8:43 PM
13
points
3
comments
1
min read
LW
link
(rootsofprogress.org)
Postmortem: Trying out for Manifold Markets
Milli | Martin
and
Austin Chen
Sep 8, 2022, 5:54 PM
24
points
0
comments
3
min read
LW
link
Thoughts on AGI consciousness / sentience
Steven Byrnes
Sep 8, 2022, 4:40 PM
38
points
37
comments
6
min read
LW
link
A rough idea for solving ELK: An approach for training generalist agents like GATO to make plans and describe them to humans clearly and honestly.
Michael Soareverix
Sep 8, 2022, 3:20 PM
2
points
2
comments
2
min read
LW
link
What Should AI Owe To Us? Accountable and Aligned AI Systems via Contractualist AI Alignment
xuan
Sep 8, 2022, 3:04 PM
26
points
16
comments
25
min read
LW
link
ACX Book Review Discussion
Screwtape
Sep 8, 2022, 2:22 PM
5
points
0
comments
1
min read
LW
link
Covid 9/8/22: Booster Boosting
Zvi
Sep 8, 2022, 1:50 PM
34
points
9
comments
24
min read
LW
link
(thezvi.wordpress.com)
Solar Blackout Resistance
jefftk
Sep 8, 2022, 1:30 PM
69
points
32
comments
3
min read
LW
link
(www.jefftk.com)
All AGI safety questions welcome (especially basic ones) [Sept 2022]
plex
Sep 8, 2022, 11:56 AM
22
points
48
comments
3
min read
LW
link
[Question]
Sequences/Eliezer essays beyond those in AI to Zombies?
Domenic
Sep 8, 2022, 5:05 AM
4
points
4
comments
1
min read
LW
link
Linkpost: Github Copilot productivity experiment
Daniel Kokotajlo
Sep 8, 2022, 4:41 AM
88
points
4
comments
1
min read
LW
link
(github.blog)
OpenPrinciples Bootcamp (Free) -- Reflect & Act on your Rationality Principles.
ti_guo
Sep 8, 2022, 3:06 AM
6
points
3
comments
4
min read
LW
link
Searching for Modularity in Large Language Models
NickyP
and
Stephen Fowler
Sep 8, 2022, 2:25 AM
44
points
3
comments
14
min read
LW
link
90% of anything should be bad (& the precision-recall tradeoff)
cartografie
Sep 8, 2022, 1:20 AM
33
points
22
comments
6
min read
LW
link
How to Do Research. v1
Pablo Repetto
Sep 8, 2022, 1:08 AM
29
points
4
comments
41
min read
LW
link
(pabloernesto.github.io)
Galaxy Trucker Needs a New Second Half
jefftk
Sep 7, 2022, 8:10 PM
13
points
7
comments
1
min read
LW
link
(www.jefftk.com)
[Question]
In a lack of data, how should you weigh credences in theoretical physics’s Theories of Everything, or TOEs?
Noosphere89
Sep 7, 2022, 6:25 PM
7
points
11
comments
1
min read
LW
link
Generators Of Disagreement With AI Alignment
George3d6
Sep 7, 2022, 6:15 PM
27
points
9
comments
9
min read
LW
link
(www.epistem.ink)
Shrödinger’s lottery or: Why you are going to live forever
Chase Dowdell
Sep 7, 2022, 6:13 PM
1
point
2
comments
4
min read
LW
link
Is training data going to be diluted by AI-generated content?
Hannes Thurnherr
Sep 7, 2022, 6:13 PM
10
points
7
comments
1
min read
LW
link
It’s (not) how you use it
Eleni Angelou
Sep 7, 2022, 5:15 PM
8
points
1
comment
2
min read
LW
link
First we shape our social graph; then it shapes us
Henrik Karlsson
Sep 7, 2022, 3:50 PM
53
points
6
comments
8
min read
LW
link
(escapingflatland.substack.com)
AI-assisted list of ten concrete alignment things to do right now
lemonhope
Sep 7, 2022, 8:38 AM
8
points
5
comments
4
min read
LW
link
Can “Reward Economics” solve AI Alignment?
Q Home
Sep 7, 2022, 7:58 AM
3
points
15
comments
18
min read
LW
link
Is there a list of projects to get started with Interpretability?
Franziska Fischer
Sep 7, 2022, 4:27 AM
8
points
2
comments
1
min read
LW
link
Progress Report 7: making GPT go hurrdurr instead of brrrrrrr
Nathan Helm-Burger
Sep 7, 2022, 3:28 AM
21
points
0
comments
4
min read
LW
link
Framing AI Childhoods
David Udell
Sep 6, 2022, 11:40 PM
37
points
8
comments
4
min read
LW
link
Deleted comments archive
Said Achmiz
Sep 6, 2022, 9:54 PM
9
points
3
comments
1
min read
LW
link
Guitar Pedals on Fiddle
jefftk
Sep 6, 2022, 7:30 PM
10
points
0
comments
2
min read
LW
link
(www.jefftk.com)
Rejected Early Drafts of Newcomb’s Problem
zahmahkibo
Sep 6, 2022, 7:04 PM
114
points
5
comments
3
min read
LW
link
[Question]
How can we secure more research positions at our universities for x-risk researchers?
Neil Crawford
Sep 6, 2022, 5:17 PM
11
points
0
comments
1
min read
LW
link
Community Building for Graduate Students: A Targeted Approach
Neil Crawford
Sep 6, 2022, 5:17 PM
6
points
0
comments
4
min read
LW
link
How Josiah became an AI safety researcher
Neil Crawford
Sep 6, 2022, 5:17 PM
4
points
0
comments
1
min read
LW
link
No, human brains are not (much) more efficient than computers
Jesse Hoogland
6 Sep 2022 13:53 UTC
22
points
21
comments
3
min read
LW
link
(www.jessehoogland.com)
On oxytocin-sensitive neurons in auditory cortex
Steven Byrnes
6 Sep 2022 12:54 UTC
32
points
6
comments
12
min read
LW
link
EA & LW Forums Weekly Summary (28 Aug − 3 Sep 22’)
Zoe Williams
6 Sep 2022 11:06 UTC
51
points
2
comments
14
min read
LW
link
Alex Lawsen On Forecasting AI Progress
Michaël Trazzi
6 Sep 2022 9:32 UTC
18
points
0
comments
2
min read
LW
link
(theinsideview.ai)
What are you for?
lsusr
6 Sep 2022 3:32 UTC
39
points
5
comments
1
min read
LW
link
The Power (and limits?) of Chunking
Nicholas / Heather Kross
6 Sep 2022 2:26 UTC
8
points
2
comments
1
min read
LW
link
Another Unphrased B-part
jefftk
6 Sep 2022 1:30 UTC
10
points
0
comments
2
min read
LW
link
(www.jefftk.com)
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel