Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
2
Covid 10/27/22: Another Origin Story
Zvi
Oct 27, 2022, 1:40 PM
32
points
1
comment
13
min read
LW
link
(thezvi.wordpress.com)
[Question]
Why are probabilities represented as real numbers instead of rational numbers?
Yaakov T
Oct 27, 2022, 11:23 AM
5
points
9
comments
1
min read
LW
link
Five Areas I Wish EAs Gave More Focus
Prometheus
Oct 27, 2022, 6:13 AM
13
points
18
comments
LW
link
Apply to the Redwood Research Mechanistic Interpretability Experiment (REMIX), a research program in Berkeley
maxnadeau
,
Xander Davies
,
Buck
and
Nate Thomas
Oct 27, 2022, 1:32 AM
135
points
14
comments
12
min read
LW
link
[Question]
Quantum Suicide and Aumann’s Agreement Theorem
Isaac King
Oct 27, 2022, 1:32 AM
14
points
20
comments
1
min read
LW
link
Reslab Request for Information: EA hardware projects
Joel Becker
Oct 26, 2022, 9:13 PM
10
points
0
comments
LW
link
A list of Petrov buttons
philh
Oct 26, 2022, 8:50 PM
19
points
8
comments
5
min read
LW
link
(reasonableapproximation.net)
The Game of Antonyms
Faustify
Oct 26, 2022, 7:26 PM
4
points
6
comments
8
min read
LW
link
Paper: In-context Reinforcement Learning with Algorithm Distillation [Deepmind]
LawrenceC
Oct 26, 2022, 6:45 PM
29
points
5
comments
1
min read
LW
link
(arxiv.org)
[Question]
How to become more articulate?
just_browsing
Oct 26, 2022, 2:43 PM
19
points
14
comments
1
min read
LW
link
Open Bands: Leading Rhythm
jefftk
Oct 26, 2022, 2:30 PM
10
points
0
comments
4
min read
LW
link
(www.jefftk.com)
Signals of war in August 2021
yieldthought
Oct 26, 2022, 8:11 AM
70
points
16
comments
2
min read
LW
link
Trigger-based rapid checklists
VipulNaik
Oct 26, 2022, 4:05 AM
44
points
0
comments
9
min read
LW
link
Why some people believe in AGI, but I don’t.
cveres
Oct 26, 2022, 3:09 AM
−15
points
6
comments
LW
link
Intent alignment should not be the goal for AGI x-risk reduction
John Nay
Oct 26, 2022, 1:24 AM
1
point
10
comments
3
min read
LW
link
Reinforcement Learning Goal Misgeneralization: Can we guess what kind of goals are selected by default?
StefanHex
and
Julian_R
Oct 25, 2022, 8:48 PM
15
points
2
comments
4
min read
LW
link
A Walkthrough of A Mathematical Framework for Transformer Circuits
Neel Nanda
Oct 25, 2022, 8:24 PM
52
points
7
comments
1
min read
LW
link
(www.youtube.com)
Nothing.
rogersbacon
Oct 25, 2022, 4:33 PM
−10
points
4
comments
6
min read
LW
link
(www.secretorum.life)
Maps and Blueprint; the Two Sides of the Alignment Equation
Nora_Ammann
Oct 25, 2022, 4:29 PM
24
points
1
comment
5
min read
LW
link
Consider Applying to the Future Fellowship at MIT
jefftk
Oct 25, 2022, 3:40 PM
29
points
0
comments
1
min read
LW
link
(www.jefftk.com)
Beyond Kolmogorov and Shannon
Alexander Gietelink Oldenziel
and
Adam Shai
Oct 25, 2022, 3:13 PM
63
points
22
comments
5
min read
LW
link
What does it take to defend the world against out-of-control AGIs?
Steven Byrnes
Oct 25, 2022, 2:47 PM
208
points
49
comments
30
min read
LW
link
1
review
Refine: what helped me write more?
Alexander Gietelink Oldenziel
Oct 25, 2022, 2:44 PM
12
points
0
comments
2
min read
LW
link
Logical Decision Theories: Our final failsafe?
Noosphere89
Oct 25, 2022, 12:51 PM
−7
points
8
comments
1
min read
LW
link
(www.lesswrong.com)
What will the scaled up GATO look like? (Updated with questions)
Amal
Oct 25, 2022, 12:44 PM
34
points
22
comments
1
min read
LW
link
Mechanism Design for AI Safety—Reading Group Curriculum
Rubi J. Hudson
Oct 25, 2022, 3:54 AM
15
points
3
comments
LW
link
Furry Rationalists & Effective Anthropomorphism both exist
agentydragon
Oct 25, 2022, 3:37 AM
42
points
3
comments
1
min read
LW
link
EA & LW Forums Weekly Summary (17 − 23 Oct 22′)
Zoe Williams
Oct 25, 2022, 2:57 AM
10
points
0
comments
LW
link
Dance Weekends: Tests not Masks
jefftk
Oct 25, 2022, 2:10 AM
12
points
0
comments
2
min read
LW
link
(www.jefftk.com)
[Question]
What is good Cyber Security Advice?
Gunnar_Zarncke
Oct 24, 2022, 11:27 PM
30
points
12
comments
2
min read
LW
link
Connections between Mind-Body Problem & Civilizations
oblivion
Oct 24, 2022, 9:55 PM
−3
points
1
comment
1
min read
LW
link
[Question]
Rationalism and money
David K
Oct 24, 2022, 9:22 PM
−5
points
2
comments
1
min read
LW
link
[Question]
Game semantics
David K
Oct 24, 2022, 9:22 PM
2
points
2
comments
1
min read
LW
link
A Good Future (rough draft)
Michael Soareverix
Oct 24, 2022, 8:45 PM
10
points
5
comments
3
min read
LW
link
A Barebones Guide to Mechanistic Interpretability Prerequisites
Neel Nanda
Oct 24, 2022, 8:45 PM
64
points
12
comments
3
min read
LW
link
(neelnanda.io)
POWERplay: An open-source toolchain to study AI power-seeking
Edouard Harris
Oct 24, 2022, 8:03 PM
29
points
0
comments
1
min read
LW
link
(github.com)
Consider trying Vivek Hebbar’s alignment exercises
Orpheus16
Oct 24, 2022, 7:46 PM
38
points
1
comment
4
min read
LW
link
[Question]
Education not meant for mass-consumption
Tolo
Oct 24, 2022, 7:45 PM
7
points
5
comments
2
min read
LW
link
Realizations in Regards to Masculinity
nmc
Oct 24, 2022, 7:42 PM
−2
points
2
comments
2
min read
LW
link
The Futility of Religion
nmc
Oct 24, 2022, 7:42 PM
−1
points
5
comments
3
min read
LW
link
The optimal timing of spending on AGI safety work; why we should probably be spending more now
Tristan Cook
Oct 24, 2022, 5:42 PM
62
points
0
comments
LW
link
AGI in our lifetimes is wishful thinking
niknoble
Oct 24, 2022, 11:53 AM
1
point
25
comments
8
min read
LW
link
DeepMind on Stratego, an imperfect information game
sanxiyn
Oct 24, 2022, 5:57 AM
15
points
9
comments
1
min read
LW
link
(arxiv.org)
[Question]
TOMT: Post from 1-2 years ago talking about a paper on social networks
Simon Berens
Oct 24, 2022, 1:29 AM
5
points
1
comment
1
min read
LW
link
AI researchers announce NeuroAI agenda
Cameron Berg
Oct 24, 2022, 12:14 AM
37
points
12
comments
6
min read
LW
link
(arxiv.org)
Empowerment is (almost) All We Need
jacob_cannell
Oct 23, 2022, 9:48 PM
61
points
44
comments
17
min read
LW
link
“Originality is nothing but judicious imitation”—Voltaire
Vestozia
23 Oct 2022 19:00 UTC
0
points
0
comments
13
min read
LW
link
Mid-Peninsula ACX/LW Meetup [CANCELLED]
moshezadka
23 Oct 2022 17:37 UTC
1
point
0
comments
1
min read
LW
link
I am a Memoryless System
Nicholas / Heather Kross
23 Oct 2022 17:34 UTC
26
points
2
comments
9
min read
LW
link
(www.thinkingmuchbetter.com)
Accountability Buddies: Why you might want one.
Samuel Nellessen
23 Oct 2022 16:25 UTC
10
points
3
comments
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel