Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
Page
2
AI Safety in a Vulnerable World: Requesting Feedback on Preliminary Thoughts
Jordan Arel
Dec 6, 2022, 10:35 PM
4
points
2
comments
3
min read
LW
link
ChatGPT and the Human Race
Ben Reilly
Dec 6, 2022, 9:38 PM
6
points
1
comment
3
min read
LW
link
[Question]
How do finite factored sets compare with phase space?
Alex_Altair
Dec 6, 2022, 8:05 PM
15
points
1
comment
1
min read
LW
link
Mesa-Optimizers via Grokking
orthonormal
Dec 6, 2022, 8:05 PM
36
points
4
comments
6
min read
LW
link
Using GPT-Eliezer against ChatGPT Jailbreaking
Stuart_Armstrong
and
rgorman
Dec 6, 2022, 7:54 PM
170
points
85
comments
9
min read
LW
link
The Parable of the Crimp
Phosphorous
Dec 6, 2022, 6:41 PM
11
points
3
comments
3
min read
LW
link
The Categorical Imperative Obscures
Gordon Seidoh Worley
Dec 6, 2022, 5:48 PM
17
points
17
comments
2
min read
LW
link
MIRI’s “Death with Dignity” in 60 seconds.
Cleo Nardo
Dec 6, 2022, 5:18 PM
58
points
4
comments
1
min read
LW
link
Things roll downhill
awenonian
Dec 6, 2022, 3:27 PM
19
points
0
comments
1
min read
LW
link
EA & LW Forums Weekly Summary (28th Nov − 4th Dec 22′)
Zoe Williams
Dec 6, 2022, 9:38 AM
10
points
1
comment
LW
link
Take 5: Another problem for natural abstractions is laziness.
Charlie Steiner
Dec 6, 2022, 7:00 AM
31
points
4
comments
3
min read
LW
link
Verification Is Not Easier Than Generation In General
johnswentworth
Dec 6, 2022, 5:20 AM
73
points
27
comments
1
min read
LW
link
Shh, don’t tell the AI it’s likely to be evil
naterush
Dec 6, 2022, 3:35 AM
19
points
9
comments
1
min read
LW
link
[Question]
What are the major underlying divisions in AI safety?
Chris_Leong
Dec 6, 2022, 3:28 AM
5
points
2
comments
1
min read
LW
link
[Link] Why I’m optimistic about OpenAI’s alignment approach
janleike
Dec 5, 2022, 10:51 PM
98
points
15
comments
1
min read
LW
link
(aligned.substack.com)
The No Free Lunch theorem for dummies
Steven Byrnes
Dec 5, 2022, 9:46 PM
37
points
16
comments
3
min read
LW
link
ChatGPT and Ideological Turing Test
Viliam
Dec 5, 2022, 9:45 PM
42
points
1
comment
1
min read
LW
link
ChatGPT on Spielberg’s A.I. and AI Alignment
Bill Benzon
Dec 5, 2022, 9:10 PM
5
points
0
comments
4
min read
LW
link
Updating my AI timelines
Matthew Barnett
Dec 5, 2022, 8:46 PM
145
points
50
comments
2
min read
LW
link
Steering Behaviour: Testing for (Non-)Myopia in Language Models
Evan R. Murphy
and
Megan Kinniment
Dec 5, 2022, 8:28 PM
40
points
19
comments
10
min read
LW
link
College Admissions as a Brutal One-Shot Game
devansh
Dec 5, 2022, 8:05 PM
8
points
26
comments
2
min read
LW
link
Analysis of AI Safety surveys for field-building insights
Ash Jafari
Dec 5, 2022, 7:21 PM
11
points
2
comments
5
min read
LW
link
Testing Ways to Bypass ChatGPT’s Safety Features
Robert_AIZI
Dec 5, 2022, 6:50 PM
7
points
4
comments
5
min read
LW
link
(aizi.substack.com)
Foresight for AGI Safety Strategy: Mitigating Risks and Identifying Golden Opportunities
jacquesthibs
Dec 5, 2022, 4:09 PM
28
points
6
comments
8
min read
LW
link
Aligned Behavior is not Evidence of Alignment Past a Certain Level of Intelligence
Ronny Fernandez
Dec 5, 2022, 3:19 PM
19
points
5
comments
7
min read
LW
link
[Question]
How should I judge the impact of giving $5k to a family of three kids and two mentally ill parents?
Blake
Dec 5, 2022, 1:42 PM
10
points
10
comments
1
min read
LW
link
Is the “Valley of Confused Abstractions” real?
jacquesthibs
Dec 5, 2022, 1:36 PM
20
points
11
comments
2
min read
LW
link
Take 4: One problem with natural abstractions is there’s too many of them.
Charlie Steiner
Dec 5, 2022, 10:39 AM
37
points
4
comments
1
min read
LW
link
[Question]
What are some good Lesswrong-related accounts or hashtags on Mastodon that I should follow?
SpectrumDT
Dec 5, 2022, 9:42 AM
2
points
0
comments
1
min read
LW
link
[Question]
Who are some prominent reasonable people who are confident that AI won’t kill everyone?
Optimization Process
Dec 5, 2022, 9:12 AM
72
points
54
comments
1
min read
LW
link
Monthly Shorts 11/22
Celer
Dec 5, 2022, 7:30 AM
8
points
0
comments
3
min read
LW
link
(keller.substack.com)
A ChatGPT story about ChatGPT doom
SurfingOrca
Dec 5, 2022, 5:40 AM
6
points
2
comments
4
min read
LW
link
A Tentative Timeline of The Near Future (2022-2025) for Self-Accountability
Yitz
Dec 5, 2022, 5:33 AM
26
points
0
comments
4
min read
LW
link
Nook Nature
Duncan Sabien (Inactive)
Dec 5, 2022, 4:10 AM
54
points
18
comments
10
min read
LW
link
Probably good projects for the AI safety ecosystem
Ryan Kidd
Dec 5, 2022, 2:26 AM
78
points
40
comments
2
min read
LW
link
Historical Notes on Charitable Funds
jefftk
Dec 4, 2022, 11:30 PM
28
points
0
comments
3
min read
LW
link
(www.jefftk.com)
AGI as a Black Swan Event
Stephen McAleese
Dec 4, 2022, 11:00 PM
8
points
8
comments
7
min read
LW
link
South Bay ACX/LW Pre-Holiday Get-Together
IS
4 Dec 2022 22:57 UTC
10
points
0
comments
1
min read
LW
link
ChatGPT is settling the Chinese Room argument
averros
4 Dec 2022 20:25 UTC
−7
points
7
comments
1
min read
LW
link
Race to the Top: Benchmarks for AI Safety
Isabella Duan
4 Dec 2022 18:48 UTC
29
points
6
comments
1
min read
LW
link
Open & Welcome Thread—December 2022
niplav
4 Dec 2022 15:06 UTC
8
points
22
comments
1
min read
LW
link
AI can exploit safety plans posted on the Internet
Peter S. Park
4 Dec 2022 12:17 UTC
−15
points
4
comments
LW
link
ChatGPT seems overconfident to me
qbolec
4 Dec 2022 8:03 UTC
19
points
3
comments
16
min read
LW
link
Could an AI be Religious?
mk54
4 Dec 2022 5:00 UTC
−12
points
14
comments
1
min read
LW
link
Can GPT-3 Write Contra Dances?
jefftk
4 Dec 2022 3:00 UTC
6
points
4
comments
10
min read
LW
link
(www.jefftk.com)
Take 3: No indescribable heavenworlds.
Charlie Steiner
4 Dec 2022 2:48 UTC
23
points
12
comments
2
min read
LW
link
Summary of a new study on out-group hate (and how to fix it)
DirectedEvolution
4 Dec 2022 1:53 UTC
60
points
30
comments
3
min read
LW
link
(www.pnas.org)
[Question]
Will the first AGI agent have been designed as an agent (in addition to an AGI)?
nahoj
3 Dec 2022 20:32 UTC
1
point
8
comments
1
min read
LW
link
Logical induction for software engineers
Alex Flint
3 Dec 2022 19:55 UTC
163
points
8
comments
27
min read
LW
link
1
review
Utilitarianism is the only option
aelwood
3 Dec 2022 17:14 UTC
−13
points
7
comments
LW
link
Previous
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel