Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
1
Lifeguards
Orpheus16
Jun 15, 2022, 11:03 PM
12
points
3
comments
2
min read
LW
link
(forum.effectivealtruism.org)
Rationality Vienna Hike
Laszlo_Treszkai
Jun 15, 2022, 10:11 PM
3
points
0
comments
1
min read
LW
link
Contra Hofstadter on GPT-3 Nonsense
rictic
Jun 15, 2022, 9:53 PM
237
points
24
comments
2
min read
LW
link
Progress links and tweets, 2022-06-13
jasoncrawford
Jun 15, 2022, 7:47 PM
12
points
0
comments
1
min read
LW
link
(rootsofprogress.org)
I applied for a MIRI job in 2020. Here’s what happened next.
ViktoriaMalyasova
Jun 15, 2022, 7:37 PM
86
points
17
comments
7
min read
LW
link
Contextual Evil
ACrackedPot
Jun 15, 2022, 7:32 PM
1
point
12
comments
2
min read
LW
link
Multigate Priors
Adam Jermyn
Jun 15, 2022, 7:30 PM
4
points
0
comments
3
min read
LW
link
FYI: I’m working on a book about the threat of AGI/ASI for a general audience. I hope it will be of value to the cause and the community
Darren McKee
Jun 15, 2022, 6:08 PM
43
points
15
comments
2
min read
LW
link
[Question]
What are all the AI Alignment and AI Safety Communication Hubs?
Gunnar_Zarncke
Jun 15, 2022, 4:16 PM
27
points
5
comments
1
min read
LW
link
Georgism, in theory
Stuart_Armstrong
Jun 15, 2022, 3:20 PM
40
points
22
comments
4
min read
LW
link
Berlin AI Safety Open Meetup June 2022
pranomostro
Jun 15, 2022, 2:33 PM
12
points
0
comments
1
min read
LW
link
A central AI alignment problem: capabilities generalization, and the sharp left turn
So8res
Jun 15, 2022, 1:10 PM
272
points
55
comments
10
min read
LW
link
1
review
Our mental building blocks are more different than I thought
Marius Hobbhahn
Jun 15, 2022, 11:07 AM
50
points
11
comments
14
min read
LW
link
[Question]
Has there been any work on attempting to use Pascal’s Mugging to make an AGI behave?
Chris_Leong
Jun 15, 2022, 8:33 AM
7
points
17
comments
1
min read
LW
link
Alignment Risk Doesn’t Require Superintelligence
JustisMills
Jun 15, 2022, 3:12 AM
35
points
4
comments
2
min read
LW
link
A Butterfly’s View of Probability
Gabriel Wu
Jun 15, 2022, 2:14 AM
29
points
17
comments
11
min read
LW
link
[Question]
Favourite new AI productivity tools?
Gabe M
Jun 15, 2022, 1:08 AM
14
points
5
comments
1
min read
LW
link
Will vague “AI sentience” concerns do more for AI safety than anything else we might do?
Aryeh Englander
Jun 14, 2022, 11:53 PM
15
points
2
comments
1
min read
LW
link
Yes, AI research will be substantially curtailed if a lab causes a major disaster
lc
Jun 14, 2022, 10:17 PM
103
points
31
comments
2
min read
LW
link
Slow motion videos as AI risk intuition pumps
Andrew_Critch
Jun 14, 2022, 7:31 PM
241
points
41
comments
2
min read
LW
link
1
review
Cryptographic Life: How to transcend in a sub-lightspeed world via Homomorphic encryption
Golol
Jun 14, 2022, 7:22 PM
1
point
0
comments
3
min read
LW
link
Blake Richards on Why he is Skeptical of Existential Risk from AI
Michaël Trazzi
Jun 14, 2022, 7:09 PM
41
points
12
comments
4
min read
LW
link
(theinsideview.ai)
[Question]
How Do You Quantify [Physics Interfacing] Real World Capabilities?
DragonGod
Jun 14, 2022, 2:49 PM
17
points
1
comment
4
min read
LW
link
Was the Industrial Revolution The Industrial Revolution?
Davis Kedrosky
Jun 14, 2022, 2:48 PM
29
points
0
comments
12
min read
LW
link
(daviskedrosky.substack.com)
Investigating causal understanding in LLMs
Marius Hobbhahn
and
Tom Lieberum
Jun 14, 2022, 1:57 PM
28
points
6
comments
13
min read
LW
link
Why multi-agent safety is important
Akbir Khan
Jun 14, 2022, 9:23 AM
10
points
2
comments
10
min read
LW
link
[Question]
Was Eliezer Yudkowsky right to give himself 10% to succeed with HPMoR in 2010?
momom2
Jun 14, 2022, 7:00 AM
2
points
2
comments
1
min read
LW
link
Resources I send to AI researchers about AI safety
Vael Gates
Jun 14, 2022, 2:24 AM
69
points
12
comments
1
min read
LW
link
Vael Gates: Risks from Advanced AI (June 2022)
Vael Gates
Jun 14, 2022, 12:54 AM
38
points
2
comments
30
min read
LW
link
Cambridge LW Meetup: Personal Finance
Tony Wang
Jun 14, 2022, 12:12 AM
3
points
0
comments
1
min read
LW
link
OpenAI: GPT-based LLMs show ability to discriminate between its own wrong answers, but inability to explain how/why it makes that discrimination, even as model scales
Aditya Jain
Jun 13, 2022, 11:33 PM
14
points
5
comments
1
min read
LW
link
(openai.com)
[Question]
Who said something like “The fact that putting 2 apples next to 2 other apples leads to there being 4 apples there has nothing to do with the fact that 2 + 2 = 4”?
hunterglenn
Jun 13, 2022, 10:23 PM
1
point
2
comments
1
min read
LW
link
Continuity Assumptions
Jan_Kulveit
Jun 13, 2022, 9:31 PM
44
points
13
comments
4
min read
LW
link
Crypto-fed Computation
aaguirre
Jun 13, 2022, 9:20 PM
24
points
7
comments
7
min read
LW
link
A Modest Pivotal Act
anonymousaisafety
Jun 13, 2022, 7:24 PM
−16
points
1
comment
5
min read
LW
link
Contra EY: Can AGI destroy us without trial & error?
nsokolsky
Jun 13, 2022, 6:26 PM
137
points
72
comments
15
min read
LW
link
What are some smaller-but-concrete challenges related to AI safety that are impacting people today?
nonzerosum
Jun 13, 2022, 5:36 PM
4
points
3
comments
1
min read
LW
link
[Link] New SEP article on Bayesian Epistemology
Aryeh Englander
Jun 13, 2022, 3:03 PM
6
points
0
comments
1
min read
LW
link
Training Trace Priors
Adam Jermyn
Jun 13, 2022, 2:22 PM
12
points
17
comments
4
min read
LW
link
[Question]
Can you MRI a deep learning model?
Yair Halberstadt
Jun 13, 2022, 1:43 PM
3
points
3
comments
1
min read
LW
link
On A List of Lethalities
Zvi
Jun 13, 2022, 12:30 PM
165
points
50
comments
54
min read
LW
link
1
review
(thezvi.wordpress.com)
D&D.Sci June 2022 Evaluation and Ruleset
abstractapplic
Jun 13, 2022, 10:31 AM
34
points
11
comments
4
min read
LW
link
[Question]
What’s the “This AI is of moral concern.” fire alarm?
Quintin Pope
Jun 13, 2022, 8:05 AM
37
points
56
comments
2
min read
LW
link
The beautiful magical enchanted golden Dall-e Mini is underrated
p.b.
Jun 13, 2022, 7:58 AM
14
points
0
comments
1
min read
LW
link
Why so little AI risk on rationalist-adjacent blogs?
Grant Demaree
13 Jun 2022 6:31 UTC
46
points
23
comments
8
min read
LW
link
Code Quality and Rule Consequentialism
Adam Zerner
13 Jun 2022 3:12 UTC
17
points
13
comments
6
min read
LW
link
Grokking “Semi-informative priors over AI timelines”
anson.ho
12 Jun 2022 22:17 UTC
15
points
7
comments
14
min read
LW
link
[Question]
How much does cybersecurity reduce AI risk?
Darmani
12 Jun 2022 22:13 UTC
34
points
23
comments
1
min read
LW
link
[Question]
How are compute assets distributed in the world?
Chris van Merwijk
12 Jun 2022 22:13 UTC
30
points
7
comments
1
min read
LW
link
Intuitive Explanation of AIXI
Thomas Larsen
12 Jun 2022 21:41 UTC
22
points
1
comment
5
min read
LW
link
Back to top
Next
N
W
F
A
C
D
E
F
G
H
I
Customize appearance
Current theme:
default
A
C
D
E
F
G
H
I
Less Wrong (text)
Less Wrong (link)
Invert colors
Reset to defaults
OK
Cancel
Hi, I’m Bobby the Basilisk! Click on the minimize button (
) to minimize the theme tweaker window, so that you can see what the page looks like with the current tweaked values. (But remember,
the changes won’t be saved until you click “OK”!
)
Theme tweaker help
Show Bobby the Basilisk
OK
Cancel