Questions
Events
Shortform
Alignment Forum
AF Comments
Home
Featured
All
Tags
Recent
Comments
Archive
Sequences
About
Search
Log In
All
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
All
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
All
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
Page
1
[Question]
Is there a fundamental distinction between simulating a mind and simulating *being* a mind? Is this a useful and important distinction?
Thoth Hermes
8 Apr 2023 23:44 UTC
−17
points
8
comments
2
min read
LW
link
“warning about ai doom” is also “announcing capabilities progress to noobs”
the gears to ascension
8 Apr 2023 23:42 UTC
23
points
5
comments
3
min read
LW
link
Feature Request: Right Click to Copy LaTeX
DragonGod
8 Apr 2023 23:27 UTC
18
points
4
comments
1
min read
LW
link
ELCK might require nontrivial scalable alignment progress, and seems tractable enough to try
Alex Lawsen
8 Apr 2023 21:49 UTC
17
points
0
comments
2
min read
LW
link
GPTs are Predictors, not Imitators
Eliezer Yudkowsky
8 Apr 2023 19:59 UTC
409
points
99
comments
3
min read
LW
link
3
reviews
4 generations of alignment
qbolec
8 Apr 2023 19:59 UTC
1
point
0
comments
3
min read
LW
link
The surprising parameter efficiency of vision models
beren
8 Apr 2023 19:44 UTC
81
points
28
comments
4
min read
LW
link
Random Observation on AI goals
FTPickle
8 Apr 2023 19:28 UTC
−11
points
2
comments
1
min read
LW
link
Can we evaluate the “tool versus agent” AGI prediction?
Xodarap
8 Apr 2023 18:40 UTC
16
points
7
comments
1
min read
LW
link
Relative Abstracted Agency
Audere
8 Apr 2023 16:57 UTC
14
points
6
comments
5
min read
LW
link
The benevolence of the butcher
dr_s
8 Apr 2023 16:29 UTC
72
points
30
comments
6
min read
LW
link
SERI MATS—Summer 2023 Cohort
Aris
,
Ryan Kidd
and
Christian Smith
8 Apr 2023 15:32 UTC
71
points
25
comments
4
min read
LW
link
AI Proposals at ‘Two Sessions’: AGI as ‘Two Bombs, One Satellite’?
Derek M. Jones
8 Apr 2023 11:31 UTC
5
points
0
comments
1
min read
LW
link
(www.chinatalk.media)
All images from the WaitButWhy sequence on AI
trevor
8 Apr 2023 7:36 UTC
73
points
5
comments
2
min read
LW
link
Guidelines for productive discussions
ambigram
8 Apr 2023 6:00 UTC
37
points
0
comments
5
min read
LW
link
All AGI Safety questions welcome (especially basic ones) [April 2023]
steven0461
8 Apr 2023 4:21 UTC
57
points
88
comments
2
min read
LW
link
Bringing Agency Into AGI Extinction Is Superfluous
George3d6
8 Apr 2023 4:02 UTC
28
points
18
comments
5
min read
LW
link
Lagos, Nigeria—ACX Meetups Everywhere 2023
damola
8 Apr 2023 3:55 UTC
1
point
0
comments
1
min read
LW
link
Upcoming Changes in Large Language Models
Andrew Keenan Richardson
8 Apr 2023 3:41 UTC
43
points
8
comments
4
min read
LW
link
(mechanisticmind.com)
Consider The Hand Axe
ymeskhout
8 Apr 2023 1:31 UTC
142
points
16
comments
6
min read
LW
link
AGI as a new data point
Will Rodgers
8 Apr 2023 1:01 UTC
−1
points
0
comments
1
min read
LW
link
Parametrize Priority Evaluations
SilverFlame
8 Apr 2023 0:39 UTC
2
points
2
comments
6
min read
LW
link
Pausing AI Developments Isn’t Enough. We Need to Shut it All Down
Eliezer Yudkowsky
8 Apr 2023 0:36 UTC
254
points
40
comments
12
min read
LW
link
Humanitarian Phase Transition needed before Technological Singularity
Dr_What
7 Apr 2023 23:17 UTC
−9
points
5
comments
2
min read
LW
link
[Question]
Thoughts about Hugging Face?
Ariel Kwiatkowski
7 Apr 2023 23:17 UTC
7
points
0
comments
1
min read
LW
link
[Question]
Is it correct to frame alignment as “programming a good philosophy of meaning”?
Util
7 Apr 2023 23:16 UTC
2
points
3
comments
1
min read
LW
link
Select Agent Specifications as Natural Abstractions
lukemarks
7 Apr 2023 23:16 UTC
19
points
3
comments
5
min read
LW
link
n=3 AI Risk Quick Math and Reasoning
lionhearted (Sebastian Marshall)
7 Apr 2023 20:27 UTC
6
points
3
comments
4
min read
LW
link
[Question]
What are good alternatives to Predictionbook for personal prediction tracking? Edited: I originally thought it was down but it was just 500 until I though of clearing cookies.
sortega
7 Apr 2023 19:18 UTC
4
points
4
comments
1
min read
LW
link
Environments for Measuring Deception, Resource Acquisition, and Ethical Violations
Dan H
7 Apr 2023 18:40 UTC
51
points
2
comments
2
min read
LW
link
(arxiv.org)
Superintelligence Is Not Omniscience
Jeffrey Heninger
7 Apr 2023 16:30 UTC
15
points
20
comments
7
min read
LW
link
(aiimpacts.org)
An ‘AGI Emergency Eject Criteria’ consensus could be really useful.
tcelferact
7 Apr 2023 16:21 UTC
5
points
0
comments
1
min read
LW
link
Reliability, Security, and AI risk: Notes from infosec textbook chapter 1
Akash
7 Apr 2023 15:47 UTC
34
points
1
comment
4
min read
LW
link
Pre-registering a study
Robert_AIZI
7 Apr 2023 15:46 UTC
10
points
0
comments
6
min read
LW
link
(aizi.substack.com)
Live discussion at Eastercon
Douglas_Reay
7 Apr 2023 15:25 UTC
5
points
0
comments
1
min read
LW
link
[Question]
ChatGTP “Writing ” News Stories for The Guardian?
jmh
7 Apr 2023 12:16 UTC
1
point
4
comments
1
min read
LW
link
Storyteller’s convention, 2223 A.D.
plex
7 Apr 2023 11:54 UTC
8
points
0
comments
2
min read
LW
link
Stampy’s AI Safety Info—New Distillations #1 [March 2023]
markov
7 Apr 2023 11:06 UTC
42
points
0
comments
2
min read
LW
link
(aisafety.info)
Beren’s “Deconfusing Direct vs Amortised Optimisation”
DragonGod
7 Apr 2023 8:57 UTC
52
points
10
comments
3
min read
LW
link
Goal alignment without alignment on epistemology, ethics, and science is futile
Roman Leventov
7 Apr 2023 8:22 UTC
20
points
2
comments
2
min read
LW
link
Polio Lab Leak Caught with Wastewater Sampling
Cullen
7 Apr 2023 1:06 UTC
82
points
3
comments
1
min read
LW
link
Catching the Eye of Sauron
Casey B.
7 Apr 2023 0:40 UTC
221
points
68
comments
4
min read
LW
link
[Question]
How to parallelize “inherently” serial theory work?
Nicholas / Heather Kross
7 Apr 2023 0:08 UTC
16
points
6
comments
1
min read
LW
link
If Alignment is Hard, then so is Self-Improvement
PavleMiha
7 Apr 2023 0:08 UTC
21
points
20
comments
1
min read
LW
link
Anthropic is further accelerating the Arms Race?
sapphire
6 Apr 2023 23:29 UTC
82
points
22
comments
1
min read
LW
link
(techcrunch.com)
Suggestion for safe AI structure (Curated Transparent Decisions)
Kane Gregory
6 Apr 2023 22:00 UTC
5
points
6
comments
3
min read
LW
link
10 reasons why lists of 10 reasons might be a winning strategy
trevor
6 Apr 2023 21:24 UTC
109
points
7
comments
1
min read
LW
link
A Defense of Utilitarianism
Pareto Optimal
6 Apr 2023 21:09 UTC
−3
points
2
comments
5
min read
LW
link
(paretooptimal.substack.com)
One Does Not Simply Replace the Humans
JerkyTreats
6 Apr 2023 20:56 UTC
9
points
3
comments
4
min read
LW
link
(www.lesswrong.com)
[Question]
Where to begin in ML/AI?
Jake the Student
6 Apr 2023 20:45 UTC
9
points
4
comments
1
min read
LW
link
Back to top
Next