RSS

Lan­guage Models Model Us

eggsyntax17 May 2024 21:00 UTC
81 points
19 comments7 min readLW link

In­struc­tion-fol­low­ing AGI is eas­ier and more likely than value al­igned AGI

Seth Herd15 May 2024 19:38 UTC
35 points
18 comments12 min readLW link

Ilya Sutskever and Jan Leike re­sign from OpenAI [up­dated]

Zach Stein-Perlman15 May 2024 0:45 UTC
230 points
84 comments2 min readLW link

Deep­Mind’s “​​Fron­tier Safety Frame­work” is weak and unambitious

Zach Stein-Perlman18 May 2024 3:00 UTC
118 points
10 comments4 min readLW link

Us­ing GPT-3 for pre­vent­ing con­flict dur­ing mes­sag­ing — a pitch for an app

Eli_17 Mar 2022 11:02 UTC
22 points
17 comments3 min readLW link

Towards Guaran­teed Safe AI: A Frame­work for En­sur­ing Ro­bust and Reli­able AI Systems

Joar Skalse17 May 2024 19:13 UTC
47 points
2 comments2 min readLW link

“If we go ex­tinct due to mis­al­igned AI, at least na­ture will con­tinue, right? … right?”

plex18 May 2024 14:09 UTC
40 points
9 comments2 min readLW link
(aisafety.info)

Scien­tific No­ta­tion Options

jefftk18 May 2024 15:10 UTC
19 points
7 comments1 min readLW link
(www.jefftk.com)

[Question] In the con­text of AI in­terp. What is a fea­ture ex­actly?

f3mi14 May 2024 13:46 UTC
9 points
1 comment1 min readLW link

[Question] Is acausal ex­tor­tion pos­si­ble?

sisyphus11 Nov 2022 19:48 UTC
−20 points
36 comments3 min readLW link

Do you be­lieve in hun­dred dol­lar bills ly­ing on the ground? Con­sider humming

Elizabeth16 May 2024 0:00 UTC
128 points
11 comments6 min readLW link
(acesounderglass.com)

Ein­stein’s Arrogance

Eliezer Yudkowsky25 Sep 2007 1:29 UTC
155 points
90 comments3 min readLW link

Why you should learn a mu­si­cal instrument

cata15 May 2024 20:36 UTC
48 points
23 comments3 min readLW link

How much AI in­fer­ence can we do?

Benjamin_Todd14 May 2024 15:10 UTC
16 points
6 comments5 min readLW link
(benjamintodd.substack.com)

A Dozen Ways to Get More Dakka

Davidmanheim8 Apr 2024 4:45 UTC
97 points
9 comments3 min readLW link

What Are Non-Zero-Sum Games?—A Primer

James Stephen Brown18 May 2024 9:19 UTC
4 points
1 comment3 min readLW link

[Cross­post] In­tro­duc­ing the Save State Paradox

Suzie. EXE18 May 2024 17:00 UTC
1 point
0 comments7 min readLW link

Ad­vice for Ac­tivists from the His­tory of Environmentalism

Jeffrey Heninger16 May 2024 18:40 UTC
75 points
5 comments6 min readLW link
(blog.aiimpacts.org)

Davi­dad’s Bold Plan for Align­ment: An In-Depth Explanation

19 Apr 2023 16:09 UTC
154 points
32 comments21 min readLW link

What Do We Mean By “Ra­tion­al­ity”?

Eliezer Yudkowsky16 Mar 2009 22:33 UTC
333 points
18 comments6 min readLW link