RSS

Deep­Mind’s “​​Fron­tier Safety Frame­work” is weak and unambitious

Zach Stein-Perlman18 May 2024 3:00 UTC
114 points
9 comments4 min readLW link

“If we go ex­tinct due to mis­al­igned AI, at least na­ture will con­tinue, right? … right?”

plex18 May 2024 14:09 UTC
39 points
7 comments2 min readLW link
(aisafety.info)

Lan­guage Models Model Us

eggsyntax17 May 2024 21:00 UTC
79 points
18 comments7 min readLW link

Scien­tific No­ta­tion Options

jefftk18 May 2024 15:10 UTC
18 points
5 comments1 min readLW link
(www.jefftk.com)

AISafety.com – Re­sources for AI Safety

17 May 2024 15:57 UTC
65 points
2 comments1 min readLW link

Deep­Mind: Fron­tier Safety Framework

Zach Stein-Perlman17 May 2024 17:30 UTC
60 points
0 comments3 min readLW link
(deepmind.google)

Ilya Sutskever and Jan Leike re­sign from OpenAI [up­dated]

Zach Stein-Perlman15 May 2024 0:45 UTC
227 points
81 comments2 min readLW link

Towards Guaran­teed Safe AI: A Frame­work for En­sur­ing Ro­bust and Reli­able AI Systems

Joar Skalse17 May 2024 19:13 UTC
45 points
1 comment2 min readLW link

Iden­ti­fy­ing Func­tion­ally Im­por­tant Fea­tures with End-to-End Sparse Dic­tionary Learning

17 May 2024 16:25 UTC
49 points
2 comments4 min readLW link
(publications.apolloresearch.ai)

Do you be­lieve in hun­dred dol­lar bills ly­ing on the ground? Con­sider humming

Elizabeth16 May 2024 0:00 UTC
127 points
11 comments6 min readLW link
(acesounderglass.com)

[Cross­post] In­tro­duc­ing the Save State Paradox

Suzie. EXE18 May 2024 17:00 UTC
1 point
0 comments7 min readLW link

In­ter­na­tional Scien­tific Re­port on the Safety of Ad­vanced AI: Key Information

Aryeh Englander18 May 2024 1:45 UTC
23 points
0 comments13 min readLW link

Ad­vice for Ac­tivists from the His­tory of Environmentalism

Jeffrey Heninger16 May 2024 18:40 UTC
73 points
5 comments6 min readLW link
(blog.aiimpacts.org)

What Are Non-Zero-Sum Games?—A Primer

James Stephen Brown18 May 2024 9:19 UTC
4 points
1 comment3 min readLW link

Towards Guaran­teed Safe AI: A Frame­work for En­sur­ing Ro­bust and Reli­able AI Systems

Gunnar_Zarncke16 May 2024 13:09 UTC
50 points
4 comments1 min readLW link
(arxiv.org)

Good­hart in RL with KL: Appendix

Thomas Kwa18 May 2024 0:40 UTC
11 points
0 comments6 min readLW link

Is There Really a Child Penalty in the Long Run?

Maxwell Tabarrok17 May 2024 11:56 UTC
23 points
5 comments5 min readLW link
(www.maximum-progress.com)

Teach­ing CS Dur­ing Take-Off

andrew carle14 May 2024 22:45 UTC
79 points
10 comments2 min readLW link

The Dun­ning-Kruger of dis­prov­ing Dun­ning-Kruger

kromem16 May 2024 10:11 UTC
41 points
0 comments5 min readLW link

D&D.Sci (Easy Mode): On The Con­struc­tion Of Im­pos­si­ble Structures

abstractapplic17 May 2024 0:25 UTC
28 points
10 comments2 min readLW link