[Question] Monotonous Work

Gideon Bauer2 Feb 2023 21:35 UTC
1 point
0 comments1 min readLW link

Is AI risk as­sess­ment too an­thro­pocen­tric?

Craig Mattson2 Feb 2023 21:34 UTC
3 points
6 comments1 min readLW link

Hal­i­fax Monthly Meetup: In­tro­duc­tion to Effec­tive Altruism

Ideopunk2 Feb 2023 21:10 UTC
10 points
0 comments1 min readLW link

Con­di­tion­ing Pre­dic­tive Models: Outer al­ign­ment via care­ful conditioning

2 Feb 2023 20:28 UTC
72 points
15 comments57 min readLW link

Con­di­tion­ing Pre­dic­tive Models: Large lan­guage mod­els as predictors

2 Feb 2023 20:28 UTC
88 points
4 comments13 min readLW link

Nor­ma­tive vs De­scrip­tive Models of Agency

mattmacdermott2 Feb 2023 20:28 UTC
26 points
5 comments4 min readLW link

An­drew Hu­ber­man on How to Op­ti­mize Sleep

Leon Lang2 Feb 2023 20:17 UTC
37 points
6 comments6 min readLW link

[Question] How can I help in­flam­ma­tion-based nerve dam­age be tem­po­rary?

Optimization Process2 Feb 2023 19:20 UTC
17 points
4 comments1 min readLW link

More find­ings on max­i­mal data dimension

Marius Hobbhahn2 Feb 2023 18:33 UTC
27 points
1 comment11 min readLW link

Her­i­ta­bil­ity, Be­hav­iorism, and Within-Life­time RL

Steven Byrnes2 Feb 2023 16:34 UTC
39 points
3 comments4 min readLW link

Covid 2/​2/​23: The Emer­gency Ends on 5/​11

Zvi2 Feb 2023 14:00 UTC
22 points
6 comments7 min readLW link
(thezvi.wordpress.com)

You are prob­a­bly not a good al­ign­ment re­searcher, and other blatant lies

junk heap homotopy2 Feb 2023 13:55 UTC
83 points
16 comments2 min readLW link

Don’t Judge a Tool by its Aver­age Output

silentbob2 Feb 2023 13:42 UTC
5 points
2 comments4 min readLW link

Epoch Im­pact Re­port 2022

Jsevillamol2 Feb 2023 13:09 UTC
16 points
0 comments1 min readLW link

You Don’t Ex­ist, Duncan

Duncan Sabien (Deactivated)2 Feb 2023 8:37 UTC
244 points
107 comments9 min readLW link

Tem­po­rally Lay­ered Ar­chi­tec­ture for Adap­tive, Distributed and Con­tin­u­ous Control

Roman Leventov2 Feb 2023 6:29 UTC
6 points
4 comments1 min readLW link
(arxiv.org)

Re­search agenda: For­mal­iz­ing ab­strac­tions of computations

Erik Jenner2 Feb 2023 4:29 UTC
92 points
10 comments31 min readLW link

Progress links and tweets, 2023-02-01

jasoncrawford2 Feb 2023 2:25 UTC
10 points
0 comments1 min readLW link
(rootsofprogress.org)

Ret­ro­spec­tive on the AI Safety Field Build­ing Hub

Vael Gates2 Feb 2023 2:06 UTC
30 points
0 comments1 min readLW link

How to ex­port An­droid Chrome tabs to an HTML file in Linux (as of Fe­bru­ary 2023)

Adam Scherlis2 Feb 2023 2:03 UTC
7 points
3 comments2 min readLW link
(adam.scherlis.com)

Hacked Ac­count Spam

jefftk2 Feb 2023 1:50 UTC
13 points
5 comments1 min readLW link
(www.jefftk.com)

A sim­ple tech­nique to re­duce nega­tive rumination

cranberry_bear2 Feb 2023 1:33 UTC
9 points
0 comments1 min readLW link

A Brief Overview of AI Safety/​Align­ment Orgs, Fields, Re­searchers, and Re­sources for ML Researchers

Austin Witte2 Feb 2023 1:02 UTC
18 points
1 comment2 min readLW link

In­ter­views with 97 AI Re­searchers: Quan­ti­ta­tive Analysis

2 Feb 2023 1:01 UTC
23 points
0 comments7 min readLW link

“AI Risk Dis­cus­sions” web­site: Ex­plor­ing in­ter­views from 97 AI Researchers

2 Feb 2023 1:00 UTC
43 points
1 comment1 min readLW link

Pre­dict­ing re­searcher in­ter­est in AI alignment

Vael Gates2 Feb 2023 0:58 UTC
25 points
0 comments1 min readLW link

Fo­cus on the places where you feel shocked ev­ery­one’s drop­ping the ball

So8res2 Feb 2023 0:27 UTC
421 points
60 comments4 min readLW link

Ex­er­cise is Good, Actually

Gordon Seidoh Worley2 Feb 2023 0:09 UTC
91 points
27 comments3 min readLW link

Product safety is a poor model for AI governance

Richard Korzekwa 1 Feb 2023 22:40 UTC
36 points
0 comments5 min readLW link
(aiimpacts.org)

Hin­ton: “mor­tal” effi­cient ana­log hard­ware may be learned-in-place, uncopyable

the gears to ascension1 Feb 2023 22:19 UTC
12 points
3 comments1 min readLW link

Can we “cure” can­cer?

jasoncrawford1 Feb 2023 22:03 UTC
41 points
31 comments2 min readLW link
(rootsofprogress.org)

Eli Lifland on Nav­i­gat­ing the AI Align­ment Landscape

ozziegooen1 Feb 2023 21:17 UTC
9 points
1 comment31 min readLW link
(quri.substack.com)

Schizophre­nia as a defi­ciency in long-range cor­tex-to-cor­tex communication

Steven Byrnes1 Feb 2023 19:32 UTC
35 points
36 comments11 min readLW link

AI Safety Ar­gu­ments: An In­ter­ac­tive Guide

Lukas Trötzmüller1 Feb 2023 19:26 UTC
20 points
0 comments3 min readLW link

More find­ings on Me­moriza­tion and dou­ble descent

Marius Hobbhahn1 Feb 2023 18:26 UTC
53 points
2 comments19 min readLW link

Lan­guage Models can be Utility-Max­imis­ing Agents

Raymond D1 Feb 2023 18:13 UTC
22 points
1 comment2 min readLW link

Trends in the dol­lar train­ing cost of ma­chine learn­ing systems

Ben Cottier1 Feb 2023 14:48 UTC
23 points
0 comments2 min readLW link
(epochai.org)

Po­lis: Why and How to Use it

brook1 Feb 2023 14:03 UTC
5 points
0 comments1 min readLW link

Su­biti­sa­tion of Self

vitaliya1 Feb 2023 9:18 UTC
14 points
4 comments2 min readLW link

Directed Babbling

Yudhister Kumar1 Feb 2023 9:10 UTC
20 points
1 comment3 min readLW link
(www.ykumar.org)

Vot­ing Re­sults for the 2021 Review

Raemon1 Feb 2023 8:02 UTC
66 points
10 comments38 min readLW link

Ab­strac­tion As Sym­me­try and Other Thoughts

Numendil1 Feb 2023 6:25 UTC
28 points
9 comments2 min readLW link

The effect of hori­zon length on scal­ing laws

Jacob_Hilton1 Feb 2023 3:59 UTC
23 points
2 comments1 min readLW link
(arxiv.org)

Con­tra Dance Lengths

jefftk1 Feb 2023 3:30 UTC
9 points
0 comments1 min readLW link
(www.jefftk.com)

Aiming for Con­ver­gence Is Like Dis­cour­ag­ing Betting

Zack_M_Davis1 Feb 2023 0:03 UTC
60 points
17 comments11 min readLW link

On value in hu­mans, other an­i­mals, and AI

Michele Campolo31 Jan 2023 23:33 UTC
3 points
17 comments5 min readLW link

Crit­i­cism of the main frame­work in AI alignment

Michele Campolo31 Jan 2023 23:01 UTC
19 points
2 comments6 min readLW link

Nice Clothes are Good, Actually

Gordon Seidoh Worley31 Jan 2023 19:22 UTC
71 points
28 comments4 min readLW link

[Linkpost] Hu­man-nar­rated au­dio ver­sion of “Is Power-Seek­ing AI an Ex­is­ten­tial Risk?”

Joe Carlsmith31 Jan 2023 19:21 UTC
12 points
1 comment1 min readLW link

No Really, At­ten­tion is ALL You Need—At­ten­tion can do feed­for­ward networks

Robert_AIZI31 Jan 2023 18:48 UTC
29 points
7 comments6 min readLW link
(aizi.substack.com)