Hiring de­ci­sions are not suit­able for pre­dic­tion markets

SimonM8 Jan 2024 21:11 UTC
12 points
6 comments1 min readLW link

Bet­ter Anomia

jefftk8 Jan 2024 18:40 UTC
8 points
0 comments1 min readLW link
(www.jefftk.com)

A starter guide for evals

8 Jan 2024 18:24 UTC
51 points
2 comments12 min readLW link
(www.apolloresearch.ai)

Is it jus­tifi­able for non-ex­perts to have strong opinions about Gaza?

8 Jan 2024 17:31 UTC
23 points
12 comments30 min readLW link

Pro­ject ideas: Backup plans & Co­op­er­a­tive AI

Lukas Finnveden8 Jan 2024 17:19 UTC
18 points
0 comments1 min readLW link
(lukasfinnveden.substack.com)

Hackathon and Stay­ing Up-to-Date in AI

jacobhaimes8 Jan 2024 17:10 UTC
11 points
0 comments1 min readLW link
(into-ai-safety.github.io)

When “yang” goes wrong

Joe Carlsmith8 Jan 2024 16:35 UTC
72 points
6 comments13 min readLW link

Task vec­tors & anal­ogy mak­ing in LLMs

Sergii8 Jan 2024 15:17 UTC
9 points
1 comment4 min readLW link
(grgv.xyz)

[Question] How to find trans­la­tions of a book?

Viliam8 Jan 2024 14:57 UTC
9 points
8 comments1 min readLW link

[Question] Why aren’t Yud­kowsky & Bostrom get­ting more at­ten­tion now?

JoshuaFox8 Jan 2024 14:42 UTC
14 points
8 comments1 min readLW link

2023 Pre­dic­tion Evaluations

Zvi8 Jan 2024 14:40 UTC
47 points
0 comments28 min readLW link
(thezvi.wordpress.com)

There is no sharp bound­ary be­tween de­on­tol­ogy and consequentialism

quetzal_rainbow8 Jan 2024 11:01 UTC
8 points
2 comments1 min readLW link

Reflec­tions on my first year of AI safety research

Jay Bailey8 Jan 2024 7:49 UTC
53 points
3 comments1 min readLW link

Why There Is Hope For An Align­ment Solution

Darklight8 Jan 2024 6:58 UTC
10 points
0 comments12 min readLW link

Sled­ding Among Hazards

jefftk8 Jan 2024 3:30 UTC
19 points
5 comments1 min readLW link
(www.jefftk.com)

Utility is relative

CrimsonChin8 Jan 2024 2:31 UTC
2 points
4 comments2 min readLW link

A model of re­search skill

L Rudolf L8 Jan 2024 0:13 UTC
60 points
6 comments12 min readLW link
(www.strataoftheworld.com)

We shouldn’t fear su­per­in­tel­li­gence be­cause it already exists

Spencer Chubb7 Jan 2024 17:59 UTC
−22 points
14 comments1 min readLW link

(Par­tial) failure in repli­cat­ing de­cep­tive al­ign­ment experiment

claudia.biancotti7 Jan 2024 17:56 UTC
1 point
0 comments1 min readLW link

Pro­ject ideas: Sen­tience and rights of digi­tal minds

Lukas Finnveden7 Jan 2024 17:34 UTC
20 points
0 comments1 min readLW link
(lukasfinnveden.substack.com)

De­cep­tive AI ≠ De­cep­tively-al­igned AI

Steven Byrnes7 Jan 2024 16:55 UTC
96 points
19 comments6 min readLW link

Bayesi­ans Com­mit the Gam­bler’s Fallacy

Kevin Dorst7 Jan 2024 12:54 UTC
48 points
30 comments8 min readLW link
(kevindorst.substack.com)

Towards AI Safety In­fras­truc­ture: Talk & Outline

Paul Bricman7 Jan 2024 9:31 UTC
11 points
0 comments2 min readLW link
(www.youtube.com)

Defend­ing against hy­po­thet­i­cal moon life dur­ing Apollo 11

eukaryote7 Jan 2024 4:49 UTC
57 points
9 comments32 min readLW link
(eukaryotewritesblog.com)

The Se­quences on YouTube

Neil 7 Jan 2024 1:44 UTC
26 points
9 comments2 min readLW link

AI Risk and the US Pres­i­den­tial Candidates

Zane6 Jan 2024 20:18 UTC
41 points
22 comments6 min readLW link

A Challenge to Effec­tive Altru­ism’s Premises

False Name6 Jan 2024 18:46 UTC
−26 points
3 comments3 min readLW link

Lack of Spi­der-Man is ev­i­dence against the simu­la­tion hypothesis

RamblinDash6 Jan 2024 18:17 UTC
7 points
22 comments1 min readLW link

A Land Tax For Britain

A.H.6 Jan 2024 15:52 UTC
6 points
9 comments4 min readLW link

Book re­view: Trick or treat­ment (2008)

Fleece Minutia6 Jan 2024 15:40 UTC
1 point
0 comments2 min readLW link

Are we in­side a black hole?

Jay6 Jan 2024 13:30 UTC
2 points
5 comments1 min readLW link

Sur­vey of 2,778 AI au­thors: six parts in pictures

KatjaGrace6 Jan 2024 4:43 UTC
80 points
1 comment2 min readLW link

Pro­ject ideas: Epistemics

Lukas Finnveden5 Jan 2024 23:41 UTC
43 points
4 comments1 min readLW link
(lukasfinnveden.substack.com)

Al­most ev­ery­one I’ve met would be well-served think­ing more about what to fo­cus on

Henrik Karlsson5 Jan 2024 21:01 UTC
96 points
8 comments11 min readLW link
(www.henrikkarlsson.xyz)

The Next ChatGPT Mo­ment: AI Avatars

5 Jan 2024 20:14 UTC
43 points
10 comments1 min readLW link

AI Im­pacts 2023 Ex­pert Sur­vey on Progress in AI

habryka5 Jan 2024 19:42 UTC
28 points
2 comments7 min readLW link
(wiki.aiimpacts.org)

Tech­nol­ogy path de­pen­dence and eval­u­at­ing expertise

5 Jan 2024 19:21 UTC
24 points
2 comments15 min readLW link

The Hip­pie Rab­bit Hole -Nuggets of Gold in Rivers of Bullshit

Jonathan Moregård5 Jan 2024 18:27 UTC
38 points
20 comments8 min readLW link
(honestliving.substack.com)

[Question] What tech­ni­cal top­ics could help with bound­aries/​mem­branes?

Chipmonk5 Jan 2024 18:14 UTC
15 points
25 comments1 min readLW link

Catch­ing AIs red-handed

5 Jan 2024 17:43 UTC
110 points
27 comments17 min readLW link

AI Im­pacts Sur­vey: De­cem­ber 2023 Edition

Zvi5 Jan 2024 14:40 UTC
34 points
6 comments10 min readLW link
(thezvi.wordpress.com)

Fore­cast your 2024 with Fatebook

Sage Future5 Jan 2024 14:07 UTC
19 points
0 comments1 min readLW link
(fatebook.io)

Pre­dic­tive model agents are sort of corrigible

Raymond D5 Jan 2024 14:05 UTC
35 points
6 comments3 min readLW link

Strik­ing Im­pli­ca­tions for Learn­ing The­ory, In­ter­pretabil­ity — and Safety?

RogerDearnaley5 Jan 2024 8:46 UTC
37 points
4 comments2 min readLW link

If I ran the zoo

Optimization Process5 Jan 2024 5:14 UTC
18 points
0 comments2 min readLW link

Does AI care about re­al­ity or just its own per­cep­tion?

RedFishBlueFish5 Jan 2024 4:05 UTC
−6 points
8 comments1 min readLW link

MIRI 2024 Mis­sion and Strat­egy Update

Malo5 Jan 2024 0:20 UTC
222 points
44 comments8 min readLW link

Pro­ject ideas: Gover­nance dur­ing ex­plo­sive tech­nolog­i­cal growth

Lukas Finnveden4 Jan 2024 23:51 UTC
14 points
0 comments1 min readLW link
(lukasfinnveden.substack.com)

Hello

S Benfield4 Jan 2024 23:35 UTC
6 points
0 comments2 min readLW link

Us­ing Threats to Achieve So­cially Op­ti­mal Outcomes

StrivingForLegibility4 Jan 2024 23:30 UTC
8 points
0 comments3 min readLW link