Get your tick­ets to Man­i­fest 2024 by May 13th!

Saul Munn3 May 2024 23:57 UTC
18 points
0 comments1 min readLW link

Embodiment

A*3 May 2024 20:06 UTC
4 points
0 comments1 min readLW link

(Geo­met­ri­cally) Max­i­mal Lot­tery-Lot­ter­ies Exist

Lorxus3 May 2024 19:29 UTC
13 points
11 comments26 min readLW link

[Question] Were there any an­cient ra­tio­nal­ists?

OliverHayman3 May 2024 18:26 UTC
11 points
3 comments1 min readLW link

Key take­aways from our EA and al­ign­ment re­search sur­veys

3 May 2024 18:10 UTC
103 points
10 comments21 min readLW link

“AI Safety for Fleshy Hu­mans” an AI Safety ex­plainer by Nicky Case

habryka3 May 2024 18:10 UTC
84 points
10 comments4 min readLW link
(aisafety.dance)

AI Clar­ity: An Ini­tial Re­search Agenda

3 May 2024 13:54 UTC
18 points
1 comment8 min readLW link

Ap­ply to ESPR & PAIR, Ra­tion­al­ity and AI Camps for Ages 16-21

Anna Gajdova3 May 2024 12:36 UTC
58 points
5 comments1 min readLW link

On pre­cise out-of-con­text steering

Olli Järviniemi3 May 2024 9:41 UTC
9 points
6 comments3 min readLW link

LLM+Plan­ners hy­bridi­s­a­tion for friendly AGI

installgentoo3 May 2024 8:40 UTC
7 points
2 comments1 min readLW link

Mechanis­tic In­ter­pretabil­ity Work­shop Hap­pen­ing at ICML 2024!

3 May 2024 1:18 UTC
48 points
6 comments1 min readLW link

Weekly newslet­ter for AI safety events and train­ing programs

Bryce Robertson3 May 2024 0:33 UTC
29 points
0 comments1 min readLW link

CCS: Coun­ter­fac­tual Civ­i­liza­tion Simulation

Morphism2 May 2024 22:54 UTC
3 points
0 comments2 min readLW link

Let’s De­sign A School, Part 2.1 School as Ed­u­ca­tion—Structure

Sable2 May 2024 22:04 UTC
26 points
2 comments10 min readLW link
(affablyevil.substack.com)

Why I’m not do­ing PauseAI

Ariel Kwiatkowski2 May 2024 22:00 UTC
−7 points
5 comments4 min readLW link

AI #61: Meta Trouble

Zvi2 May 2024 18:40 UTC
29 points
0 comments52 min readLW link
(thezvi.wordpress.com)

[Question] How would you nav­i­gate a se­vere fi­nan­cial emer­gency with no help or re­sources?

Tigerlily2 May 2024 18:27 UTC
38 points
22 comments2 min readLW link

Why is AGI/​ASI Inevitable?

DeathlessAmaranth2 May 2024 18:27 UTC
14 points
6 comments1 min readLW link

AISN #34: New Mili­tary AI Sys­tems Plus, AI Labs Fail to Uphold Vol­un­tary Com­mit­ments to UK AI Safety In­sti­tute, and New AI Policy Pro­pos­als in the US Senate

2 May 2024 16:12 UTC
6 points
0 comments8 min readLW link
(newsletter.safe.ai)

Ai Salon: Trust­wor­thy AI Fu­tures #1

Ian Eisenberg2 May 2024 16:07 UTC
1 point
0 comments1 min readLW link

How to write Pseu­docode and why you should

Johannes C. Mayer2 May 2024 15:53 UTC
8 points
5 comments3 min readLW link

AI #62: Too Soon to Tell

Zvi2 May 2024 15:40 UTC
30 points
8 comments31 min readLW link
(thezvi.wordpress.com)

White­board Pro­gram Tra­ce­ing: De­bug a Pro­gram Be­fore you have the Code

Johannes C. Mayer2 May 2024 15:30 UTC
3 points
0 comments1 min readLW link

[Question] Which skin­care prod­ucts are ev­i­dence-based?

Vanessa Kosoy2 May 2024 15:22 UTC
117 points
47 comments1 min readLW link

Q&A on Pro­posed SB 1047

Zvi2 May 2024 15:10 UTC
74 points
8 comments44 min readLW link
(thezvi.wordpress.com)

[Question] What are the Ac­tivi­ties that make up your Re­search Pro­cess?

Johannes C. Mayer2 May 2024 15:01 UTC
4 points
0 comments1 min readLW link

Please stop pub­lish­ing ideas/​in­sights/​re­search about AI

Tamsin Leake2 May 2024 14:54 UTC
0 points
61 comments4 min readLW link

[Question] How do you Select the Right Re­search Ac­i­tivity in the Right Mo­ment?

Johannes C. Mayer2 May 2024 14:45 UTC
6 points
1 comment1 min readLW link

[Question] Can stealth air­craft be de­tected op­ti­cally?

Yair Halberstadt2 May 2024 7:47 UTC
18 points
25 comments1 min readLW link

An ex­pla­na­tion of evil in an or­ga­nized world

KatjaGrace2 May 2024 5:20 UTC
28 points
9 comments2 min readLW link
(worldspiritsockpuppet.com)

Why I stopped work­ing on AI safety

jbkjr2 May 2024 5:08 UTC
−5 points
0 comments4 min readLW link
(jbkjr.me)

[Linkpost] Silver Bul­letin: For most peo­ple, poli­tics is about fit­ting in

Gunnar_Zarncke1 May 2024 18:12 UTC
18 points
4 comments1 min readLW link
(www.natesilver.net)

Launch­ing ap­pli­ca­tions for AI Safety Ca­reers Course In­dia 2024

Axiom_Futures1 May 2024 17:55 UTC
4 points
1 comment1 min readLW link

[Question] Shane Legg’s nec­es­sary prop­er­ties for ev­ery AGI Safety plan

jacquesthibs1 May 2024 17:15 UTC
58 points
12 comments1 min readLW link

KAN: Kol­mogorov-Arnold Networks

Gunnar_Zarncke1 May 2024 16:50 UTC
18 points
15 comments1 min readLW link
(arxiv.org)

Man­i­fund Q1 Retro: Learn­ings from im­pact certs

Austin Chen1 May 2024 16:48 UTC
40 points
1 comment1 min readLW link

ACX Covid Ori­gins Post con­vinced readers

ErnestScribbler1 May 2024 13:06 UTC
77 points
7 comments2 min readLW link

LessWrong Com­mu­nity Week­end 2024, open for applications

1 May 2024 10:18 UTC
79 points
2 comments7 min readLW link

Take SCIFs, it’s dan­ger­ous to go alone

1 May 2024 8:02 UTC
42 points
1 comment3 min readLW link

AXRP Epi­sode 30 - AI Se­cu­rity with Jeffrey Ladish

DanielFilan1 May 2024 2:50 UTC
25 points
0 comments79 min readLW link

Neuro/​BCI/​WBE for Safe AI Workshop

Allison Duettmann1 May 2024 0:46 UTC
3 points
0 comments1 min readLW link

AGI: Cryp­tog­ra­phy, Se­cu­rity & Mul­tipo­lar Sce­nar­ios Workshop

Allison Duettmann1 May 2024 0:42 UTC
8 points
1 comment1 min readLW link

The for­mal goal is a pointer

Morphism1 May 2024 0:27 UTC
20 points
10 comments1 min readLW link

Arch-an­ar­chy:The­ory and practice

Peter lawless 30 Apr 2024 23:20 UTC
−5 points
0 comments2 min readLW link

“Open Source AI” is a lie, but it doesn’t have to be

jacobhaimes30 Apr 2024 23:10 UTC
18 points
5 comments6 min readLW link
(jacob-haimes.github.io)

Ques­tions for labs

Zach Stein-Perlman30 Apr 2024 22:15 UTC
77 points
11 comments8 min readLW link

Real­ity com­pre­hen­si­bil­ity: are there illog­i­cal things in re­al­ity?

DDthinker30 Apr 2024 21:30 UTC
−3 points
0 comments10 min readLW link

Mechanis­ti­cally Elic­it­ing La­tent Be­hav­iors in Lan­guage Models

30 Apr 2024 18:51 UTC
204 points
40 comments45 min readLW link

[Question] What is the eas­iest/​funnest way to build up a com­pre­hen­sive un­der­stand­ing of AI and AI Safety?

Jordan Arel30 Apr 2024 18:41 UTC
4 points
2 comments1 min readLW link

Transcoders en­able fine-grained in­ter­pretable cir­cuit anal­y­sis for lan­guage models

30 Apr 2024 17:58 UTC
69 points
14 comments17 min readLW link