RSS

Christopher King

Karma: 805

@theking@mathstodon.xyz

METR’s pre­limi­nary eval­u­a­tion of o3 and o4-mini

Christopher KingApr 16, 2025, 8:23 PM
13 points
4 comments1 min readLW link
(metr.github.io)

[Question] How far along Metr’s law can AI start au­tomat­ing or helping with al­ign­ment re­search?

Christopher KingMar 20, 2025, 3:58 PM
20 points
21 comments1 min readLW link

No, the Poly­mar­ket price does not mean we can im­me­di­ately con­clude what the prob­a­bil­ity of a bird flu pan­demic is. We also need to know the in­ter­est rate!

Christopher KingDec 28, 2024, 4:05 PM
5 points
8 comments1 min readLW link

How I saved 1 hu­man life (in ex­pec­ta­tion) with­out over­think­ing it

Christopher KingDec 22, 2024, 8:53 PM
19 points
0 comments4 min readLW link

Christo­pher King’s Shortform

Christopher KingDec 18, 2024, 9:02 PM
5 points
1 commentLW link

LDT (and ev­ery­thing else) can be irrational

Christopher KingNov 6, 2024, 4:05 AM
10 points
13 comments2 min readLW link

Acausal Now: We could to­tally acausally bar­gain with aliens at our cur­rent tech level if desired

Christopher KingAug 9, 2023, 12:50 AM
1 point
5 comments4 min readLW link

Ne­cro­mancy’s un­in­tended con­se­quences.

Christopher KingAug 9, 2023, 12:08 AM
−6 points
2 comments2 min readLW link

How do low level hy­pothe­ses con­strain high level ones? The mys­tery of the dis­ap­pear­ing di­a­mond.

Christopher KingJul 11, 2023, 7:27 PM
17 points
11 comments2 min readLW link

Challenge pro­posal: small­est pos­si­ble self-hard­en­ing back­door for RLHF

Christopher KingJun 29, 2023, 4:56 PM
7 points
0 comments2 min readLW link

An­throp­i­cally Blind: the an­thropic shadow is re­flec­tively inconsistent

Christopher KingJun 29, 2023, 2:36 AM
43 points
40 comments10 min readLW link

Solomonoff in­duc­tion still works if the uni­verse is un­com­putable, and its use­ful­ness doesn’t re­quire know­ing Oc­cam’s razor

Christopher KingJun 18, 2023, 1:52 AM
38 points
28 comments4 min readLW link

De­mys­tify­ing Born’s rule

Christopher KingJun 14, 2023, 3:16 AM
5 points
26 comments3 min readLW link

Cur­rent AI harms are also sci-fi

Christopher KingJun 8, 2023, 5:49 PM
26 points
3 comments1 min readLW link

In­fer­ence from a Math­e­mat­i­cal De­scrip­tion of an Ex­ist­ing Align­ment Re­search: a pro­posal for an outer al­ign­ment re­search program

Christopher KingJun 2, 2023, 9:54 PM
7 points
4 comments16 min readLW link

The un­spo­ken but ridicu­lous as­sump­tion of AI doom: the hid­den doom assumption

Christopher KingJun 1, 2023, 5:01 PM
−9 points
1 comment3 min readLW link

[Question] What pro­jects and efforts are there to pro­mote AI safety re­search?

Christopher KingMay 24, 2023, 12:33 AM
4 points
0 comments1 min readLW link

See­ing Ghosts by GPT-4

Christopher KingMay 20, 2023, 12:11 AM
−13 points
0 comments1 min readLW link

We are mis­al­igned: the sad­den­ing idea that most of hu­man­ity doesn’t in­trin­si­cally care about x-risk, even on a per­sonal level

Christopher KingMay 19, 2023, 4:12 PM
3 points
5 comments2 min readLW link

Pro­posal: we should start refer­ring to the risk from un­al­igned AI as a type of *ac­ci­dent risk*

Christopher KingMay 16, 2023, 3:18 PM
22 points
6 comments2 min readLW link