RSS

Max H

Karma: 2,422

Most of my posts and comments are about AI and alignment. Posts I’m most proud of, which also provide a good introduction to my worldview:

I also created Forum Karma, and wrote a longer self-introduction here.

PMs and private feedback are always welcome.

NOTE: I am not Max Harms, author of Crystal Society. I’d prefer for now that my LW postings not be attached to my full name when people Google me for other reasons, but you can PM me here or on Discord (m4xed) if you want to know who I am.

Bayesian up­dat­ing in real life is mostly about un­der­stand­ing your hypotheses

Max HJan 1, 2024, 12:10 AM
63 points
4 comments11 min readLW link

Em­mett Shear to be in­terim CEO of OpenAI

Max HNov 20, 2023, 5:40 AM
21 points
5 comments1 min readLW link
(www.theverge.com)

Con­crete pos­i­tive vi­sions for a fu­ture with­out AGI

Max HNov 8, 2023, 3:12 AM
41 points
28 comments8 min readLW link

Try­ing to de­con­fuse some core AI x-risk problems

Oct 17, 2023, 6:36 PM
34 points
13 comments27 min readLW link

An ex­pla­na­tion for ev­ery to­ken: us­ing an LLM to sam­ple an­other LLM

Max HOct 11, 2023, 12:53 AM
35 points
5 comments11 min readLW link

Ac­tu­ally, “per­sonal at­tacks af­ter ob­ject-level ar­gu­ments” is a pretty good rule of epistemic conduct

Max HSep 17, 2023, 8:25 PM
37 points
15 comments7 min readLW link

Fo­rum Karma: view stats and find highly-rated com­ments for any LW user

Max HJul 1, 2023, 3:36 PM
60 points
16 comments2 min readLW link
(forumkarma.com)

10 quick takes about AGI

Max HJun 20, 2023, 2:22 AM
35 points
17 comments7 min readLW link

Four lev­els of un­der­stand­ing de­ci­sion theory

Max HJun 1, 2023, 8:55 PM
12 points
11 comments4 min readLW link

Without a tra­jec­tory change, the de­vel­op­ment of AGI is likely to go badly

Max HMay 29, 2023, 11:42 PM
16 points
2 comments13 min readLW link

Where do you lie on two axes of world ma­nipu­la­bil­ity?

Max HMay 26, 2023, 3:04 AM
30 points
15 comments3 min readLW link

Re­ward is the op­ti­miza­tion tar­get (of ca­pa­bil­ities re­searchers)

Max HMay 15, 2023, 3:22 AM
32 points
4 comments5 min readLW link

Max H’s Shortform

Max HMay 13, 2023, 12:17 AM
5 points
7 comments1 min readLW link

Gra­di­ent hack­ing via ac­tual hacking

Max HMay 10, 2023, 1:57 AM
12 points
7 comments3 min readLW link

LLM cog­ni­tion is prob­a­bly not hu­man-like

Max HMay 8, 2023, 1:22 AM
26 points
15 comments7 min readLW link

A test of your ra­tio­nal­ity skills

Max HApr 20, 2023, 1:19 AM
11 points
11 comments4 min readLW link

Pay­ing the cor­rigi­bil­ity tax

Max HApr 19, 2023, 1:57 AM
14 points
1 comment13 min readLW link

“Aligned” foun­da­tion mod­els don’t im­ply al­igned systems

Max HApr 13, 2023, 4:13 AM
39 points
11 comments5 min readLW link

A decade of lurk­ing, a month of posting

Max HApr 9, 2023, 12:21 AM
70 points
4 comments5 min readLW link

Eliezer on The Lu­nar So­ciety podcast

Max HApr 6, 2023, 4:18 PM
40 points
5 comments1 min readLW link
(www.dwarkeshpatel.com)