RSS

LessWrong’s UX may not be liv­ing up to its ideas

neo18 Mar 2026 20:16 UTC
5 points
0 comments2 min readLW link

Two Skil­lsets You Need to Launch an Im­pact­ful AI Safety Project

18 Mar 2026 20:09 UTC
17 points
0 comments6 min readLW link

“Act-based ap­proval-di­rected agents”, for IDA skeptics

Steven Byrnes18 Mar 2026 18:47 UTC
39 points
2 comments5 min readLW link

“Lost in the Mid­dle” Repli­cates

Stephen Martin18 Mar 2026 16:09 UTC
18 points
0 comments1 min readLW link

Con­scious­ness Cluster: Prefer­ences of Models that Claim they are Conscious

18 Mar 2026 16:06 UTC
55 points
2 comments5 min readLW link

Solv­ing Slop

Noam Makavy18 Mar 2026 5:18 UTC
−1 points
0 comments1 min readLW link

Sy­co­phancy Towards Re­searchers Drives Perfor­ma­tive Misalignment

18 Mar 2026 4:59 UTC
70 points
0 comments21 min readLW link

New­comb’s Para­dox Simulation

shakelush18 Mar 2026 4:30 UTC
6 points
2 comments1 min readLW link
(www.onlineide.pro)

The Psy­chopa­thy Spectrum

Dawn Drescher17 Mar 2026 21:36 UTC
36 points
0 comments1 min readLW link
(impartial-priorities.org)

LLMs as Gi­ant Lookup-Tables of Shal­low Circuits

17 Mar 2026 21:35 UTC
85 points
25 comments7 min readLW link

Re­quiem for a Tran­shu­man Timeline

Ihor Kendiukhov17 Mar 2026 21:27 UTC
145 points
11 comments5 min readLW link

There is No One There: A sim­ple ex­per­i­ment to con­vince your­self that LLMs prob­a­bly are not conscious

Peter Kuhn17 Mar 2026 17:26 UTC
12 points
17 comments5 min readLW link

Re­search note on win­dow shift­ing training

17 Mar 2026 15:58 UTC
26 points
1 comment15 min readLW link

Ma­halo, Kailua-Kona

jenn17 Mar 2026 15:54 UTC
28 points
0 comments9 min readLW link

[Linkpost] “Ex­ploit­ing” the In-Kind Dona­tion Bias

jknowak17 Mar 2026 8:27 UTC
1 point
0 comments1 min readLW link
(epsilonaligned.substack.com)

[Linkpost] Let’s Start Eventmaxxing

jknowak17 Mar 2026 8:22 UTC
0 points
0 comments1 min readLW link
(epsilonaligned.substack.com)

[Linkpost] Many Faces Of Prob­lem Decomposition

jknowak17 Mar 2026 8:17 UTC
1 point
0 comments1 min readLW link
(epsilonaligned.substack.com)

How to not do de­ci­sion the­ory backwards

Anthony DiGiovanni17 Mar 2026 7:22 UTC
14 points
0 comments16 min readLW link

I made a job-level AI ca­pa­bil­ity es­ti­ma­tor by ask­ing “Where is AI do­ing similar work to­day?”

Alex A17 Mar 2026 4:45 UTC
6 points
1 comment1 min readLW link

The tech bro ethos, and the end of my job

PatrickDFarley17 Mar 2026 4:13 UTC
11 points
5 comments3 min readLW link