Tun­ing your Cog­ni­tive Strategies

Raemon27 Apr 2023 20:32 UTC
155 points
57 comments9 min readLW link
(bewelltuned.com)

The LW cross­roads of purpose

Caerulea-Lawrence27 Apr 2023 19:53 UTC
11 points
2 comments2 min readLW link

Me­tac­u­lus Event: Fore­cast Fri­day, April 28th at 12pm ET — Speed Fore­cast­ing Ses­sion!

ChristianWilliams27 Apr 2023 19:50 UTC
0 points
0 comments1 min readLW link

In­frafunc­tions Proofs

Diffractor27 Apr 2023 19:25 UTC
12 points
1 comment10 min readLW link

In­frafunc­tions and Ro­bust Optimization

Diffractor27 Apr 2023 19:25 UTC
61 points
11 comments15 min readLW link

What are the limits of su­per­in­tel­li­gence?

rainy27 Apr 2023 18:29 UTC
4 points
3 comments5 min readLW link

A Pro­posal for AI Align­ment: Us­ing Directly Op­pos­ing Models

Arne B27 Apr 2023 18:05 UTC
0 points
5 comments3 min readLW link

My views on “doom”

paulfchristiano27 Apr 2023 17:50 UTC
248 points
37 comments2 min readLW link1 review
(ai-alignment.com)

[un­ti­tled post]

NeuralSystem_e5e127 Apr 2023 17:37 UTC
3 points
0 comments1 min readLW link

An In­ter­na­tional Man­hat­tan Pro­ject for Ar­tifi­cial Intelligence

Glenn Clayton27 Apr 2023 17:34 UTC
−11 points
2 comments5 min readLW link

Quote quiz: “drift­ing into de­pen­dence”

jasoncrawford27 Apr 2023 15:13 UTC
7 points
6 comments1 min readLW link
(rootsofprogress.org)

Se­cond-Level Em­piri­cism: Refram­ing the Two-Child Puzzle

Richard Henage27 Apr 2023 15:04 UTC
16 points
5 comments3 min readLW link

In­ter­view with Paul Chris­ti­ano: How We Prevent the AI’s from Killing us

Dalmert27 Apr 2023 14:39 UTC
12 points
0 comments1 min readLW link
(www.youtube.com)

AI #9: The Merge and the Million Tokens

Zvi27 Apr 2023 14:20 UTC
36 points
8 comments53 min readLW link
(thezvi.wordpress.com)

AI doom from an LLM-plateau-ist perspective

Steven Byrnes27 Apr 2023 13:58 UTC
157 points
24 comments6 min readLW link

Ro­mance, mi­s­un­der­stand­ing, so­cial stances, and the hu­man LLM

Kaj_Sotala27 Apr 2023 12:59 UTC
69 points
32 comments16 min readLW link

“A Note on the Com­pat­i­bil­ity of Differ­ent Ro­bust Pro­gram Equil­ibria of the Pri­soner’s Dilemma”

the gears to ascension27 Apr 2023 7:34 UTC
18 points
5 comments1 min readLW link
(arxiv.org)

AI chat­bots don’t know why they did it

skybrian27 Apr 2023 6:57 UTC
18 points
11 comments2 min readLW link
(skybrian.substack.com)

The Great Ide­olog­i­cal Con­flict: In­tu­ition­ists vs. Establishmentarians

Thoth Hermes27 Apr 2023 1:49 UTC
3 points
0 comments11 min readLW link
(thothhermes.substack.com)

Au­tomat­ing the Breath Pulse

jefftk27 Apr 2023 0:10 UTC
11 points
0 comments1 min readLW link
(www.jefftk.com)

Free­dom Is All We Need

Leo Glisic27 Apr 2023 0:09 UTC
−1 points
8 comments10 min readLW link

Con­tra Yud­kowsky on Doom from Foom #2

jacob_cannell27 Apr 2023 0:07 UTC
93 points
76 comments6 min readLW link

A very non-tech­ni­cal ex­pla­na­tion of the ba­sics of in­fra-Bayesianism

David Matolcsi26 Apr 2023 22:57 UTC
62 points
9 comments9 min readLW link

LM Si­tu­a­tional Aware­ness, Eval­u­a­tion Pro­posal: Vio­lat­ing Imitation

Jacob Pfau26 Apr 2023 22:53 UTC
16 points
2 comments2 min readLW link

Re­cent Database Mi­gra­tion—Re­port Bugs

RobertM26 Apr 2023 22:19 UTC
38 points
2 comments1 min readLW link

In­fra-Bayesi­anism nat­u­rally leads to the mono­ton­ic­ity prin­ci­ple, and I think this is a problem

David Matolcsi26 Apr 2023 21:39 UTC
19 points
6 comments4 min readLW link

Un­der­stand­ing new terms via etymology

corruptedCatapillar26 Apr 2023 20:48 UTC
4 points
1 comment2 min readLW link
(forum.effectivealtruism.org)

Chad Jones pa­per mod­el­ing AI and x-risk vs. growth

jasoncrawford26 Apr 2023 20:07 UTC
39 points
7 comments2 min readLW link
(web.stanford.edu)

I was Wrong, Si­mu­la­tor The­ory is Real

Robert_AIZI26 Apr 2023 17:45 UTC
75 points
7 comments3 min readLW link
(aizi.substack.com)

$250 prize for check­ing Jake Can­nell’s Brain Efficiency

Alexander Gietelink Oldenziel26 Apr 2023 16:21 UTC
123 points
170 comments2 min readLW link

My ver­sion of Si­mu­lacra Levels

Daniel Kokotajlo26 Apr 2023 15:50 UTC
42 points
15 comments3 min readLW link

[Question] Is the fact that we don’t ob­serve any ob­vi­ous glitch ev­i­dence that we’re not in a simu­la­tion?

Jim Buhler26 Apr 2023 14:57 UTC
8 points
16 comments1 min readLW link

Tran­script and Brief Re­sponse to Twit­ter Con­ver­sa­tion be­tween Yann LeCunn and Eliezer Yudkowsky

Zvi26 Apr 2023 13:10 UTC
190 points
51 comments10 min readLW link
(thezvi.wordpress.com)

What comes af­ter?

rogersbacon26 Apr 2023 12:44 UTC
3 points
0 comments2 min readLW link
(www.secretorum.life)

Ac­ci­den­tal Terraforming

Sable26 Apr 2023 6:49 UTC
9 points
16 comments5 min readLW link
(affablyevil.substack.com)

Philos­o­phy by Paul Gra­ham Link

EniScien26 Apr 2023 5:36 UTC
21 points
4 comments1 min readLW link

Box­ing at the gym

yakimoff26 Apr 2023 5:10 UTC
1 point
0 comments1 min readLW link

Si­be­lius + drinks

yakimoff26 Apr 2023 5:08 UTC
1 point
0 comments1 min readLW link

A sim­ple pre­sen­ta­tion of AI risk arguments

Seth Herd26 Apr 2023 2:19 UTC
16 points
0 comments2 min readLW link

Archety­pal Trans­fer Learn­ing: a Pro­posed Align­ment Solu­tion that solves the In­ner & Outer Align­ment Prob­lem while adding Cor­rigible Traits to GPT-2-medium

MiguelDev26 Apr 2023 1:37 UTC
14 points
5 comments10 min readLW link

[Question] How Many Bits Of Op­ti­miza­tion Can One Bit Of Ob­ser­va­tion Un­lock?

johnswentworth26 Apr 2023 0:26 UTC
62 points
32 comments3 min readLW link

Believe in Your­self and don’t stop Improving

Johannes C. Mayer25 Apr 2023 22:34 UTC
0 points
0 comments1 min readLW link

Should LW have an offi­cial list of norms?

Ruby25 Apr 2023 21:20 UTC
58 points
31 comments5 min readLW link

Im­ple­ment­ing a Trans­former from scratch in PyTorch—a write-up on my experience

Mislav Jurić25 Apr 2023 20:51 UTC
20 points
0 comments10 min readLW link

Ex­plor­ing the Lot­tery Ticket Hypothesis

Rauno Arike25 Apr 2023 20:06 UTC
54 points
3 comments11 min readLW link

Ge­netic Se­quenc­ing of Wastew­a­ter: Prevalence to Rel­a­tive Abundance

jefftk25 Apr 2023 19:30 UTC
17 points
2 comments2 min readLW link
(www.jefftk.com)

[Feed­back please] New User’s Guide to LessWrong

Ruby25 Apr 2023 18:54 UTC
38 points
18 comments6 min readLW link

Refram­ing the bur­den of proof: Com­pa­nies should prove that mod­els are safe (rather than ex­pect­ing au­di­tors to prove that mod­els are dan­ger­ous)

Akash25 Apr 2023 18:49 UTC
27 points
11 comments3 min readLW link
(childrenoficarus.substack.com)

LLMs for on­line dis­cus­sion moderation

Dave Lindbergh25 Apr 2023 16:53 UTC
12 points
3 comments3 min readLW link

AI Safety Newslet­ter #3: AI policy pro­pos­als and a new challenger approaches

ozhang25 Apr 2023 16:15 UTC
33 points
0 comments1 min readLW link