Piling bounded arguments

momom219 Sep 2024 22:27 UTC
7 points
0 comments4 min readLW link

We Don’t Know Our Own Values, but Re­ward Bridges The Is-Ought Gap

19 Sep 2024 22:22 UTC
45 points
47 comments5 min readLW link

In­ter­ested in Cog­ni­tive Boot­camp?

Raemon19 Sep 2024 22:12 UTC
48 points
0 comments2 min readLW link

Just How Good Are Modern Chess Com­put­ers?

nem19 Sep 2024 18:57 UTC
10 points
1 comment6 min readLW link

RLHF is the worst pos­si­ble thing done when fac­ing the al­ign­ment problem

tailcalled19 Sep 2024 18:56 UTC
32 points
10 comments6 min readLW link

AISafety.info: What are In­duc­tive Bi­ases?

Algon19 Sep 2024 17:26 UTC
11 points
4 comments2 min readLW link
(aisafety.info)

Physics of Lan­guage mod­els (part 2.1)

Nathan Helm-Burger19 Sep 2024 16:48 UTC
9 points
2 comments1 min readLW link
(youtu.be)

Why good things of­ten don’t lead to bet­ter outcomes

DMMF19 Sep 2024 16:37 UTC
16 points
1 comment4 min readLW link
(danfrank.ca)

To CoT or not to CoT? Chain-of-thought helps mainly on math and sym­bolic reasoning

Bogdan Ionut Cirstea19 Sep 2024 16:13 UTC
21 points
1 comment1 min readLW link
(arxiv.org)

Laz­i­ness death spirals

PatrickDFarley19 Sep 2024 15:58 UTC
242 points
34 comments8 min readLW link

[In­tu­itive self-mod­els] 1. Preliminaries

Steven Byrnes19 Sep 2024 13:45 UTC
86 points
20 comments15 min readLW link

AI #82: The Gover­nor Ponders

Zvi19 Sep 2024 13:30 UTC
50 points
8 comments27 min readLW link
(thezvi.wordpress.com)

Slave Mo­ral­ity: A place for ev­ery man and ev­ery man in his place

Martin Sustrik19 Sep 2024 4:20 UTC
16 points
7 comments2 min readLW link
(250bpm.substack.com)

Which LessWrong/​Align­ment top­ics would you like to be tu­tored in? [Poll]

Ruby19 Sep 2024 1:35 UTC
43 points
12 comments1 min readLW link

The Oblique­ness Thesis

jessicata19 Sep 2024 0:26 UTC
75 points
16 comments17 min readLW link

How to choose what to work on

jasoncrawford18 Sep 2024 20:39 UTC
22 points
6 comments4 min readLW link
(blog.rootsofprogress.org)

In­ten­tion-to-Treat (Re: How harm­ful is mu­sic, re­ally?)

kqr18 Sep 2024 18:44 UTC
11 points
0 comments5 min readLW link
(entropicthoughts.com)

The case for a nega­tive al­ign­ment tax

18 Sep 2024 18:33 UTC
79 points
20 comments7 min readLW link

En­doge­nous Growth and Hu­man Intelligence

Nicholas D.18 Sep 2024 14:05 UTC
3 points
0 comments2 min readLW link

In­quisi­tive vs. ad­ver­sar­ial rationality

gb18 Sep 2024 13:50 UTC
6 points
9 comments2 min readLW link

Pro­nouns are Annoying

ymeskhout18 Sep 2024 13:30 UTC
15 points
21 comments4 min readLW link
(www.ymeskhout.com)

Is “su­per­hu­man” AI fore­cast­ing BS? Some ex­per­i­ments on the “539″ bot from the Cen­tre for AI Safety

titotal18 Sep 2024 13:07 UTC
78 points
3 comments1 min readLW link
(open.substack.com)

Knowl­edge’s practicability

Ted Nguyễn18 Sep 2024 2:31 UTC
−5 points
0 comments7 min readLW link
(tednguyen.substack.com)

Skills from a year of Pur­pose­ful Ra­tion­al­ity Practice

Raemon18 Sep 2024 2:05 UTC
178 points
18 comments7 min readLW link

[Question] Where to find re­li­able re­views of AI prod­ucts?

Elizabeth17 Sep 2024 23:48 UTC
29 points
6 comments1 min readLW link

Su­per­po­si­tion through Ac­tive Learn­ing Lens

akankshanc17 Sep 2024 17:32 UTC
1 point
0 comments10 min readLW link

Sur­vey—Psy­cholog­i­cal Im­pact of Long-Term AI Engagement

Manuela García17 Sep 2024 17:31 UTC
2 points
0 comments1 min readLW link

Sur­vey—Psy­cholog­i­cal Im­pact of Long-Term AI Engagement

Manuela García17 Sep 2024 17:31 UTC
1 point
1 comment1 min readLW link

[Question] What does it mean for an event or ob­ser­va­tion to have prob­a­bil­ity 0 or 1 in Bayesian terms?

Noosphere8917 Sep 2024 17:28 UTC
1 point
22 comments1 min readLW link

How harm­ful is mu­sic, re­ally?

dkl917 Sep 2024 14:53 UTC
10 points
6 comments3 min readLW link
(dkl9.net)

Monthly Roundup #22: Septem­ber 2024

Zvi17 Sep 2024 12:20 UTC
35 points
10 comments45 min readLW link
(thezvi.wordpress.com)

I fi­nally got ChatGPT to sound like me

lsusr17 Sep 2024 9:39 UTC
46 points
18 comments6 min readLW link

Food, Pri­son & Ex­otic An­i­mals: Sparse Au­toen­coders De­tect 6.5x Perform­ing Youtube Thumbnails

Louka Ewington-Pitsos17 Sep 2024 3:52 UTC
6 points
2 comments7 min readLW link

Head in the Cloud: Why an Upload of Your Mind is Not You

xhq17 Sep 2024 0:25 UTC
−11 points
3 comments14 min readLW link

[Question] How does some­one prove that their gen­eral in­tel­li­gence is above av­er­age?

M. Y. Zuo16 Sep 2024 21:01 UTC
−3 points
12 comments1 min readLW link

[Question] Does life ac­tu­ally lo­cally *in­crease* en­tropy?

tailcalled16 Sep 2024 20:30 UTC
10 points
27 comments1 min readLW link

Book re­view: Xenosystems

jessicata16 Sep 2024 20:17 UTC
48 points
18 comments37 min readLW link
(unstableontology.com)

MIRI’s Septem­ber 2024 newsletter

Harlan16 Sep 2024 18:15 UTC
46 points
0 comments1 min readLW link
(intelligence.org)

Gen­er­a­tive ML in chem­istry is bot­tle­necked by synthesis

Abhishaike Mahajan16 Sep 2024 16:31 UTC
38 points
2 comments14 min readLW link
(www.owlposting.com)

Se­cret Col­lu­sion: Will We Know When to Un­plug AI?

16 Sep 2024 16:07 UTC
55 points
7 comments31 min readLW link

GPT-o1

Zvi16 Sep 2024 13:40 UTC
86 points
34 comments46 min readLW link
(thezvi.wordpress.com)

[Question] Can sub­junc­tive de­pen­dence emerge from a sim­plic­ity prior?

Daniel C16 Sep 2024 12:39 UTC
6 points
0 comments1 min readLW link

Longevity and the Mind

George3d616 Sep 2024 9:43 UTC
5 points
2 comments10 min readLW link

[Question] What’s the Deal with Log­i­cal Uncer­tainty?

Ape in the coat16 Sep 2024 8:11 UTC
31 points
23 comments2 min readLW link

Re­in­force­ment Learn­ing from In­for­ma­tion Bazaar Feed­back, and other uses of in­for­ma­tion markets

Abhimanyu Pallavi Sudhir16 Sep 2024 1:04 UTC
5 points
1 comment5 min readLW link

Hyperpolation

Gunnar_Zarncke15 Sep 2024 21:37 UTC
22 points
6 comments1 min readLW link
(arxiv.org)

[Question] If I wanted to spend WAY more on AI, what would I spend it on?

Logan Zoellner15 Sep 2024 21:24 UTC
48 points
16 comments1 min readLW link

Su­per­in­tel­li­gence Can’t Solve the Prob­lem of De­cid­ing What You’ll Do

Vladimir_Nesov15 Sep 2024 21:03 UTC
27 points
11 comments1 min readLW link

For Limited Su­per­in­tel­li­gences, Epistemic Ex­clu­sion is Harder than Ro­bust­ness to Log­i­cal Exploitation

Lorec15 Sep 2024 20:49 UTC
3 points
9 comments3 min readLW link

Why I funded PIBBSS

Ryan Kidd15 Sep 2024 19:56 UTC
115 points
21 comments3 min readLW link