Test Cases for Im­pact Reg­u­lari­sa­tion Methods

DanielFilan6 Feb 2019 21:50 UTC
72 points
5 comments13 min readLW link
(danielfilan.com)

A ten­ta­tive solu­tion to a cer­tain mytholog­i­cal beast of a problem

Edward Knox6 Feb 2019 20:42 UTC
−11 points
9 comments1 min readLW link

AI Align­ment is Alchemy.

Jeevan6 Feb 2019 20:32 UTC
−9 points
20 comments1 min readLW link

My use of the phrase “Su­per-Hu­man Feed­back”

David Scott Krueger (formerly: capybaralet)6 Feb 2019 19:11 UTC
13 points
0 comments1 min readLW link

Thoughts on Ben Garfinkel’s “How sure are we about this AI stuff?”

David Scott Krueger (formerly: capybaralet)6 Feb 2019 19:09 UTC
25 points
17 comments1 min readLW link

Show LW: (video) how to re­mem­ber ev­ery­thing you learn

ArthurLidia6 Feb 2019 19:02 UTC
3 points
0 comments1 min readLW link

Does the EA com­mu­nity do “ba­sic sci­ence” grants? How do I get one?

Jameson Quinn6 Feb 2019 18:10 UTC
7 points
6 comments1 min readLW link

Is the World Get­ting Bet­ter? A brief sum­mary of re­cent debate

ErickBall6 Feb 2019 17:38 UTC
35 points
8 comments2 min readLW link
(capx.co)

Se­cu­rity amplification

paulfchristiano6 Feb 2019 17:28 UTC
21 points
2 comments13 min readLW link

Align­ment Newslet­ter #44

Rohin Shah6 Feb 2019 8:30 UTC
18 points
0 comments9 min readLW link
(mailchi.mp)

South Bay Meetup March 2nd

David Friedman6 Feb 2019 6:48 UTC
1 point
0 comments1 min readLW link

[Question] If Ra­tion­al­ity can be likened to a ‘Mar­tial Art’, what would be the Forms?

Bae's Theorem6 Feb 2019 5:48 UTC
21 points
10 comments1 min readLW link

Com­plex­ity Penalties in Statis­ti­cal Learning

michael_h6 Feb 2019 4:13 UTC
31 points
3 comments6 min readLW link

Au­to­mated Nomic Game 2

jefftk5 Feb 2019 22:11 UTC
19 points
2 comments2 min readLW link

Should we bait crim­i­nals us­ing clones ?

Aël Chappuit5 Feb 2019 21:13 UTC
−23 points
3 comments1 min readLW link

De­scribing things: par­si­mony, fruit­ful­ness, and adapt­abil­ity

Mary Chernyshenko5 Feb 2019 20:59 UTC
1 point
0 comments1 min readLW link

Philos­o­phy as low-en­ergy approximation

Charlie Steiner5 Feb 2019 19:34 UTC
40 points
20 comments3 min readLW link

When to use quantilization

RyanCarey5 Feb 2019 17:17 UTC
65 points
5 comments4 min readLW link

(notes on) Policy Desider­ata for Su­per­in­tel­li­gent AI: A Vec­tor Field Approach

Ben Pace4 Feb 2019 22:08 UTC
43 points
5 comments7 min readLW link

SSC Paris Meetup, 09/​02/​18

fbreton4 Feb 2019 19:54 UTC
1 point
0 comments1 min readLW link

Jan­uary 2019 gw­ern.net newsletter

gwern4 Feb 2019 15:53 UTC
15 points
0 comments1 min readLW link
(www.gwern.net)

My athe­ism story

Pausecafe4 Feb 2019 14:33 UTC
26 points
3 comments7 min readLW link

(Why) Does the Basilisk Ar­gu­ment fail?

Lookingforyourlogic3 Feb 2019 23:50 UTC
0 points
11 comments2 min readLW link

Con­struct­ing Goodhart

johnswentworth3 Feb 2019 21:59 UTC
29 points
10 comments3 min readLW link

Con­clu­sion to the se­quence on value learning

Rohin Shah3 Feb 2019 21:05 UTC
51 points
20 comments5 min readLW link

AI Safety Pr­ereq­ui­sites Course: Re­vamp and New Lessons

philip_b3 Feb 2019 21:04 UTC
24 points
5 comments1 min readLW link

[Question] What are some of bizarre the­o­ries based on an­thropic rea­son­ing?

Dr. Jamchie3 Feb 2019 18:48 UTC
21 points
13 comments1 min readLW link

Ra­tion­al­ity: What’s the point?

Hazard3 Feb 2019 16:34 UTC
12 points
11 comments1 min readLW link

Quan­tify­ing Hu­man Suffer­ing and “Every­day Suffer­ing”

willfranks3 Feb 2019 13:07 UTC
7 points
3 comments1 min readLW link

[Question] How to stay con­cen­trated for a long pe­riod of time?

infinickel3 Feb 2019 5:24 UTC
6 points
15 comments1 min readLW link

How to no­tice be­ing mind-hacked

Shmi2 Feb 2019 23:13 UTC
18 points
22 comments2 min readLW link

De­pres­sion philosophizing

aaq2 Feb 2019 22:54 UTC
6 points
2 comments1 min readLW link

LessWrong DC: Metameetup

rusalkii2 Feb 2019 18:50 UTC
1 point
0 comments1 min readLW link

SSC At­lanta Meetup

Steve French2 Feb 2019 3:11 UTC
2 points
0 comments1 min readLW link

[Question] How does Gra­di­ent Des­cent In­ter­act with Good­hart?

Scott Garrabrant2 Feb 2019 0:14 UTC
68 points
19 comments4 min readLW link

Philadelphia SSC Meetup

Majuscule1 Feb 2019 23:51 UTC
1 point
0 comments1 min readLW link

STRUCTURE: Real­ity and ra­tio­nal best practice

Hazard1 Feb 2019 23:51 UTC
5 points
2 comments1 min readLW link

An At­tempt To Ex­plain No-Self In Sim­ple Terms

Justin Vriend1 Feb 2019 23:50 UTC
1 point
0 comments3 min readLW link

STRUCTURE: How the So­cial Affects your rationality

Hazard1 Feb 2019 23:35 UTC
0 points
0 comments1 min readLW link

STRUCTURE: A Crash Course in Your Brain

Hazard1 Feb 2019 23:17 UTC
6 points
4 comments1 min readLW link

Fe­bru­ary Nashville SSC Meetup

Dude McDude1 Feb 2019 22:36 UTC
1 point
0 comments1 min readLW link

[Question] What kind of in­for­ma­tion would serve as the best ev­i­dence for re­solv­ing the de­bate of whether a cen­trist or leftist Demo­cratic nom­i­nee is like­lier to take the White House in 2020?

Evan_Gaensbauer1 Feb 2019 18:40 UTC
10 points
10 comments3 min readLW link

Ur­gent & im­por­tant: How (not) to do your to-do list

bfinn1 Feb 2019 17:44 UTC
51 points
20 comments13 min readLW link

Who wants to be a Million­aire?

Bucky1 Feb 2019 14:02 UTC
29 points
1 comment11 min readLW link

What is Wrong?

Inyuki1 Feb 2019 12:02 UTC
1 point
2 comments2 min readLW link

Drexler on AI Risk

PeterMcCluskey1 Feb 2019 5:11 UTC
35 points
10 comments9 min readLW link
(www.bayesianinvestor.com)

Boundaries—A map and ter­ri­tory ex­per­i­ment. [post-ra­tio­nal­ity]

Elo1 Feb 2019 2:08 UTC
−18 points
14 comments2 min readLW link

[Question] Why is this util­i­tar­ian calcu­lus wrong? Or is it?

EconomicModel31 Jan 2019 23:57 UTC
15 points
21 comments1 min readLW link

Small hope for less bias and more practability

ArthurLidia31 Jan 2019 22:09 UTC
0 points
0 comments1 min readLW link

Reli­a­bil­ity am­plifi­ca­tion

paulfchristiano31 Jan 2019 21:12 UTC
24 points
3 comments7 min readLW link