Align­ment Newslet­ter #34

Rohin ShahNov 26, 2018, 11:10 PM
24 points
0 comments10 min readLW link
(mailchi.mp)

Boltz­mann Brains, Si­mu­la­tions and self re­fut­ing hy­poth­e­sis

Donald HobsonNov 26, 2018, 7:09 PM
1 point
9 comments1 min readLW link

Quan­tum Me­chan­ics, Noth­ing to do with Consciousness

Donald HobsonNov 26, 2018, 6:59 PM
5 points
27 comments3 min readLW link

Sta­tus model

BuckyNov 26, 2018, 3:05 PM
26 points
7 comments3 min readLW link

Hu­mans Con­sult­ing HCH

paulfchristianoNov 25, 2018, 11:18 PM
39 points
9 comments1 min readLW link

Ap­proval-di­rected bootstrapping

paulfchristianoNov 25, 2018, 11:18 PM
24 points
0 comments1 min readLW link

How rapidly are GPUs im­prov­ing in price perfor­mance?

gallabytesNov 25, 2018, 7:54 PM
31 points
9 commentsLW link
(mediangroup.org)

Values Weren’t Com­plex, Once.

DavidmanheimNov 25, 2018, 9:17 AM
36 points
13 comments2 min readLW link

A cul­ture of ex­ploita­tion?

Bae's TheoremNov 24, 2018, 10:00 PM
1 point
3 comments1 min readLW link

Fixed Point Discussion

Scott GarrabrantNov 24, 2018, 8:53 PM
45 points
2 comments4 min readLW link

Four fac­tors that mod­er­ate the in­ten­sity of emotions

RubyNov 24, 2018, 8:40 PM
63 points
11 comments8 min readLW link

deluks917 on On­line Weirdos

Jacob FalkovichNov 24, 2018, 5:03 PM
24 points
3 comments10 min readLW link

[Mon­treal] Towards High-As­surance Ad­vanced AI Sys­tems by Richard Mallah

Mati_RoyNov 24, 2018, 6:24 AM
3 points
0 comments1 min readLW link

Up­com­ing: Open Questions

RaemonNov 24, 2018, 1:39 AM
41 points
7 comments2 min readLW link

A Dragon Con­fronts the Terasem Movement

AlephywrNov 24, 2018, 1:31 AM
−4 points
10 comments25 min readLW link
(dancefighterredux.wordpress.com)

What if peo­ple sim­ply fore­casted your fu­ture choices?

ozziegooenNov 23, 2018, 10:52 AM
16 points
6 comments6 min readLW link

Over­sight of Un­safe Sys­tems via Dy­namic Safety En­velopes

DavidmanheimNov 23, 2018, 8:37 AM
10 points
2 comments2 min readLW link

On MIRI’s new re­search directions

Rob BensingerNov 22, 2018, 11:42 PM
53 points
12 comments1 min readLW link
(intelligence.org)

LW Up­date 2018-11-22 – Abridged Comments

RaemonNov 22, 2018, 10:11 PM
11 points
16 comments1 min readLW link

Ap­proval-di­rected agents

paulfchristianoNov 22, 2018, 9:15 PM
31 points
10 comments15 min readLW link

Believ­ing oth­ers’ priors

rkNov 22, 2018, 8:44 PM
8 points
19 comments7 min readLW link

Spec­u­la­tive Evopsych, Ep. 1

Optimization ProcessNov 22, 2018, 7:00 PM
41 points
9 comments1 min readLW link

If You Want to Win, Stop Conceding

Davis_KingsleyNov 22, 2018, 6:10 PM
47 points
15 comments3 min readLW link

Re­view: Artifact

ZviNov 22, 2018, 3:00 PM
21 points
3 comments13 min readLW link
(thezvi.wordpress.com)

Per­spec­tive Rea­son­ing and the Sleep­ing Beauty Problem

dadadarrenNov 22, 2018, 11:55 AM
6 points
10 comments2 min readLW link

The Se­man­tic Man

namespaceNov 22, 2018, 8:38 AM
19 points
4 comments1 min readLW link
(www.generalsemantics.org)

Je­sus Made Me Ra­tional (An In­tro­duc­tion)

MotasaurusNov 22, 2018, 5:09 AM
−14 points
56 comments3 min readLW link

Iter­a­tion Fixed Point Exercises

Nov 22, 2018, 12:35 AM
33 points
12 comments3 min readLW link

Sugges­tion: New ma­te­rial shouldn’t be re­leased too fast

Chris_LeongNov 21, 2018, 4:39 PM
23 points
7 comments1 min readLW link

EA Bris­tol Strat­egy Meet­ing

thegreatnickNov 21, 2018, 10:57 AM
1 point
0 comments1 min readLW link

Ra­tion­al­ity Café No. 6 - The Se­quences, Part 1; Sec­tion B Repeat

thegreatnickNov 21, 2018, 10:54 AM
8 points
2 comments1 min readLW link

EA Funds: Long-Term Fu­ture fund is open to ap­pli­ca­tions un­til Novem­ber 24th (this Satur­day)

habrykaNov 21, 2018, 3:39 AM
37 points
0 comments1 min readLW link

In­cor­rect hy­pothe­ses point to cor­rect observations

Kaj_SotalaNov 20, 2018, 9:10 PM
169 points
40 comments4 min readLW link
(kajsotala.fi)

Preschool: Much Less Than You Wanted To Know

ZviNov 20, 2018, 7:30 PM
65 points
15 comments2 min readLW link
(thezvi.wordpress.com)

New safety re­search agenda: scal­able agent al­ign­ment via re­ward modeling

VikaNov 20, 2018, 5:29 PM
34 points
12 comments1 min readLW link
(medium.com)

Pro­saic AI alignment

paulfchristianoNov 20, 2018, 1:56 PM
48 points
10 comments8 min readLW link

Moscow LW meetup in “Nauchka” library

Alexander230Nov 20, 2018, 12:19 PM
2 points
0 comments1 min readLW link

[Insert clever in­tro here]

Bae's TheoremNov 20, 2018, 3:26 AM
18 points
13 comments1 min readLW link

Align­ment Newslet­ter #33

Rohin ShahNov 19, 2018, 5:20 PM
23 points
0 comments9 min readLW link
(mailchi.mp)

Games in Kocherga club: Fal­la­cy­ma­nia, Tower of Chaos, Scien­tific Discovery

Alexander230Nov 19, 2018, 2:23 PM
2 points
0 comments1 min readLW link

Let­ting Others Be Vulnerable

lifelonglearnerNov 19, 2018, 2:59 AM
34 points
6 comments7 min readLW link

Click­bait might not be de­stroy­ing our gen­eral Intelligence

Donald HobsonNov 19, 2018, 12:13 AM
25 points
13 comments2 min readLW link

South Bay Meetup 12/​8

DavidFriedmanNov 19, 2018, 12:04 AM
3 points
0 comments1 min readLW link

[Link] “They go to­gether: Free­dom, Pros­per­ity, and Big Govern­ment”

CronoDASNov 18, 2018, 4:51 PM
11 points
3 comments1 min readLW link

Col­lab­o­ra­tion-by-De­sign ver­sus Emer­gent Collaboration

DavidmanheimNov 18, 2018, 7:22 AM
11 points
2 comments2 min readLW link

Di­ag­o­nal­iza­tion Fixed Point Exercises

Nov 18, 2018, 12:31 AM
40 points
25 comments3 min readLW link

Ia! Ia! Ex­tradi­men­sional Cephalo­pod Nafl’fh­tagn!

ExCephNov 17, 2018, 11:00 PM
14 points
5 comments1 min readLW link

Effec­tive Altru­ism, YouTube, and AI (talk by Lê Nguyên Hoang)

Paperclip MinimizerNov 17, 2018, 7:21 PM
3 points
0 commentsLW link
(www.youtube.com)

An un­al­igned benchmark

paulfchristianoNov 17, 2018, 3:51 PM
31 points
0 comments9 min readLW link

On Ri­gor­ous Er­ror Handling

Martin SustrikNov 17, 2018, 9:20 AM
13 points
4 comments6 min readLW link
(250bpm.com)