[Question] What are MIRI’s big achieve­ments in AI al­ign­ment?

tailcalled7 Mar 2023 21:30 UTC
29 points
7 comments1 min readLW link

A Brief Defense of Ath­let­i­cism

Wofsen7 Mar 2023 20:48 UTC
46 points
5 comments1 min readLW link

[Question] How “grifty” is the Fore­sight In­sti­tute? Are they mak­ing but­ton soup?

Cedar7 Mar 2023 19:43 UTC
7 points
3 comments1 min readLW link

[Question] What‘s in your list of un­solved prob­lems in AI al­ign­ment?

jacquesthibs7 Mar 2023 18:58 UTC
60 points
9 comments1 min readLW link

In­tro­duc­ing AI Align­ment Inc., a Cal­ifor­nia pub­lic benefit cor­po­ra­tion...

TherapistAI7 Mar 2023 18:47 UTC
1 point
4 comments1 min readLW link

Abuse in LessWrong and ra­tio­nal­ist com­mu­ni­ties in Bloomberg News

whistleblower677 Mar 2023 18:45 UTC
1 point
72 comments7 min readLW link
(www.bloomberg.com)

Test post for formatting

Solenoid_Entity7 Mar 2023 17:48 UTC
0 points
2 comments1 min readLW link

The Pinnacle

nem7 Mar 2023 17:07 UTC
11 points
0 comments8 min readLW link

Pod­cast Tran­script: Daniela and Dario Amodei on Anthropic

remember7 Mar 2023 16:47 UTC
46 points
2 comments79 min readLW link
(futureoflife.org)

The View from 30,000 Feet: Pre­face to the Se­cond EleutherAI Retrospective

7 Mar 2023 16:22 UTC
14 points
0 comments4 min readLW link
(blog.eleuther.ai)

Break­ing Rank (Cal­ibra­tion Game)

jenn7 Mar 2023 15:40 UTC
11 points
0 comments2 min readLW link

Ou­trangeous (Cal­ibra­tion Game)

jenn7 Mar 2023 15:29 UTC
36 points
3 comments9 min readLW link

[Linkpost] Some high-level thoughts on the Deep­Mind al­ign­ment team’s strategy

7 Mar 2023 11:55 UTC
128 points
13 comments5 min readLW link
(drive.google.com)

Align­ment works both ways

Karl von Wendt7 Mar 2023 10:41 UTC
23 points
21 comments2 min readLW link

Google’s PaLM-E: An Em­bod­ied Mul­ti­modal Lan­guage Model

SandXbox7 Mar 2023 4:11 UTC
87 points
7 comments1 min readLW link
(palm-e.github.io)

GÖDEL GOING DOWN

Jimdrix_Hendri6 Mar 2023 23:06 UTC
−9 points
3 comments1 min readLW link

Against ubiquitous al­ign­ment taxes

beren6 Mar 2023 19:50 UTC
56 points
10 comments2 min readLW link

Ad­den­dum: ba­sic facts about lan­guage mod­els dur­ing training

beren6 Mar 2023 19:24 UTC
22 points
2 comments5 min readLW link

Un­der­stand­ing The Roots Of Math­e­mat­ics Be­fore Find­ing The Roots Of A Func­tion.

LiesLaris6 Mar 2023 18:47 UTC
2 points
0 comments1 min readLW link

Dis­cus­sion: LLaMA Leak & Whistle­blow­ing in pre-AGI era

jirahim6 Mar 2023 18:47 UTC
1 point
4 comments1 min readLW link

[Question] Are we too con­fi­dent about un­al­igned AGI kil­ling off hu­man­ity?

RomanS6 Mar 2023 16:19 UTC
21 points
63 comments1 min readLW link

In­tro­duc­ing Leap Labs, an AI in­ter­pretabil­ity startup

Jessica Rumbelow6 Mar 2023 16:16 UTC
103 points
12 comments1 min readLW link

Monthly Roundup #4: March 2023

Zvi6 Mar 2023 14:10 UTC
31 points
0 comments24 min readLW link
(thezvi.wordpress.com)

Fun­da­men­tal Uncer­tainty: Chap­ter 6 - How can we be cer­tain about the truth?

Gordon Seidoh Worley6 Mar 2023 13:52 UTC
10 points
18 comments16 min readLW link

The idea

JNS6 Mar 2023 13:42 UTC
3 points
0 comments9 min readLW link

Hon­esty, Open­ness, Trust­wor­thi­ness, and Secrets

NormanPerlmutter6 Mar 2023 9:03 UTC
13 points
0 comments9 min readLW link

EA & LW Fo­rum Weekly Sum­mary (27th Feb − 5th Mar 2023)

Zoe Williams6 Mar 2023 3:18 UTC
12 points
0 comments1 min readLW link

The Type II In­ner-Com­pass Theorem

Tristan Miano6 Mar 2023 2:35 UTC
−16 points
0 comments22 min readLW link

AGI’s Im­pact on Em­ploy­ment

TheUnkown 6 Mar 2023 1:56 UTC
1 point
1 comment1 min readLW link
(www.apricitas.io)

Why did you trash the old HPMOR.com?

AnnoyedReader6 Mar 2023 1:55 UTC
55 points
68 comments2 min readLW link

Cap Model Size for AI Safety

research_prime_space6 Mar 2023 1:11 UTC
0 points
4 comments1 min readLW link

What should we do about net­work-effect mo­nop­o­lies?

benkuhn6 Mar 2023 0:50 UTC
31 points
7 comments1 min readLW link
(www.benkuhn.net)

Who Aligns the Align­ment Re­searchers?

Ben Smith5 Mar 2023 23:22 UTC
48 points
0 comments11 min readLW link

Star­tups are like firewood

Adam Zerner5 Mar 2023 23:09 UTC
26 points
2 comments3 min readLW link

A con­cern­ing ob­ser­va­tion from me­dia cov­er­age of AI in­dus­try dynamics

Justin Olive5 Mar 2023 21:38 UTC
8 points
3 comments3 min readLW link

Steven Pinker on ChatGPT and AGI (Feb 2023)

Evan R. Murphy5 Mar 2023 21:34 UTC
11 points
8 comments1 min readLW link
(news.harvard.edu)

Is it time to talk about AI dooms­day prep­ping yet?

bokov5 Mar 2023 21:17 UTC
0 points
8 comments1 min readLW link

Co­or­di­na­tion ex­plo­sion be­fore in­tel­li­gence ex­plo­sion...?

tailcalled5 Mar 2023 20:48 UTC
47 points
9 comments2 min readLW link

The Ogdoad

Tristan Miano5 Mar 2023 20:01 UTC
−15 points
1 comment37 min readLW link

[Question] What are some good ways to heighten my emo­tions?

oh543215 Mar 2023 18:06 UTC
5 points
5 comments1 min readLW link

Re­search pro­posal: Lev­er­ag­ing Jun­gian archetypes to cre­ate val­ues-based models

MiguelDev5 Mar 2023 17:39 UTC
5 points
2 comments2 min readLW link

Abus­ing Snap Cir­cuits IC

jefftk5 Mar 2023 17:00 UTC
19 points
3 comments3 min readLW link
(www.jefftk.com)

Do hu­mans de­rive val­ues from fic­ti­tious im­puted co­her­ence?

TsviBT5 Mar 2023 15:23 UTC
45 points
8 comments14 min readLW link

The In­ner-Com­pass Theorem

Tristan Miano5 Mar 2023 15:21 UTC
−18 points
12 comments16 min readLW link

Hal­i­fax Monthly Meetup: AI Safety Discussion

Ideopunk5 Mar 2023 12:42 UTC
10 points
0 comments1 min readLW link

Why kill ev­ery­one?

arisAlexis5 Mar 2023 11:53 UTC
7 points
5 comments2 min readLW link

Selec­tive, Cor­rec­tive, Struc­tural: Three Ways of Mak­ing So­cial Sys­tems Work

Said Achmiz5 Mar 2023 8:45 UTC
99 points
13 comments2 min readLW link

Sub­sti­tute goods for leisure are abundant

Adam Zerner5 Mar 2023 3:45 UTC
20 points
7 comments5 min readLW link

[Question] Does polyamory at a work­place turn nepo­tism up to eleven?

Viliam5 Mar 2023 0:57 UTC
45 points
11 comments2 min readLW link

Why We MUST Build an (al­igned) Ar­tifi­cial Su­per­in­tel­li­gence That Takes Over Hu­man So­ciety—A Thought Experiment

twkaiser5 Mar 2023 0:47 UTC
−13 points
12 comments2 min readLW link