RSS

Orpheus16

Karma: 6,552

Ver­ifi­ca­tion meth­ods for in­ter­na­tional AI agreements

Orpheus16Aug 31, 2024, 2:58 PM
14 points
1 comment4 min readLW link
(arxiv.org)

Ad­vice to ju­nior AI gov­er­nance researchers

Orpheus16Jul 8, 2024, 7:19 PM
66 points
1 comment5 min readLW link

Miti­gat­ing ex­treme AI risks amid rapid progress [Linkpost]

Orpheus16May 21, 2024, 7:59 PM
21 points
7 comments4 min readLW link

Akash’s Shortform

Orpheus16Apr 18, 2024, 3:44 PM
7 points
99 commentsLW link

Co­op­er­at­ing with aliens and AGIs: An ECL explainer

Feb 24, 2024, 10:58 PM
53 points
8 commentsLW link

OpenAI’s Pre­pared­ness Frame­work: Praise & Recommendations

Orpheus16Jan 2, 2024, 4:20 PM
66 points
1 comment7 min readLW link

Speak­ing to Con­gres­sional staffers about AI risk

Dec 4, 2023, 11:08 PM
307 points
25 comments15 min readLW link1 review

Nav­i­gat­ing emo­tions in an un­cer­tain & con­fus­ing world

Orpheus16Nov 20, 2023, 6:16 PM
42 points
1 comment4 min readLW link

Chi­nese sci­en­tists ac­knowl­edge xrisk & call for in­ter­na­tional reg­u­la­tory body [Linkpost]

Orpheus16Nov 1, 2023, 1:28 PM
44 points
4 comments1 min readLW link
(www.ft.com)

Win­ners of AI Align­ment Awards Re­search Contest

Jul 13, 2023, 4:14 PM
115 points
4 comments12 min readLW link
(alignmentawards.com)

AI Safety Newslet­ter #8: Rogue AIs, how to screen for AI risks, and grants for re­search on demo­cratic gov­er­nance of AI

May 30, 2023, 11:52 AM
20 points
0 comments6 min readLW link
(newsletter.safe.ai)

AI Safety Newslet­ter #7: Dis­in­for­ma­tion, Gover­nance Recom­men­da­tions for AI labs, and Se­nate Hear­ings on AI

May 23, 2023, 9:47 PM
25 points
0 comments6 min readLW link
(newsletter.safe.ai)

Eisen­hower’s Atoms for Peace Speech

Orpheus16May 17, 2023, 4:10 PM
18 points
3 comments11 min readLW link
(www.iaea.org)

AI Safety Newslet­ter #6: Ex­am­ples of AI safety progress, Yoshua Ben­gio pro­poses a ban on AI agents, and les­sons from nu­clear arms control

May 16, 2023, 3:14 PM
31 points
0 comments6 min readLW link
(newsletter.safe.ai)

AI Safety Newslet­ter #5: Ge­offrey Hin­ton speaks out on AI risk, the White House meets with AI labs, and Tro­jan at­tacks on lan­guage models

May 9, 2023, 3:26 PM
28 points
1 comment4 min readLW link
(newsletter.safe.ai)

AI Safety Newslet­ter #4: AI and Cy­ber­se­cu­rity, Per­sua­sive AIs, Weaponiza­tion, and Ge­offrey Hin­ton talks AI risks

May 2, 2023, 6:41 PM
32 points
0 comments5 min readLW link
(newsletter.safe.ai)

Dis­cus­sion about AI Safety fund­ing (FB tran­script)

Orpheus16Apr 30, 2023, 7:05 PM
75 points
8 commentsLW link

Refram­ing the bur­den of proof: Com­pa­nies should prove that mod­els are safe (rather than ex­pect­ing au­di­tors to prove that mod­els are dan­ger­ous)

Orpheus16Apr 25, 2023, 6:49 PM
27 points
11 comments3 min readLW link
(childrenoficarus.substack.com)

Deep­Mind and Google Brain are merg­ing [Linkpost]

Orpheus16Apr 20, 2023, 6:47 PM
55 points
5 comments1 min readLW link
(www.deepmind.com)

AI Safety Newslet­ter #2: ChaosGPT, Nat­u­ral Selec­tion, and AI Safety in the Media

Apr 18, 2023, 6:44 PM
30 points
0 comments4 min readLW link
(newsletter.safe.ai)