Scal­ing pre­dic­tion mar­kets with meta-markets

DentosalOct 10, 2024, 9:17 PM
1 point
0 comments2 min readLW link

Startup Suc­cess Rates Are So Low Be­cause the Re­wards Are So Large

AppliedDivinityStudiesOct 10, 2024, 8:22 PM
42 points
6 comments2 min readLW link

Can AI Out­pre­dict Hu­mans? Re­sults From Me­tac­u­lus’s Q3 AI Fore­cast­ing Benchmark

ChristianWilliamsOct 10, 2024, 6:58 PM
50 points
2 comments1 min readLW link
(www.metaculus.com)

Ra­tion­al­ity Quotes—Fall 2024

ScrewtapeOct 10, 2024, 6:37 PM
79 points
26 comments1 min readLW link

[Question] why won’t this al­ign­ment plan work?

KvmanThinkingOct 10, 2024, 3:44 PM
8 points
7 comments1 min readLW link

AI #85: AI Wins the No­bel Prize

ZviOct 10, 2024, 1:40 PM
30 points
6 comments31 min readLW link
(thezvi.wordpress.com)

Be­hav­ioral red-team­ing is un­likely to pro­duce clear, strong ev­i­dence that mod­els aren’t scheming

BuckOct 10, 2024, 1:36 PM
100 points
4 comments13 min readLW link

Joshua Achiam Public State­ment Analysis

ZviOct 10, 2024, 12:50 PM
73 points
14 comments21 min readLW link
(thezvi.wordpress.com)

Do you want to do a de­bate on youtube? I’m look­ing for po­lite, truth-seek­ing par­ti­ci­pants.

Nathan YoungOct 10, 2024, 9:32 AM
12 points
0 comments1 min readLW link

Ra­tion­al­ist Gnosticism

tailcalledOct 10, 2024, 9:06 AM
9 points
10 comments3 min readLW link

The deep­est athe­ist: Sam Altman

Trey EdwinOct 10, 2024, 3:27 AM
14 points
2 comments4 min readLW link

Values Are Real Like Harry Potter

Oct 9, 2024, 11:42 PM
83 points
20 comments5 min readLW link

Mo­men­tum of Light in Glass

BenOct 9, 2024, 8:19 PM
143 points
44 comments11 min readLW link

vgillioz’s Shortform

vgilliozOct 9, 2024, 7:31 PM
1 point
2 comments1 min readLW link

Hamil­to­nian Dy­nam­ics in AI: A Novel Ap­proach to Op­ti­miz­ing Rea­son­ing in Lan­guage Models

Javier Marin ValenzuelaOct 9, 2024, 7:14 PM
3 points
0 comments10 min readLW link

Tri­an­gu­lat­ing My In­ter­pre­ta­tion of Meth­ods: Black Boxes by Marco J. Nathan

adamShimiOct 9, 2024, 7:13 PM
8 points
0 comments6 min readLW link
(formethods.substack.com)

Scaf­fold­ing for “Notic­ing Me­tacog­ni­tion”

RaemonOct 9, 2024, 5:54 PM
80 points
4 comments17 min readLW link

Safe Pre­dic­tive Agents with Joint Scor­ing Rules

Rubi J. HudsonOct 9, 2024, 4:38 PM
55 points
10 comments17 min readLW link

Demis Hass­abis and Ge­offrey Hin­ton Awarded No­bel Prizes

Anna GajdovaOct 9, 2024, 12:56 PM
48 points
14 comments1 min readLW link

Hu­mans are (mostly) metarational

Yair HalberstadtOct 9, 2024, 5:51 AM
14 points
6 comments3 min readLW link

[Job Ad] MATS is hiring!

Oct 9, 2024, 2:17 AM
10 points
0 comments5 min readLW link

Pal­isade is hiring: Exec As­sis­tant, Con­tent Lead, Ops Lead, and Policy Lead

Charlie Rogers-SmithOct 9, 2024, 12:04 AM
11 points
0 comments4 min readLW link

AGI & Con­scious­ness—Joscha Bach

Rahul ChandOct 8, 2024, 10:51 PM
1 point
0 comments10 min readLW link

Video and tran­script of pre­sen­ta­tion on Oth­er­ness and con­trol in the age of AGI

Joe CarlsmithOct 8, 2024, 10:30 PM
35 points
1 comment27 min readLW link

From seeded com­plex­ity to con­scious­ness—yes, it’s all the same.

eschatailOct 8, 2024, 9:31 PM
−23 points
0 comments2 min readLW link

Limits of safe and al­igned AI

ShivamOct 8, 2024, 9:30 PM
2 points
0 comments4 min readLW link

[Question] What con­sti­tutes an in­fo­haz­ard?

K1r4d4rk.v1Oct 8, 2024, 9:29 PM
−4 points
8 comments1 min readLW link

[Question] What makes one a “ra­tio­nal­ist”?

mathyoufOct 8, 2024, 8:25 PM
7 points
5 comments3 min readLW link

[In­tu­itive self-mod­els] 4. Trance

Steven ByrnesOct 8, 2024, 1:30 PM
75 points
7 comments24 min readLW link

Schel­ling game eval­u­a­tions for AI control

Olli JärviniemiOct 8, 2024, 12:01 PM
65 points
5 comments11 min readLW link

Think­ing About a Pedalboard

jefftkOct 8, 2024, 11:50 AM
9 points
2 comments1 min readLW link
(www.jefftk.com)

Overview of strong hu­man in­tel­li­gence am­plifi­ca­tion methods

TsviBTOct 8, 2024, 8:37 AM
271 points
142 comments10 min readLW link

Near-death experiences

Declan MolonyOct 8, 2024, 6:34 AM
3 points
1 comment2 min readLW link

The un­rea­son­able effec­tive­ness of plas­mid se­quenc­ing as a service

Abhishaike MahajanOct 8, 2024, 2:02 AM
23 points
2 comments13 min readLW link
(www.owlposting.com)

There is a globe in your LLM

jacob_droriOct 8, 2024, 12:43 AM
86 points
4 comments1 min readLW link

MATS AI Safety Strat­egy Cur­ricu­lum v2

Oct 7, 2024, 10:44 PM
42 points
6 comments13 min readLW link

2025 Color Trends

sarahconstantinOct 7, 2024, 9:20 PM
40 points
7 comments6 min readLW link
(sarahconstantin.substack.com)

Clar­ify­ing Align­ment Fun­da­men­tals Through the Lens of Ontology

eternal/ephemeraOct 7, 2024, 8:57 PM
12 points
4 comments24 min readLW link

Ethics on Cos­mic Scale, Outer Space Treaty, Directed Pansper­mia, For­wards-Con­tam­i­na­tion, Tech­nol­ogy Assess­ment, Plane­tary Pro­tec­tion, and Fermi’s Paradox

MrFantasticOct 7, 2024, 8:56 PM
−12 points
0 comments1 min readLW link

Do­main-spe­cific SAEs

jacob_droriOct 7, 2024, 8:15 PM
27 points
0 comments5 min readLW link

Me­tac­u­lus Is Open Source

ChristianWilliamsOct 7, 2024, 7:55 PM
13 points
0 comments1 min readLW link
(www.metaculus.com)

Re­search up­date: Towards a Law of Iter­ated Ex­pec­ta­tions for Heuris­tic Estimators

Eric NeymanOct 7, 2024, 7:29 PM
87 points
2 comments22 min readLW link

AI Model Registries: A Foun­da­tional Tool for AI Governance

Oct 7, 2024, 7:27 PM
20 points
1 comment4 min readLW link
(www.convergenceanalysis.org)

Eval­u­at­ing the truth of state­ments in a world of am­bigu­ous lan­guage.

HastingsOct 7, 2024, 6:08 PM
48 points
19 comments2 min readLW link

Ad­vice for journalists

Nathan YoungOct 7, 2024, 4:46 PM
100 points
53 comments9 min readLW link
(nathanpmyoung.substack.com)

Time Effi­cient Re­sis­tance Training

romeostevensitOct 7, 2024, 3:15 PM
42 points
10 comments3 min readLW link

A Nar­row Path: a plan to deal with AI ex­tinc­tion risk

Oct 7, 2024, 1:02 PM
73 points
12 comments2 min readLW link
(www.narrowpath.co)

Toy Models of Fea­ture Ab­sorp­tion in SAEs

Oct 7, 2024, 9:56 AM
49 points
8 comments10 min readLW link

An ar­gu­ment that con­se­quen­tial­ism is incomplete

cousin_itOct 7, 2024, 9:45 AM
32 points
27 comments1 min readLW link

An X-Ray is Worth 15 Fea­tures: Sparse Au­toen­coders for In­ter­pretable Ra­diol­ogy Re­port Generation

Oct 7, 2024, 8:53 AM
38 points
0 comments5 min readLW link
(arxiv.org)