RSS

Refine

TagLast edit: Sep 1, 2022, 2:27 PM by NeuralBets

Refine is a conceptual research incubator hosted by Conjecture.

Refine’s Se­cond Blog Post Day

adamShimiAug 20, 2022, 1:01 PM
19 points
0 comments1 min readLW link

Bench­mark­ing Pro­pos­als on Risk Scenarios

Paul BricmanAug 20, 2022, 10:01 AM
25 points
2 comments14 min readLW link

What if we ap­proach AI safety like a tech­ni­cal en­g­ineer­ing safety problem

zeshenAug 20, 2022, 10:29 AM
36 points
4 comments7 min readLW link

What I Learned Run­ning Refine

adamShimiNov 24, 2022, 2:49 PM
108 points
5 comments4 min readLW link

Refine Blog­post Day #3: The short­forms I did write

Alexander Gietelink OldenzielSep 16, 2022, 9:03 PM
23 points
0 comments1 min readLW link

Refine’s Third Blog Post Day/​Week

adamShimiSep 17, 2022, 5:03 PM
18 points
0 comments1 min readLW link

Rep­re­sen­ta­tional Tethers: Ty­ing AI La­tents To Hu­man Ones

Paul BricmanSep 16, 2022, 2:45 PM
30 points
0 comments16 min readLW link

Epistemic Arte­facts of (con­cep­tual) AI al­ign­ment research

Aug 19, 2022, 5:18 PM
31 points
1 comment5 min readLW link

Over­sight Leagues: The Train­ing Game as a Feature

Paul BricmanSep 9, 2022, 10:08 AM
20 points
6 comments10 min readLW link

Ide­olog­i­cal In­fer­ence Eng­ines: Mak­ing Deon­tol­ogy Differ­en­tiable*

Paul BricmanSep 12, 2022, 12:00 PM
6 points
0 comments14 min readLW link

Levels of goals and alignment

zeshenSep 16, 2022, 4:44 PM
27 points
4 comments6 min readLW link

Cat­a­logu­ing Pri­ors in The­ory and Practice

Paul BricmanOct 13, 2022, 12:36 PM
13 points
8 comments7 min readLW link

Refine: what helped me write more?

Alexander Gietelink OldenzielOct 25, 2022, 2:44 PM
12 points
0 comments2 min readLW link

Embed­ding safety in ML development

zeshenOct 31, 2022, 12:27 PM
24 points
1 comment18 min readLW link

A new­comer’s guide to the tech­ni­cal AI safety field

zeshenNov 4, 2022, 2:29 PM
42 points
3 comments10 min readLW link

In­ter­lude: But Who Op­ti­mizes The Op­ti­mizer?

Paul BricmanSep 23, 2022, 3:30 PM
15 points
0 comments10 min readLW link

Sum­mary of ML Safety Course

zeshenSep 27, 2022, 1:05 PM
7 points
0 comments6 min readLW link

My Thoughts on the ML Safety Course

zeshenSep 27, 2022, 1:15 PM
50 points
3 comments17 min readLW link

(Struc­tural) Sta­bil­ity of Cou­pled Optimizers

Paul BricmanSep 30, 2022, 11:28 AM
25 points
0 comments10 min readLW link

Refine’s First Blog Post Day

adamShimiAug 13, 2022, 10:23 AM
55 points
3 comments1 min readLW link

How I think about alignment

Linda LinseforsAug 13, 2022, 10:01 AM
31 points
11 comments5 min readLW link

Steelmin­ing via Analogy

Paul BricmanAug 13, 2022, 9:59 AM
24 points
0 comments2 min readLW link
(paulbricman.com)

I missed the crux of the al­ign­ment prob­lem the whole time

zeshenAug 13, 2022, 10:11 AM
53 points
7 comments3 min readLW link

All the posts I will never write

Alexander Gietelink OldenzielAug 14, 2022, 6:29 PM
54 points
8 comments8 min readLW link

Refine: An In­cu­ba­tor for Con­cep­tual Align­ment Re­search Bets

adamShimiApr 15, 2022, 8:57 AM
144 points
13 comments4 min readLW link

How to Diver­sify Con­cep­tual Align­ment: the Model Be­hind Refine

adamShimiJul 20, 2022, 10:44 AM
87 points
11 comments8 min readLW link
No comments.