RSS

jacquesthibs

Karma: 2,416

I work primarily on AI Alignment. Scroll down to my pinned Shortform for an idea of my current work and who I’d like to collaborate with.

Website: https://​​jacquesthibodeau.com

Twitter: https://​​twitter.com/​​JacquesThibs

GitHub: https://​​github.com/​​JayThibs

LinkedIn: https://​​www.linkedin.com/​​in/​​jacques-thibodeau/​​

How much I’m pay­ing for AI pro­duc­tivity soft­ware (and the fu­ture of AI use)

jacquesthibsOct 11, 2024, 5:11 PM
59 points
18 comments8 min readLW link
(jacquesthibodeau.com)

[Question] Shane Legg’s nec­es­sary prop­er­ties for ev­ery AGI Safety plan

jacquesthibsMay 1, 2024, 5:15 PM
58 points
12 comments1 min readLW link

AISC Pro­ject: Bench­marks for Stable Reflectivity

jacquesthibsNov 13, 2023, 2:51 PM
17 points
0 comments8 min readLW link

Re­search agenda: Su­per­vis­ing AIs im­prov­ing AIs

Apr 29, 2023, 5:09 PM
76 points
5 comments19 min readLW link

Paus­ing AI Devel­op­ments Isn’t Enough. We Need to Shut it All Down by Eliezer Yudkowsky

jacquesthibsMar 29, 2023, 11:16 PM
291 points
297 comments3 min readLW link
(time.com)

Prac­ti­cal Pit­falls of Causal Scrubbing

Mar 27, 2023, 7:47 AM
87 points
17 comments13 min readLW link

[Question] Can in­de­pen­dent re­searchers get a spon­sored visa for the US or UK?

jacquesthibsMar 24, 2023, 6:10 AM
23 points
1 comment1 min readLW link

[Question] What‘s in your list of un­solved prob­lems in AI al­ign­ment?

jacquesthibsMar 7, 2023, 6:58 PM
60 points
9 comments1 min readLW link

[Si­mu­la­tors sem­i­nar se­quence] #2 Semiotic physics—revamped

Feb 27, 2023, 12:25 AM
24 points
23 comments13 min readLW link

Kolb’s: an ap­proach to con­sciously get bet­ter at anything

jacquesthibsJan 3, 2023, 6:16 PM
12 points
1 comment6 min readLW link

[Si­mu­la­tors sem­i­nar se­quence] #1 Back­ground & shared assumptions

Jan 2, 2023, 11:48 PM
50 points
4 comments3 min readLW link

But is it re­ally in Rome? An in­ves­ti­ga­tion of the ROME model edit­ing technique

jacquesthibsDec 30, 2022, 2:40 AM
104 points
2 comments18 min readLW link

Re­sults from a sur­vey on tool use and work­flows in al­ign­ment research

Dec 19, 2022, 3:19 PM
79 points
2 comments19 min readLW link

[Question] How is ARC plan­ning to use ELK?

jacquesthibsDec 15, 2022, 8:11 PM
24 points
5 comments1 min readLW link

Fore­sight for AGI Safety Strat­egy: Miti­gat­ing Risks and Iden­ti­fy­ing Golden Opportunities

jacquesthibsDec 5, 2022, 4:09 PM
28 points
6 comments8 min readLW link

Is the “Valley of Con­fused Ab­strac­tions” real?

jacquesthibsDec 5, 2022, 1:36 PM
20 points
11 comments2 min readLW link

jacquesthibs’s Shortform

jacquesthibsNov 21, 2022, 12:04 PM
2 points
307 commentsLW link

A de­scrip­tive, not pre­scrip­tive, overview of cur­rent AI Align­ment Research

Jun 6, 2022, 9:59 PM
139 points
21 comments7 min readLW link

AI Align­ment YouTube Playlists

May 9, 2022, 9:33 PM
30 points
4 comments1 min readLW link

A sur­vey of tool use and work­flows in al­ign­ment research

Mar 23, 2022, 11:44 PM
45 points
4 comments1 min readLW link