RSS

Caleb Biddulph

Karma: 711

Ves­ti­gial rea­son­ing in RL

Caleb BiddulphApr 13, 2025, 3:40 PM
50 points
7 comments9 min readLW link

Caleb Bid­dulph’s Shortform

Caleb BiddulphJan 30, 2025, 9:35 PM
4 points
23 comments1 min readLW link

[Question] Why not train rea­son­ing mod­els with RLHF?

Caleb BiddulphJan 30, 2025, 7:58 AM
4 points
4 comments1 min readLW link

Wor­ries about la­tent rea­son­ing in LLMs

Caleb BiddulphJan 20, 2025, 9:09 AM
42 points
3 comments7 min readLW link

5 ways to im­prove CoT faithfulness

Caleb BiddulphOct 5, 2024, 8:17 PM
42 points
40 comments6 min readLW link

OpenAI’s Sora is an agent

Caleb BiddulphFeb 16, 2024, 7:35 AM
96 points
25 comments4 min readLW link

Is Me­taethics Un­nec­es­sary Given In­tent-Aligned AI?

Caleb BiddulphSep 2, 2023, 9:48 AM
10 points
0 comments7 min readLW link

Prepar­ing for AI-as­sisted al­ign­ment re­search: we need data!

Caleb BiddulphJan 17, 2023, 3:28 AM
31 points
3 commentsLW link

The Ra­tional Utili­tar­ian Love Move­ment (A His­tor­i­cal Ret­ro­spec­tive)

Caleb BiddulphNov 3, 2022, 7:11 AM
3 points
0 commentsLW link