It was in the references that initially didn’t make the cut. After further thought, it’s indeed worth adding. I referenced the Distill article AI Safety Needs Social Scientists, which spends more time on the motivating arguments, and linked to the paper in the note.
No worries. As much as I think less has been written on debate than amplification (Paul has a lot of blog posts on IDA), it seems to me like most of the work Paul’s team at OpenAI is doing is working on debates rather than IDA.
I don’t know whether this is on purpose, but I’d think that AI Safety Via Debate (original paper: https://arxiv.org/abs/1805.00899; recent progress report: https://www.lesswrong.com/posts/Br4xDbYu4Frwrb64a/writeup-progress-on-ai-safety-via-debate-1) should get a mention, probably in the Technical agendas focused on possible solutions section? I’d argue it’s different enough from IDA to have it’s own subititle.
It was in the references that initially didn’t make the cut. After further thought, it’s indeed worth adding. I referenced the Distill article AI Safety Needs Social Scientists, which spends more time on the motivating arguments, and linked to the paper in the note.
Thanks for your feedback!
No worries. As much as I think less has been written on debate than amplification (Paul has a lot of blog posts on IDA), it seems to me like most of the work Paul’s team at OpenAI is doing is working on debates rather than IDA.