I think AI control agendas are defined in such a way such that this metric isn’t as relevant as you think it is:
I think it is unlikely for control work to buy humanity much time until someone builds a very powerful unaligned AI system, at least at our present levels of coordination tools.
Because the agenda isn’t trying to make AIs alignable, but to make them useful and not break out of labs, so the question of the timeline to unaligned AI is less relevant than it is for most methods of making safe AI.
I think AI control agendas are defined in such a way such that this metric isn’t as relevant as you think it is:
Because the agenda isn’t trying to make AIs alignable, but to make them useful and not break out of labs, so the question of the timeline to unaligned AI is less relevant than it is for most methods of making safe AI.