I am optimistic that further thinking on automation prospects could identify other automation-tractable areas of alignment and control (e.g. see here for previous work).
This tag might be helpful: https://www.lesswrong.com/w/ai-assisted-alignment
Here’s a recent shortform on the topic: https://www.lesswrong.com/posts/mKgbawbJBxEmQaLSJ/davekasten-s-shortform?commentId=32jReMrHDd5vkDBwt
I wonder about getting an LLM to process LW archive posts, and tag posts which contain alignment ideas that seem automatable.
If people start losing jobs from automation, that could finally build political momentum for serious regulation.
Suggested in Zvi’s comments the other month (22 likes):
Source: https://thezvi.substack.com/p/the-paris-ai-anti-safety-summit/comment/92963364
Just skimming the thread, I didn’t see anyone offer a serious attempt at counterargument, either.