We have just released an academic survey of AI alignment.
We identify four main categories of alignment research:
Learning from feedback (e.g. scalable oversight)
Learning under distribution shift
Assurance (e.g. interpretability)
Governance
We mainly focused on academic references but also included some posts from LessWrong and other forums. We would love to hear from the community about any references we missed or anything that was unclear or misstated. We hope that this can be a good starting point for AI researchers who might be unfamiliar with current efforts in AI alignment.
AI Alignment: A Comprehensive Survey
Link post
We have just released an academic survey of AI alignment.
We identify four main categories of alignment research:
Learning from feedback (e.g. scalable oversight)
Learning under distribution shift
Assurance (e.g. interpretability)
Governance
We mainly focused on academic references but also included some posts from LessWrong and other forums. We would love to hear from the community about any references we missed or anything that was unclear or misstated. We hope that this can be a good starting point for AI researchers who might be unfamiliar with current efforts in AI alignment.