Curated. Short and sweet summary of a research approach being pursued in AI alignment, that I think has not been written up like this before (especially on LessWrong).
The slightly longer and slightly worse version is the pair of Human-AI Interaction and Reward uncertainty.
Nonetheless, thanks for the links :)
Curated. Short and sweet summary of a research approach being pursued in AI alignment, that I think has not been written up like this before (especially on LessWrong).
The slightly longer and slightly worse version is the pair of Human-AI Interaction and Reward uncertainty.
Nonetheless, thanks for the links :)