Thanks Koen for your feedback! You make a great point about a clearer call to action for RL researchers. I think an immediate call to action is to be aware of the following:
there is a broader scope of aligned RL agent design
there are difficult unsolved problems in this broader scope
for sufficiently advanced agents, these problems need general solutions rather than ad-hoc ones
Then a long-term call to action (if/when they are in the position to deploy an advanced AI system) is to consider the broader scope and look for general solutions to specification problems rather than deploying ad-hoc solutions. For those general solutions, they could refer to the safety literature and/or consult the safety community.
Thanks Koen for your feedback! You make a great point about a clearer call to action for RL researchers. I think an immediate call to action is to be aware of the following:
there is a broader scope of aligned RL agent design
there are difficult unsolved problems in this broader scope
for sufficiently advanced agents, these problems need general solutions rather than ad-hoc ones
Then a long-term call to action (if/when they are in the position to deploy an advanced AI system) is to consider the broader scope and look for general solutions to specification problems rather than deploying ad-hoc solutions. For those general solutions, they could refer to the safety literature and/or consult the safety community.