I really like this in that it’s approaching an issue I view as currently neglected within AI safety research: how to determine human values to be learned. Like Kaj I find this a bit hard to engage with specific issues to give feedback, but I look forward to where this goes since I expect us to eventually need more formal approaches to axiology, even if they are only “adequate”.
I really like this in that it’s approaching an issue I view as currently neglected within AI safety research: how to determine human values to be learned. Like Kaj I find this a bit hard to engage with specific issues to give feedback, but I look forward to where this goes since I expect us to eventually need more formal approaches to axiology, even if they are only “adequate”.