Yeah, it’s hard to extrapolate how an algorithm “should” handle inputs it wasn’t designed for, especially if you’re the algorithm :-/ As far as I understand Wei’s position, he hopes that we can formalize human philosophical abilities and delegate to that, rather than extend preferences directly which would be vulnerable. See this old thread, especially steven0461′s reply, and this post by Wei which fleshes out the idea.
Yeah, it’s hard to extrapolate how an algorithm “should” handle inputs it wasn’t designed for, especially if you’re the algorithm :-/ As far as I understand Wei’s position, he hopes that we can formalize human philosophical abilities and delegate to that, rather than extend preferences directly which would be vulnerable. See this old thread, especially steven0461′s reply, and this post by Wei which fleshes out the idea.
Thanks for those links.