Acknowledging that humans have a variety of values doesn’t invalidate the alignment problem; it makes it harder. Sophisticated discussions of the alignment problem address this additional difficulty.
(read abstracts fast, maybe level1 skim papers that you’re skeptical of but sound interesting, in order to use the above link. hit “more like this” if one sounds kinda right but not quite like what you seek.)
The alignment problem assumes there is such a thing as a human value system, rather than lots of individual ones.
Acknowledging that humans have a variety of values doesn’t invalidate the alignment problem; it makes it harder. Sophisticated discussions of the alignment problem address this additional difficulty.
...means there are at least two alignment problems.
alignment is fundamentally a question of multi-agent systems first, in my view.
(read abstracts fast, maybe level1 skim papers that you’re skeptical of but sound interesting, in order to use the above link. hit “more like this” if one sounds kinda right but not quite like what you seek.)