“Work on the safety of an ecosystem made up of a large number of in-some-ways-superhuman-and-in-other-ways-not AIs” seems like a very different problem than “ensure that when you build a single coherent, effectively-omniscient agent, you give it a goal that does not ruin everything when it optimizes really hard for that goal”.
There are definitely parallels between the two scenarios, but I’m not sure a solution for the second scenario would even work to prevent an organization of AIs with cognitive blind spots from going off the rails.
My model of jacob_cannell’s model is that the medium-term future looks something like “ad-hoc organizations of mostly-cooperating organizations of powerful-but-not-that-powerful agents, with the first organization to reach a given level of capability being the one that focused its resources on finding and using better coordination mechanisms between larger numbers of individual processes rather than the one that focused on raw predictive power”, and that his model of Eliezer goes “no, actually focusing on raw predictive power is the way to go”.
And I think the two different scenarios do in fact suggest different strategies.
“Work on the safety of an ecosystem made up of a large number of in-some-ways-superhuman-and-in-other-ways-not AIs” seems like a very different problem than “ensure that when you build a single coherent, effectively-omniscient agent, you give it a goal that does not ruin everything when it optimizes really hard for that goal”.
There are definitely parallels between the two scenarios, but I’m not sure a solution for the second scenario would even work to prevent an organization of AIs with cognitive blind spots from going off the rails.
My model of jacob_cannell’s model is that the medium-term future looks something like “ad-hoc organizations of mostly-cooperating organizations of powerful-but-not-that-powerful agents, with the first organization to reach a given level of capability being the one that focused its resources on finding and using better coordination mechanisms between larger numbers of individual processes rather than the one that focused on raw predictive power”, and that his model of Eliezer goes “no, actually focusing on raw predictive power is the way to go”.
And I think the two different scenarios do in fact suggest different strategies.