I’m just trying to get the problem you’re presenting. Is it that in the event of a foom, a self-improving AI always presents a threat of having its values drift far enough away from humanity’s that it will endanger the human race? And your goal is to create the set of values that allow for both self-improvement and friendliness? And to do this, you must not only create the AI architecture but influence the greater system of AI creation as well? I’m not involved in AI research in any capacity, I just want to see if I understand the fundamentals of what you’re discussing.
I’m just trying to get the problem you’re presenting. Is it that in the event of a foom, a self-improving AI always presents a threat of having its values drift far enough away from humanity’s that it will endanger the human race? And your goal is to create the set of values that allow for both self-improvement and friendliness? And to do this, you must not only create the AI architecture but influence the greater system of AI creation as well? I’m not involved in AI research in any capacity, I just want to see if I understand the fundamentals of what you’re discussing.