What I expect from formal “analytic philosophy” methods:
1) A useful decomposition of the issue into problems and subproblems (eg AI goal stability, AI agency, reduced impact, correct physical models on the universe, correct models of fuzzy human concepts such as human beings, convergence or divergence of goals, etc...)
2) Full or partial solutions some of the subproblems, ideally of general applicability (so they can be added easily to any AI design).
3) A good understanding of the remaining holes.
and lastly:
4) Exposing the implicit assumptions in proposed (non-analytic) solutions to the AI risk problem, so that the naive approaches can be discarded and the better approaches improved.
My response in the comment section:
What I expect from formal “analytic philosophy” methods:
1) A useful decomposition of the issue into problems and subproblems (eg AI goal stability, AI agency, reduced impact, correct physical models on the universe, correct models of fuzzy human concepts such as human beings, convergence or divergence of goals, etc...)
2) Full or partial solutions some of the subproblems, ideally of general applicability (so they can be added easily to any AI design).
3) A good understanding of the remaining holes.
and lastly:
4) Exposing the implicit assumptions in proposed (non-analytic) solutions to the AI risk problem, so that the naive approaches can be discarded and the better approaches improved.
Ben expanded his original article by editing a reply to your points into the end.