Musk is strangely insightful about some things (electric cars becoming mainstream, reusable rockets being economically feasible), and strangely thoughtless about other things. If I’ve learned anything from digesting a few hundred blogs and podcasts on AI over the past 7 years, it’s that there is no single simple objective that captures what we actually want AI to do.
Curiosity is not going to cut it. Nor is “freedom”, which is what Musk was talking with Stuart Russel about maximizing a year or two ago. Human values are messy and context-dependent and not even internally consistent. If we actually want AI that fulfills the desires that humans would want after long reflection on the outcomes, it’s going to involve a lot of learning from human behavior, likely with design mistakes along the way which we will have to correct.
The hard part is not finding a single thing to maximize. The hard part is training an agent that is corrigible so that when we inevitably mess up, we don’t just die or enter a permanent dystopia on the first attempt.
On the Twitter spaces 2 days ago, a lot of emphasis seemed put on understanding which to me has a more humble conotation to me. Still I agree I would not bet on their luck with a choice of a single value to build their systems upon.( Although they have a luckers track record.)
Musk is strangely insightful about some things (electric cars becoming mainstream, reusable rockets being economically feasible), and strangely thoughtless about other things. If I’ve learned anything from digesting a few hundred blogs and podcasts on AI over the past 7 years, it’s that there is no single simple objective that captures what we actually want AI to do.
Curiosity is not going to cut it. Nor is “freedom”, which is what Musk was talking with Stuart Russel about maximizing a year or two ago. Human values are messy and context-dependent and not even internally consistent. If we actually want AI that fulfills the desires that humans would want after long reflection on the outcomes, it’s going to involve a lot of learning from human behavior, likely with design mistakes along the way which we will have to correct.
The hard part is not finding a single thing to maximize. The hard part is training an agent that is corrigible so that when we inevitably mess up, we don’t just die or enter a permanent dystopia on the first attempt.
On the Twitter spaces 2 days ago, a lot of emphasis seemed put on understanding which to me has a more humble conotation to me.
Still I agree I would not bet on their luck with a choice of a single value to build their systems upon.( Although they have a luckers track record.)