Suppose we had a matrix where each column corresponds to an AI safety problem and each row corresponds to a proposed safety measure.
Anyone have a sense of what distinct set of AI safety problems we’re faced with? I’ve tried my hand at this but the way I usually think about AI safety is a mapping between one very big problem (getting AI not to do things we don’t want it to do) and tens or so of low-probability-of-success safety measures trying to address it (in the veins of defining the things we don’t want AI to do and limiting the number of things that AI can do).
Should I be doing this differently? Maybe we could try to explicitly define the kinds of superintelligent AI we expect humanity to build and then design specific safety measures for each use-case. This would let us build a matrix but I’m also worried that superintelligent AI applications are too general to actually carve out a finite set of use-cases from the spectrum of possible AIs we could develop.
Anyone have a sense of what distinct set of AI safety problems we’re faced with?
See section 4, “Problems with AGI”, in this review for a list of lists.
However, I suspect the thing would work best in conjunction with a particular proposed FAI design where each column corresponds to a potential safety problem people are worried about with it.
Anyone have a sense of what distinct set of AI safety problems we’re faced with? I’ve tried my hand at this but the way I usually think about AI safety is a mapping between one very big problem (getting AI not to do things we don’t want it to do) and tens or so of low-probability-of-success safety measures trying to address it (in the veins of defining the things we don’t want AI to do and limiting the number of things that AI can do).
Should I be doing this differently? Maybe we could try to explicitly define the kinds of superintelligent AI we expect humanity to build and then design specific safety measures for each use-case. This would let us build a matrix but I’m also worried that superintelligent AI applications are too general to actually carve out a finite set of use-cases from the spectrum of possible AIs we could develop.
See section 4, “Problems with AGI”, in this review for a list of lists.
However, I suspect the thing would work best in conjunction with a particular proposed FAI design where each column corresponds to a potential safety problem people are worried about with it.