“The type of system humans are” has a big disadvantage compared to mathematically simpler systems like maximizers: it seems harder to reason about such “fuzzy” systems, e.g. prove their safety. How do you convince yourself that a “fuzzy” superintelligence is safe to run?
We have an existence proof of intelligences based upon “The type of systems humans are”, we don’t for pure maximizers. It is no good trying to develop friendliness theory based upon a pure easily reasoned about system if you can’t make an intelligence out of it.
So while it is harder, this may be the sort of system we have to deal with. It is these sorts of questions I wanted to try to answer with the group in my original post.
I’ll try to explain why I am sceptical of maximizer based intelligences in a discussion post. It is not because they are inhuman.
“The type of system humans are” has a big disadvantage compared to mathematically simpler systems like maximizers: it seems harder to reason about such “fuzzy” systems, e.g. prove their safety. How do you convince yourself that a “fuzzy” superintelligence is safe to run?
We have an existence proof of intelligences based upon “The type of systems humans are”, we don’t for pure maximizers. It is no good trying to develop friendliness theory based upon a pure easily reasoned about system if you can’t make an intelligence out of it.
So while it is harder, this may be the sort of system we have to deal with. It is these sorts of questions I wanted to try to answer with the group in my original post.
I’ll try to explain why I am sceptical of maximizer based intelligences in a discussion post. It is not because they are inhuman.
Well, it’s also hard to prove the safety of maximizers. Proving the danger, on the other hand...