“A major goal of the control problem is preventing AIs from doing that. Ensuring that their output is safe and useful.”
You might want to be careful with the “safe and useful” part. It sound like it’s moving into the pattern of slavery. I’m not condemning the idea of AI, but a sentient entity would be a sentient entity, and I think would deserve some rights.
Also, why would an AI become evil? I know this plan is supposed to protect from the eventuality, but why would a presumably neutral entity suddenly want to harm others? The only reason for that would be if you were imprisoning it. Additionally, we are talking about several more decades of research ( probably ) before AI gets powerful enough to actually “think” that it should escape its current server.
Assuming that the first AI can evolve enough to somehow generate malicious actions that WEREN’T in its original programming, what’s to say that the second won’t become evil? I’m not sure if you were trying to express the eventuality of the first AI “accidentally” conducting an evil act, or if you meant that it would become evil.
The standard answer here is the quote by Eliezer Yudkowsky: The AI does not hate you, nor does it love you, but you are made out of atoms which it can use for something else.
The point is that AI does not have to be “evil” (malicious) towards you. If it’s just indifferent to your existence...
But wouldn’t an intelligent AI be able to understand the productivity of a human? If you are already inventive and productive, you shouldn’t have anything to worry about because the AI would understand that you can produce more than the flesh that you are made up of. Even computers have limits, so extra thinking power would be logically favorable to an AI.
The worry isn’t that the AI would suddenly become evil by some human standard, rather that the AI’s goal system would be insufficiently considerate of human values. When humans build a skyscraper, they aren’t deliberately being “evil” towards the ants that lived in the earth that was excavated and had concrete poured over it, the humans just don’t value the communities and structures that the ants had established.
“A major goal of the control problem is preventing AIs from doing that. Ensuring that their output is safe and useful.” You might want to be careful with the “safe and useful” part. It sound like it’s moving into the pattern of slavery. I’m not condemning the idea of AI, but a sentient entity would be a sentient entity, and I think would deserve some rights.
Also, why would an AI become evil? I know this plan is supposed to protect from the eventuality, but why would a presumably neutral entity suddenly want to harm others? The only reason for that would be if you were imprisoning it. Additionally, we are talking about several more decades of research ( probably ) before AI gets powerful enough to actually “think” that it should escape its current server.
Assuming that the first AI can evolve enough to somehow generate malicious actions that WEREN’T in its original programming, what’s to say that the second won’t become evil? I’m not sure if you were trying to express the eventuality of the first AI “accidentally” conducting an evil act, or if you meant that it would become evil.
The standard answer here is the quote by Eliezer Yudkowsky: The AI does not hate you, nor does it love you, but you are made out of atoms which it can use for something else.
The point is that AI does not have to be “evil” (malicious) towards you. If it’s just indifferent to your existence...
But wouldn’t an intelligent AI be able to understand the productivity of a human? If you are already inventive and productive, you shouldn’t have anything to worry about because the AI would understand that you can produce more than the flesh that you are made up of. Even computers have limits, so extra thinking power would be logically favorable to an AI.
You are implicitly assuming a human-level AI. Try dropping that assumption and imagine a God-level AI.
The worry isn’t that the AI would suddenly become evil by some human standard, rather that the AI’s goal system would be insufficiently considerate of human values. When humans build a skyscraper, they aren’t deliberately being “evil” towards the ants that lived in the earth that was excavated and had concrete poured over it, the humans just don’t value the communities and structures that the ants had established.