I should probably make clear that most of my knowledge of AI comes from LW posts, I do not work with it professionally, and that this discussion is on my part motivated by curiosity and desire to learn.
Emotions are a core part of how humans make decisions.
Agreed.
Your assessment is probably more accurate than mine.
My original line was of thinking was that while AIs might use quick-and-imprecise thinking shortcuts triggered by pattern-matching (which is sort of how I see emotions), human emotions are too inconveniently packaged to be much of use in AI design. (While being necessary, they also misfire a lot; coping with emotions is an important skill to learn; in some situations emotions do more harm than good; all in all this doesn’t seem like good mind design). So I was wondering if whatever AI uses for its thinking, we would even recognize as emotions.
My assessment now is that even if AI uses different thinking shortcuts than humans do, they might still misfire. For example, I can imagine a pattern activation triggering more patterns, which in turn trigger more and more patterns, resulting in a cascade effect not unlike emotional over-stimulation/breakdown in humans. So I think it’s possible that we might see AI having what we would describe as emotions (perhaps somewhat uncanny emotions, but emotions all the same).
P. S. For the sake of completeness: my mental model also includes biological organisms needing emotions in order to create motivation (rather than just drawing conclusions). (example: fear creating motivation to escape danger). AI should already have a supergoal so it does not need “motivation”. However it would need to see how its current context connects to its supergoal, and create/activate subgoals that apply to the current situation, and here once again thinking shortcuts might be useful, perhaps not too unlike human emotions.
Example: AI sees a fast-moving object that it predicts will intersect its current location, and a thinking shortcut activates a dodging strategy. This is a subgoal to the goal of surviving, which is in turn is a subgoal to the AI’s supergoal (whatever that is).
Having a thinking shortcut (this one we might call “reflex” rather “emotion”) results in faster thinking. Slow thinking might be inefficient to the point of fatal “Hm… that object seems to be moving mighty fast in my direction… if it hits me it might damage/destroy me. Would that be a good thing? No, I guess not—I need to functional in order to achieve my supergoal. So I should probably dodg.. ”
AI should already have a supergoal so it does not need “motivation”.
We know relatively little about what it takes to create a AGI. Saying that an AGI should have feature X or feature Y to be a functioning AGI is drawing to much conclusions from the data we have.
On the other hand we know that the architecture on which humans run produces “intelligence” so that at least one possible architecture that could be implemented in a computer.
Bootstraping AGI from Whole Brain Emulations is on of the ideas that is in discussion even on LessWrong.
I should probably make clear that most of my knowledge of AI comes from LW posts, I do not work with it professionally, and that this discussion is on my part motivated by curiosity and desire to learn.
Agreed.
Your assessment is probably more accurate than mine.
My original line was of thinking was that while AIs might use quick-and-imprecise thinking shortcuts triggered by pattern-matching (which is sort of how I see emotions), human emotions are too inconveniently packaged to be much of use in AI design. (While being necessary, they also misfire a lot; coping with emotions is an important skill to learn; in some situations emotions do more harm than good; all in all this doesn’t seem like good mind design). So I was wondering if whatever AI uses for its thinking, we would even recognize as emotions.
My assessment now is that even if AI uses different thinking shortcuts than humans do, they might still misfire. For example, I can imagine a pattern activation triggering more patterns, which in turn trigger more and more patterns, resulting in a cascade effect not unlike emotional over-stimulation/breakdown in humans.
So I think it’s possible that we might see AI having what we would describe as emotions (perhaps somewhat uncanny emotions, but emotions all the same).
P. S. For the sake of completeness: my mental model also includes biological organisms needing emotions in order to create motivation (rather than just drawing conclusions). (example: fear creating motivation to escape danger).
AI should already have a supergoal so it does not need “motivation”. However it would need to see how its current context connects to its supergoal, and create/activate subgoals that apply to the current situation, and here once again thinking shortcuts might be useful, perhaps not too unlike human emotions.
Example: AI sees a fast-moving object that it predicts will intersect its current location, and a thinking shortcut activates a dodging strategy. This is a subgoal to the goal of surviving, which is in turn is a subgoal to the AI’s supergoal (whatever that is).
Having a thinking shortcut (this one we might call “reflex” rather “emotion”) results in faster thinking. Slow thinking might be inefficient to the point of fatal “Hm… that object seems to be moving mighty fast in my direction… if it hits me it might damage/destroy me. Would that be a good thing? No, I guess not—I need to functional in order to achieve my supergoal. So I should probably dodg.. ”
We know relatively little about what it takes to create a AGI. Saying that an AGI should have feature X or feature Y to be a functioning AGI is drawing to much conclusions from the data we have.
On the other hand we know that the architecture on which humans run produces “intelligence” so that at least one possible architecture that could be implemented in a computer.
Bootstraping AGI from Whole Brain Emulations is on of the ideas that is in discussion even on LessWrong.