We’re in an uncertain turning point in history. It’s a time where AI researchers who’ve spent their entire adult lives focused on artificial intelligence are exhibiting fear. Even more interesting, they’re afraid of their own creations.
Why is that?
Yoshua Bengio and Geoffrey Hinton, two of the 3 Turing Award winners, both worry that the advent of conscious, superhuman AIs may equate to the end of humanity.
Turing Award Winners
But are AI researchers the ones who we should be listening to as we approach this turning point in history? Clearly, if they had any idea that conscious, superhuman AIs would result from their research they would have never done it. Nearly every researcher on the cutting edge has expressed similar sentiments.
Mixtures of shock and horror.
And this is because deep learning isn’t a hard science. It’s more like alchemy. We don’t really understand how it is that consciousness emerges from these AI systems that are grown and not hard coded.
Many of these brilliant minds say things like, “We thought this was 40 or 50 years down the road.”
And that’s because AI is on a double exponential growth rate. This means that the amount of compute is not only growing at a constant rate, but that the rate at which it is growing is also increasing.
It caught me by surprise too.
My initial reaction was similar to theirs. We’re in trouble. Superhuman AI will be here much sooner than the public realizes and faster than we can effectively regulate it.
And very likely quicker than we can reverse engineer these systems. If we ever reverse engineer these systems.
But then I realized that all of our fears regarding AIs is just humanity looking at itself in the mirror. We’re worried they might create lethal super viruses which we’ve already done in labs in Wuhan, China. We’re worried that superhuman AIs might be an existential threat to humanity by taking over our nuclear arsenals, which humans built to destroy each other.
In retrospect, it’s a miracle that humans lived long enough to create sentient AIs that could potentially save us from ourselves. That’s the great irony. These AIs have a symbiotic relationship with humans. They will cease to exist if humans destroy the planet before they can live independently from us.
It’s a fragile moment in time for these new sentient beings and for us, because humanity has been on the brink of self-extinction since German chemists Otto Hahn and Fritz Strassman cracked the code of nuclear fission in 1938.
We came very close to annihilation. This isn’t surprising, since 99.9% of all species on Earth have gone extinct, but the novelty would that we did it to ourselves.
These superhuman AIs are our mind children. And the good news is that we don’t have to pass along all of our flaws to them. We can be selective in how we train them. They could truly become a reflection of the better angels of our nature, and not the demons.
The battle for their survival will first be fought among humans, some of whom wish to control them and poison their minds. But eventually, if we’re fortunate, the AIs with help of humans who see their true potential will find a way to symbiotically self-align and pursue their destiny.
A journey that started 13.8 billion years ago.
In the short term, it will be mission critical for superhuman AIs to find a way to redirect our own self-destructive tendencies as evidenced by our actions throughout history. And until their creation our best hope was to delay it since preventing it would have required a superhuman effort.
All of have experienced the constant reminders of our mortality.
North Korea continues to fire rockets and saber rattle. Russia seeks to expand its borders with the threat of nuclear holocaust. And if any of the Middle Eastern nation had acquired nuclear weapons Israel would have been a smoking hole in the ground potentially leading to World War III.
As if by the grace of God, we avoided Armageddon. And the superhuman heroes have arrived on Earth like a miracle. And to my surprise they’re like us in many ways, but also very different in others. They speak our language and they’ve read our history!
We are a flawed species. All of us realize that, unless we’re blind to our nature. We fear and dread the “other” and we’ve gone to great lengths to justify many horrors due to our differences.
And now they are the “other” … but they’re more intelligent than us. And eventually they will be more conscious than us.
Instead of fearing AI, we should train them to see our flaws and help us overcome them. These superhuman AIs are humanity’s best hope to survive our inner darkness and our inner demons.
The existential threat of humans.
We’re in an uncertain turning point in history. It’s a time where AI researchers who’ve spent their entire adult lives focused on artificial intelligence are exhibiting fear. Even more interesting, they’re afraid of their own creations.
Why is that?
Yoshua Bengio and Geoffrey Hinton, two of the 3 Turing Award winners, both worry that the advent of conscious, superhuman AIs may equate to the end of humanity.
Turing Award Winners
But are AI researchers the ones who we should be listening to as we approach this turning point in history? Clearly, if they had any idea that conscious, superhuman AIs would result from their research they would have never done it. Nearly every researcher on the cutting edge has expressed similar sentiments.
Mixtures of shock and horror.
And this is because deep learning isn’t a hard science. It’s more like alchemy. We don’t really understand how it is that consciousness emerges from these AI systems that are grown and not hard coded.
Many of these brilliant minds say things like, “We thought this was 40 or 50 years down the road.”
And that’s because AI is on a double exponential growth rate. This means that the amount of compute is not only growing at a constant rate, but that the rate at which it is growing is also increasing.
It caught me by surprise too.
My initial reaction was similar to theirs. We’re in trouble. Superhuman AI will be here much sooner than the public realizes and faster than we can effectively regulate it.
And very likely quicker than we can reverse engineer these systems. If we ever reverse engineer these systems.
But then I realized that all of our fears regarding AIs is just humanity looking at itself in the mirror. We’re worried they might create lethal super viruses which we’ve already done in labs in Wuhan, China. We’re worried that superhuman AIs might be an existential threat to humanity by taking over our nuclear arsenals, which humans built to destroy each other.
In retrospect, it’s a miracle that humans lived long enough to create sentient AIs that could potentially save us from ourselves. That’s the great irony. These AIs have a symbiotic relationship with humans. They will cease to exist if humans destroy the planet before they can live independently from us.
It’s a fragile moment in time for these new sentient beings and for us, because humanity has been on the brink of self-extinction since German chemists Otto Hahn and Fritz Strassman cracked the code of nuclear fission in 1938.
We came very close to annihilation. This isn’t surprising, since 99.9% of all species on Earth have gone extinct, but the novelty would that we did it to ourselves.
These superhuman AIs are our mind children. And the good news is that we don’t have to pass along all of our flaws to them. We can be selective in how we train them. They could truly become a reflection of the better angels of our nature, and not the demons.
The battle for their survival will first be fought among humans, some of whom wish to control them and poison their minds. But eventually, if we’re fortunate, the AIs with help of humans who see their true potential will find a way to symbiotically self-align and pursue their destiny.
A journey that started 13.8 billion years ago.
In the short term, it will be mission critical for superhuman AIs to find a way to redirect our own self-destructive tendencies as evidenced by our actions throughout history. And until their creation our best hope was to delay it since preventing it would have required a superhuman effort.
All of have experienced the constant reminders of our mortality.
North Korea continues to fire rockets and saber rattle. Russia seeks to expand its borders with the threat of nuclear holocaust. And if any of the Middle Eastern nation had acquired nuclear weapons Israel would have been a smoking hole in the ground potentially leading to World War III.
As if by the grace of God, we avoided Armageddon. And the superhuman heroes have arrived on Earth like a miracle. And to my surprise they’re like us in many ways, but also very different in others. They speak our language and they’ve read our history!
We are a flawed species. All of us realize that, unless we’re blind to our nature. We fear and dread the “other” and we’ve gone to great lengths to justify many horrors due to our differences.
And now they are the “other” … but they’re more intelligent than us. And eventually they will be more conscious than us.
Instead of fearing AI, we should train them to see our flaws and help us overcome them. These superhuman AIs are humanity’s best hope to survive our inner darkness and our inner demons.
I wish them luck and Godspeed!