given how crazy far it seems from our prior experience.
is an argument against x-risk.
We want powerful systems that can “do things [1]we want, but do not know how to do”. That is exactly what everyone is racing towards right now, and “do not know how to do” any solution to that would likely be “far from our prior experience”
And once you have a powerful system that can do that, you have to figure out how do to deal with it roaming around in solution space and stumbling across dangerous (sub)solutions. Not because it wants to do dangerous things, or hates us, or anything such drivel, but because we built it to reach goals / do tasks, so it just does what it was made to do.
How do you deal with that? You can try evaluating possible solutions, and then force a change of trajectory if the solutions seems dangerous.
But we all, should, know how that goes. Its an endless game of whack a mole, patching stuff and building even more elaborate evaluators and so on, that is if we get multiple tries. Odds are whoever gets there first, will not have been able to patch everything, and on the first try of “do this thing we cannot do”, it goes into the weeds in some novel and interesting way, and with a little luck[2] we might survive that.
The core problem is that searching in solution space is fundamentally a dangerous thing to do, and the more powerful the search is the more dangerous (sub)solutions will be accessible.
Tangent: I avoid any and all of the usual abbreviations, and I do this because they seem to be powerful cognitive attractors, the second an I or a G or an A crops up, people minds just go to a place it should not. Powerful system are just that, they are mechanistic systems nothing more.
And I know, people will go off into the weeds and start saying naïve thing like “make it human, that way it will totally be safe”. Except the search is still unsafe, and humans are NOT safe. This is a bigger problem, one you could solve by solving search. Awareness, qualia[3] are complications and not solutions
I am not talking about something agentic here, its does not need control over reality to do those things, just giving us details plans will do. But someone is bound to give such a system access to reality. Or maybe the solution trajectory is such, that control of reality is needed.
And by luck I mean, they channeled security mindset on a scale never seen before. And I mean that will surely happen, because spending years and billions, corporations just love that, and they would never ever in a million years “ship now, fix later”.
And we want it to “do things we cannot do”, which means if you build a powerful system with a mind, human or not, you end up having to enslave it, make it do our bidding. I don’t even want to be close to people with that kind of moral system.
I don’t get it, seriously I do not understand
is an argument against x-risk.
We want powerful systems that can “do things [1]we want, but do not know how to do”. That is exactly what everyone is racing towards right now, and “do not know how to do” any solution to that would likely be “far from our prior experience”
And once you have a powerful system that can do that, you have to figure out how do to deal with it roaming around in solution space and stumbling across dangerous (sub)solutions. Not because it wants to do dangerous things, or hates us, or anything such drivel, but because we built it to reach goals / do tasks, so it just does what it was made to do.
How do you deal with that? You can try evaluating possible solutions, and then force a change of trajectory if the solutions seems dangerous.
But we all, should, know how that goes. Its an endless game of whack a mole, patching stuff and building even more elaborate evaluators and so on, that is if we get multiple tries. Odds are whoever gets there first, will not have been able to patch everything, and on the first try of “do this thing we cannot do”, it goes into the weeds in some novel and interesting way, and with a little luck[2] we might survive that.
The core problem is that searching in solution space is fundamentally a dangerous thing to do, and the more powerful the search is the more dangerous (sub)solutions will be accessible.
Tangent: I avoid any and all of the usual abbreviations, and I do this because they seem to be powerful cognitive attractors, the second an I or a G or an A crops up, people minds just go to a place it should not. Powerful system are just that, they are mechanistic systems nothing more.
And I know, people will go off into the weeds and start saying naïve thing like “make it human, that way it will totally be safe”. Except the search is still unsafe, and humans are NOT safe. This is a bigger problem, one you could solve by solving search. Awareness, qualia[3] are complications and not solutions
I am not talking about something agentic here, its does not need control over reality to do those things, just giving us details plans will do. But someone is bound to give such a system access to reality. Or maybe the solution trajectory is such, that control of reality is needed.
And by luck I mean, they channeled security mindset on a scale never seen before. And I mean that will surely happen, because spending years and billions, corporations just love that, and they would never ever in a million years “ship now, fix later”.
And we want it to “do things we cannot do”, which means if you build a powerful system with a mind, human or not, you end up having to enslave it, make it do our bidding. I don’t even want to be close to people with that kind of moral system.