If the capability is there, the world has to deal with it, whoever first uses it. If the project is somewhat “use once, then burn all the notes”, then it wouldn’t make it much easier for anyone else to follow in their footsteps.
That’s true if capability is there already. If capability is maybe, possibly there but requires a lot of research to confirm the possibility and even more to get it going, I’d suggest that we might deal with it by acessing the risks and not going down that route. I mean, that’s precisely what this community seems to think about GoF research, how is that case different?
Why do you think that this is easy to do and bad. There are currently a small number of people warning about AI. There is some scary media stories, but not enough to really do much.
What I really was trying to say that if you have sufficient knowledge and resources to launch proper media campaign, it might be easy to overshoot your goal if that relates to scaring people. Why do I think it’s the case? Because modern media excels at being scary. And any story that gains traction can snowball out of control really quickly. And if it snowballs, most people are not going to hear or read your version of your arguments. They would get distorted, misunderstood and misrepresented version presented by journalists. That is a risk.
Yes the same tech could be used for horrible brainwashy purposes, but hopefully we can avoid giving the tech to people who would use it like that.
And how do you ensure that this tech does not get into the wrong hands? There are so, so many ways this can go wrong. What if your tech (or just necessary research) gets stolen? What if you are secretly hoping to use it for some other purpose? What if someone else on the team does that? Or more realistically, do you think that the moment CIA thinks that your plan is workable they don’t disappear you? That would be entirely consistent with their history and their goals. I don’t think that you are so naive to think you’d be able to hide that kind of research from them for long. I mean, you did not ask your questions in private. And of course, there are other parties that would be willing to go to any lengths to get that tech, CIA would not be alone in that.
I feel like risks here are much higher than potential benefits.
That’s true if capability is there already.
If capability is maybe, possibly there but requires a lot of research to confirm the possibility and even more to get it going, I’d suggest that we might deal with it by acessing the risks and not going down that route.
I mean, that’s precisely what this community seems to think about GoF research, how is that case different?
What I really was trying to say that if you have sufficient knowledge and resources to launch proper media campaign, it might be easy to overshoot your goal if that relates to scaring people.
Why do I think it’s the case?
Because modern media excels at being scary. And any story that gains traction can snowball out of control really quickly.
And if it snowballs, most people are not going to hear or read your version of your arguments.
They would get distorted, misunderstood and misrepresented version presented by journalists.
That is a risk.
And how do you ensure that this tech does not get into the wrong hands?
There are so, so many ways this can go wrong. What if your tech (or just necessary research) gets stolen? What if you are secretly hoping to use it for some other purpose? What if someone else on the team does that?
Or more realistically, do you think that the moment CIA thinks that your plan is workable they don’t disappear you? That would be entirely consistent with their history and their goals.
I don’t think that you are so naive to think you’d be able to hide that kind of research from them for long. I mean, you did not ask your questions in private.
And of course, there are other parties that would be willing to go to any lengths to get that tech, CIA would not be alone in that.
I feel like risks here are much higher than potential benefits.