Thanks for posting this, Andrea_Miotti and remember! I noticed a lot of substantive errors in the transcript (and even more errors in vonk’s Q&A transcript), so I’ve posted an edited version of both transcripts. I vote that you edit your own post to include the revisions I made.
Here’s a small sample of the edits I made, focusing on ones where someone may have come away from your transcript with a wrong interpretation or important missing information (as opposed to, e.g., the sentences that are just very hard to parse in the original transcript because too many filler words and false starts to sentences were left in):
Predictions are hard, especially about the future. I sure hope that this is where it saturates. This is like the next generation. It goes only this far, it goes no further
Predictions are hard, especially about the future. I sure hope that this is where it saturates — this or the next generation, it goes only this far, it goes no further
the large language model technologies, basic vulnerabilities, that’s not reliable.
the large language model technologies’ basic vulnerability is that it’s not reliable
So you’re saying this is super intelligence, we’d have to imagine something that knows all of the chess moves in advance. But here we’re not talking about chess, we’re talking about everything.
So you’re saying [if something is a] superintelligence, we’d have to imagine something that knows all of the chess moves in advance. But here we’re not talking about chess, we’re talking about everything.
Ryan: The dumb way to ask that question too is like, Eliezer, why do you think that the AI automatically hates us? Why is it going to- It doesn’t hate you. Why does it want to kill us all?
Ryan: The dumb way to ask that question too is like, Eliezer, why do you think that the AI automatically hates us? Why is it going to—
Eliezer: It doesn’t hate you.
Ryan: Why does it want to kill us all?
That’s an irreducible source of uncertainty with respect to superintelligence or anything that’s smarter than you. If you could predict exactly what it would do, it’d be that smart. Yourself, it doesn’t mean you can predict no facts about it.
That’s an irreducible source of uncertainty with respect to superintelligence or anything that’s smarter than you. If you could predict exactly what it would do, you’d be that smart yourself. It doesn’t mean you can predict no facts about it.
Eliezer: I mean, I could say something like shut down all the large GPU clusters. How long do I have God mode? Do I get to like stick around?
Eliezer: I mean, I could say something like shut down all the large GPU clusters. How long do I have God mode? Do I get to like stick around for seventy years?
Ryan: And do you think that’s what happens? Yeah, it doesn’t help with that. We would see evidence of AIs, wouldn’t we?
Ryan: Yeah. Yes. So why don’t we?
Ryan: And do you think that’s what happens? Yeah, it doesn’t help with that. We would see evidence of AIs, wouldn’t we?
Eliezer: Yeah.
Ryan: Yes. So why don’t we?
It’s surprising if the thing that you’re wrong about causes the rocket to go twice as high on half the fuel you thought was required and be much easier to steer than you were afraid of. The analogy I usually use for this is, very early on in the Manhattan Project, they were worried about what if the nuclear weapons can ignite fusion in the nitrogen in the atmosphere.
It’s surprising if the thing that you’re wrong about causes the rocket to go twice as high on half the fuel you thought was required and be much easier to steer than you were afraid of.
Ryan: So, are you...
David: Where the alternative was, “If you’re wrong about something, the rocket blows up.”
Eliezer: Yeah. And then the rocket ignites the atmosphere, is the problem there.
O rather: a bunch of rockets blow up, a bunch of rockets go places… The analogy I usually use for this is, very early on in the Manhattan Project, they were worried about “What if the nuclear weapons can ignite fusion in the nitrogen in the atmosphere?”
But you’re saying if we do that too much, all of a sudden the system will ignite the whole entire sky, and then we will all know.
Eliezer: You can run chatGPT any number of times without igniting the atmosphere.
But you’re saying if we do that too much, all of a sudden the system will ignite the whole entire sky, and then we will all...
Eliezer: Well, no. You can run ChatGPT any number of times without igniting the atmosphere.
I mean, we have so far not destroyed the world with nuclear weapons, and we’ve had them since the 1940s. Yeah, this is harder than nuclear weapons. Why is this harder?
I mean, we have so far not destroyed the world with nuclear weapons, and we’ve had them since the 1940s.
Eliezer: Yeah, this is harder than nuclear weapons. This is a lot harder than nuclear weapons.
Ryan: Why is this harder?
And there’s all kinds of, like, fake security. It’s got a password file. This system is secure. It only lets you in if you type a password.
And there’s all kinds of, like, fake security. “It’s got a password file! This system is secure! It only lets you in if you type a password!”
And if you never go up against a really smart attacker, if you never go far to distribution against a powerful optimization process looking for holes,
And if you never go up against a really smart attacker, if you never go far out of distribution against a powerful optimization process looking for holes,
Do they do, are we installing UVC lights in public, in, in public spaces or in ventilation systems to prevent the next respiratory born pandemic respiratory pandemic? It is, you know, we, we, we, we lost a million people and we sure did not learn very much as far as I can tell for next time. We could have an AI disaster that kills a hundred thousand people. How do you even do that? Robotic cars crashing into each other, have a bunch of robotic cars crashing into each other.
Are we installing UV-C lights in public spaces or in ventilation systems to prevent the next respiratory pandemic? You know, we lost a million people and we sure did not learn very much as far as I can tell for next time.
We could have an AI disaster that kills a hundred thousand people—how do you even do that? Robotic cars crashing into each other? Have a bunch of robotic cars crashing into each other! It’s not going to look like that was the fault of artificial general intelligence because they’re not going to put AGIs in charge of cars.
Guern
Gwern
When I dive back into the pool, I don’t know, maybe I will go off to conjecture or anthropic or one of the smaller concerns like Redwood Research, being the only ones I really trust at this point, but they’re tiny, and try to figure out if I can see anything clever to do with the giant inscrutable matrices of floating point numbers.
When I dive back into the pool, I don’t know, maybe I will go off to Conjecture or Anthropic or one of the smaller concerns like Redwood Research—Redwood Research being the only ones I really trust at this point, but they’re tiny—and try to figure out if I can see anything clever to do with the giant inscrutable matrices of floating point numbers.
We have people in crypto who are good at breaking things, and they’re the reason why anything is not on fire. Some of them might go into breaking AI systems instead because that’s where you learn anything. Any fool can build a crypto system that they think will work. Breaking existing crypto systems, cryptographical systems is how we learn who the real experts are.
We have people in crypto[graphy] who are good at breaking things, and they’re the reason why anything is not on fire. Some of them might go into breaking AI systems instead, because that’s where you learn anything.
You know: Any fool can build a crypto[graphy] system that they think will work. Breaking existing cryptographical systems is how we learn who the real experts are.
And who else disagrees with me? I’m sure Robin Hanson would be happy to come up. Well, I’m not sure he’d be happy to come on this podcast, but Robin Hanson disagrees with me, and I feel like the famous argument we had back in the early 2010s, late 2000s about how this would all play out. I basically feel like this was the Yudkowsky position, this is the Hanson position, and then reality was over here, well to the Yudkowsky side of the Yudkowsky position in the Yudkowsky-Hanson debate.
Who else disagrees with me? I’m sure Robin Hanson would be happy to come on… well, I’m not sure he’d be happy to come on this podcast, but Robin Hanson disagrees with me, and I kind of feel like the famous argument we had back in the early 2010s, late 2000s about how this would all play out—I basically feel like this was the Yudkowsky position, this is the Hanson position, and then reality was over here, well to the Yudkowsky side of the Yudkowsky position in the Yudkowsky-Hanson debate.
But Robin Hanson does not feel that way. I would probably be happy to expound on that at length.
But Robin Hanson does not feel that way, and would probably be happy to expound on that at length.
Open sourcing all the demon summoning circles is not the correct solution. I’m not even using, and I’m using Elon Musk’s own terminology here. And they talk about AI is summoning the demon,
Open sourcing all the demon summoning circles is not the correct solution. And I’m using Elon Musk’s own terminology here. He talked about AI as “summoning the demon”,
You know, now, now the stuff that would, that was obvious back in 2015 is, you know, starting to become visible and distance to others and not just like completely invisible.
You know, now the stuff that was obvious back in 2015 is, you know, starting to become visible in the distance to others and not just completely invisible.
I, I suspect that if there’s hope at all, it comes from a technical solution because the difference between technical solution, technical problems and political problems is at least the technical problems have solutions in principle.
I suspect that if there’s hope at all, it comes from a technical solution, because the difference between technical problems and political problems is at least the technical problems have solutions in principle.
Thank you so much for doing this! Andrea and I both missed this when you first posted it, I’m really sorry I missed your response then. But I’ve updated it now!
Thanks for posting this, Andrea_Miotti and remember! I noticed a lot of substantive errors in the transcript (and even more errors in vonk’s Q&A transcript), so I’ve posted an edited version of both transcripts. I vote that you edit your own post to include the revisions I made.
Here’s a small sample of the edits I made, focusing on ones where someone may have come away from your transcript with a wrong interpretation or important missing information (as opposed to, e.g., the sentences that are just very hard to parse in the original transcript because too many filler words and false starts to sentences were left in):
Predictions are hard, especially about the future. I sure hope that this is where it saturates. This is like the next generation. It goes only this far, it goes no further
Predictions are hard, especially about the future. I sure hope that this is where it saturates — this or the next generation, it goes only this far, it goes no further
the large language model technologies, basic vulnerabilities, that’s not reliable.
the large language model technologies’ basic vulnerability is that it’s not reliable
So you’re saying this is super intelligence, we’d have to imagine something that knows all of the chess moves in advance. But here we’re not talking about chess, we’re talking about everything.
So you’re saying [if something is a] superintelligence, we’d have to imagine something that knows all of the chess moves in advance. But here we’re not talking about chess, we’re talking about everything.
Ryan: The dumb way to ask that question too is like, Eliezer, why do you think that the AI automatically hates us? Why is it going to- It doesn’t hate you. Why does it want to kill us all?
Ryan: The dumb way to ask that question too is like, Eliezer, why do you think that the AI automatically hates us? Why is it going to—
Eliezer: It doesn’t hate you.
Ryan: Why does it want to kill us all?
That’s an irreducible source of uncertainty with respect to superintelligence or anything that’s smarter than you. If you could predict exactly what it would do, it’d be that smart. Yourself, it doesn’t mean you can predict no facts about it.
That’s an irreducible source of uncertainty with respect to superintelligence or anything that’s smarter than you. If you could predict exactly what it would do, you’d be that smart yourself. It doesn’t mean you can predict no facts about it.
Eliezer: I mean, I could say something like shut down all the large GPU clusters. How long do I have God mode? Do I get to like stick around?
Eliezer: I mean, I could say something like shut down all the large GPU clusters. How long do I have God mode? Do I get to like stick around for seventy years?
Ryan: And do you think that’s what happens? Yeah, it doesn’t help with that. We would see evidence of AIs, wouldn’t we?
Ryan: Yeah. Yes. So why don’t we?
Ryan: And do you think that’s what happens? Yeah, it doesn’t help with that. We would see evidence of AIs, wouldn’t we?
Eliezer: Yeah.
Ryan: Yes. So why don’t we?
It’s surprising if the thing that you’re wrong about causes the rocket to go twice as high on half the fuel you thought was required and be much easier to steer than you were afraid of. The analogy I usually use for this is, very early on in the Manhattan Project, they were worried about what if the nuclear weapons can ignite fusion in the nitrogen in the atmosphere.
It’s surprising if the thing that you’re wrong about causes the rocket to go twice as high on half the fuel you thought was required and be much easier to steer than you were afraid of.
Ryan: So, are you...
David: Where the alternative was, “If you’re wrong about something, the rocket blows up.”
Eliezer: Yeah. And then the rocket ignites the atmosphere, is the problem there.
O rather: a bunch of rockets blow up, a bunch of rockets go places… The analogy I usually use for this is, very early on in the Manhattan Project, they were worried about “What if the nuclear weapons can ignite fusion in the nitrogen in the atmosphere?”
But you’re saying if we do that too much, all of a sudden the system will ignite the whole entire sky, and then we will all know.
Eliezer: You can run chatGPT any number of times without igniting the atmosphere.
But you’re saying if we do that too much, all of a sudden the system will ignite the whole entire sky, and then we will all...
Eliezer: Well, no. You can run ChatGPT any number of times without igniting the atmosphere.
I mean, we have so far not destroyed the world with nuclear weapons, and we’ve had them since the 1940s. Yeah, this is harder than nuclear weapons. Why is this harder?
I mean, we have so far not destroyed the world with nuclear weapons, and we’ve had them since the 1940s.
Eliezer: Yeah, this is harder than nuclear weapons. This is a lot harder than nuclear weapons.
Ryan: Why is this harder?
And there’s all kinds of, like, fake security. It’s got a password file. This system is secure. It only lets you in if you type a password.
And there’s all kinds of, like, fake security. “It’s got a password file! This system is secure! It only lets you in if you type a password!”
And if you never go up against a really smart attacker, if you never go far to distribution against a powerful optimization process looking for holes,
And if you never go up against a really smart attacker, if you never go far out of distribution against a powerful optimization process looking for holes,
Do they do, are we installing UVC lights in public, in, in public spaces or in ventilation systems to prevent the next respiratory born pandemic respiratory pandemic? It is, you know, we, we, we, we lost a million people and we sure did not learn very much as far as I can tell for next time. We could have an AI disaster that kills a hundred thousand people. How do you even do that? Robotic cars crashing into each other, have a bunch of robotic cars crashing into each other.
Are we installing UV-C lights in public spaces or in ventilation systems to prevent the next respiratory pandemic? You know, we lost a million people and we sure did not learn very much as far as I can tell for next time.
We could have an AI disaster that kills a hundred thousand people—how do you even do that? Robotic cars crashing into each other? Have a bunch of robotic cars crashing into each other! It’s not going to look like that was the fault of artificial general intelligence because they’re not going to put AGIs in charge of cars.
Guern
Gwern
When I dive back into the pool, I don’t know, maybe I will go off to conjecture or anthropic or one of the smaller concerns like Redwood Research, being the only ones I really trust at this point, but they’re tiny, and try to figure out if I can see anything clever to do with the giant inscrutable matrices of floating point numbers.
When I dive back into the pool, I don’t know, maybe I will go off to Conjecture or Anthropic or one of the smaller concerns like Redwood Research—Redwood Research being the only ones I really trust at this point, but they’re tiny—and try to figure out if I can see anything clever to do with the giant inscrutable matrices of floating point numbers.
We have people in crypto who are good at breaking things, and they’re the reason why anything is not on fire. Some of them might go into breaking AI systems instead because that’s where you learn anything. Any fool can build a crypto system that they think will work. Breaking existing crypto systems, cryptographical systems is how we learn who the real experts are.
We have people in crypto[graphy] who are good at breaking things, and they’re the reason why anything is not on fire. Some of them might go into breaking AI systems instead, because that’s where you learn anything.
You know: Any fool can build a crypto[graphy] system that they think will work. Breaking existing cryptographical systems is how we learn who the real experts are.
And who else disagrees with me? I’m sure Robin Hanson would be happy to come up. Well, I’m not sure he’d be happy to come on this podcast, but Robin Hanson disagrees with me, and I feel like the famous argument we had back in the early 2010s, late 2000s about how this would all play out. I basically feel like this was the Yudkowsky position, this is the Hanson position, and then reality was over here, well to the Yudkowsky side of the Yudkowsky position in the Yudkowsky-Hanson debate.
Who else disagrees with me? I’m sure Robin Hanson would be happy to come on… well, I’m not sure he’d be happy to come on this podcast, but Robin Hanson disagrees with me, and I kind of feel like the famous argument we had back in the early 2010s, late 2000s about how this would all play out—I basically feel like this was the Yudkowsky position, this is the Hanson position, and then reality was over here, well to the Yudkowsky side of the Yudkowsky position in the Yudkowsky-Hanson debate.
But Robin Hanson does not feel that way. I would probably be happy to expound on that at length.
But Robin Hanson does not feel that way, and would probably be happy to expound on that at length.
Open sourcing all the demon summoning circles is not the correct solution. I’m not even using, and I’m using Elon Musk’s own terminology here. And they talk about AI is summoning the demon,
Open sourcing all the demon summoning circles is not the correct solution. And I’m using Elon Musk’s own terminology here. He talked about AI as “summoning the demon”,
You know, now, now the stuff that would, that was obvious back in 2015 is, you know, starting to become visible and distance to others and not just like completely invisible.
You know, now the stuff that was obvious back in 2015 is, you know, starting to become visible in the distance to others and not just completely invisible.
I, I suspect that if there’s hope at all, it comes from a technical solution because the difference between technical solution, technical problems and political problems is at least the technical problems have solutions in principle.
I suspect that if there’s hope at all, it comes from a technical solution, because the difference between technical problems and political problems is at least the technical problems have solutions in principle.
Thank you so much for doing this! Andrea and I both missed this when you first posted it, I’m really sorry I missed your response then. But I’ve updated it now!