The claim is not that all AI alignment has bad epistemics—I haven’t read all of it, and I know some at least make empirical hypotheses, and like and enjoy the parts which do—the claim is that a lot do. Including, alarmingly, the genealogically first, most influential, and most doomy.
The central example is of course MIRI, which afaict maintains that there is a > 99% chance of doom because of deep, hard-to-communicate insight into the nature of intelligence, agency, and optimization, while refraining from making any non-standard predictions about intelligence in the world prior to doom. If your causal model based on understanding intelligence gives a tight 99% probability of doom, then it should also give you some tight predictions before then, and this is an absolutely immense red flag if it does not. Point me to some predictions if I’m wrong.
(It’s an immense red flag even if Yann LeCunn or some non-doomers also have sucky epistemics, make non falsifiable claims, etc.)
I agree with Jacob that the early broad correct claims about the future seem non-specific to MIRI / early AI alignment concerns—I think it probably makes more sense to defer to Kurzweil than Yud off those (haven’t read K a ton). And many early MIRI-specific claims seem just false—I just reread IEM and it heavily thinks, for instance, that clever algorithms rather than compute is the path to intelligence, which looks pretty wrong from the current view.
I’m happy to criticize Yudkowsky, I think there’s some truth to various of your criticisms in fact. I do think you are being way too uncharitable to him though. For example, ‘clever algorithms’ are in fact responsible for approximately half of the progress towards AGI in recent decades, IIRC, and I expect them to continue to play a role, and I don’t think IEM was making any claims along the lines of ‘compute is irrelevant’ or ‘compute is fairly unimportant.’
But in general, on average, the epistemics of the doomers are better than the epistemics of the non-doomers. I claim. And moreover the epistemics of Yudkowsky are better than the epistemics of most non-doomers.
I think you are confused about how to use falsifiability. Here, perhaps this meme is a succinct way explain what I mean:
So if someone prophesied doom by asteroid, we’d be unsatisfied with someone saying “there will be a giant explosion from an asteroid.” We’d probably want them to be able to predict locations in the sky where you could see the asteroid, before we shifted society to live in underground bunkers to protect us from the blast. Why do we expect them to be able to predict the doom, but no non-doom? Why would their knowledge extend to one thing, but not the other?
Similarly, if someone prophesied doom from a supernatural occurrence, we’d also want to see some miracles, proofs, signs—or at least other supernatural predictions—before we covered ourselves in sackcloth and ashes and started flagellating ourselves to protect ourselves from God’s wrath. Why would we expect them to be able to predicted doom, but nothing else? Not adhering to this standard historically would result in you selling all your possessions quite unreasonably.
I could go on for quite some time. I think this is a super reasonable standard. I think not adhering to it, historically, makes you fall for things that are straightforward cults. It’s basically a prior against theories of doom getting to by coincidence not make any predictions prior to doom. Yeah, sure.
So, yeah, I’ve seen the above meme a lot on Twitter—I think it’s evidence that the epistemics of many doomers are just bad, inasmuch as it is shared and liked and retweeted. You can construe it either as (1) mocking people for not seeing something which is just obvious, or its (2) just asserting that you don’t need to make any other predictions, doom is plenty. (2) is just a bad epistemic standard, as history suggests.
And mocking people is, well, for sure a move you can do.
I analogized AI doom to saying-the-fire-will-kill-the-dog; you analogize AI doom to saying-supernatural-shit-will-happen. Which analogy is more apt? That’s the crux.
OK, let’s zoom in on the fire analogy from the meme. The point I was making is that the doomers have been making plenty of falsifiable predictions which have in fact been coming true—e.g. that the fire will grow and grow, that the room will get hotter and brighter, and eventually the dog will burn to death. Falsifiability is relative. The fire-doomers have been making plenty of falsifiable predictions relative to interlocutors who were saying that the fire would disappear or diminish, but not making any falsifiable predictions relative to interlocutors who were saying that the fire would burn brighter and hotter but then somehow fail to kill the dog for some reason or other.
Similarly, the AGI doomers have been making plenty of falsifiable predictions relative to most people, e.g. people who didn’t think AGI was possible in our lifetimes, or people who thought it would be tool-like rather than agent-like, or modular rather than unified. But their predictions have been unfalsifiable relative to certain non-doomers who also predicted unified agentic AGI that for one reason or another wouldn’t cause doom.
OK. Next point: Falsifiability is symmetric. If my predictions are unfalsifiable relative to yours, yours are also unfalsifiable relative to me. Fundamentally what’s going on is simply that there are two groups of people whose expectations/predictions are the same (at least for now). This sort of thing happens all the time, and in fact is inescapable due to underdetermination of theory by data.
Next point: We have more resources to draw from besides falsifiability. We can evaluate arguments, we can build models and examine their assumptions, we can compare the simplicity and elegance of different models… All the usual tools of science for adjudicating disputes between different views that are both consistent with what we’ve seen so far.
Final point: AGI doomers have been unusually good about trying to stick their necks out and find ways that their views differ from the views of others. AGI doomers have been making models and arguments, and putting them in writing, and letting others attack premises and cast doubt on assumptions and poke holes in logic and cheer when a prediction turns out false.… more so than AGI non-doomers. Imagine what it would look like if the shoe was on the other foot: there’d be an “AGI non-doom” camp that would be making big multi-premise arguments about why AGI will probably go well by default, and making various predictions, and then a dispersed, disunified gaggle of doomer skeptics poking holes in the argument and saying “we don’t have enough evidence for such a crazy view” and “show me a falsifiable claim” and cheering when some prediction turns out false.
Putting it together… IMO you are using “y’all aren’t making falsifiable predictions” as a rhetorical cudgel with which to beat your opponents, when actually they would be equally justified in using it against you. It’s a symmetric weapon, it isn’t actually a component of good epistemology (at least not when used the way you are using it.)
(0) What I hint at in (1-3) seems like a very standard thing, for which there should be a standard reference post. Does anybody know it? If not, somebody should write on it.
(1) Pascal’s wagers in general: Am I interperting you correctly as saying that AI risk, doom by asteroid, belief in Christianity, doom by supernatural occurence, etc., all belong to the same class of “Pascal’s wager-like” decision problems?
If yes, then hurray, I definitely agree. However, I believe this is where you should bite the bullet and treat every Pascal’s wager seriously, and only dismiss it once you understand it somewhat. In particular, I claim that the heuristic “somebody claims X has extremely high or low payoffs ==> dismiss X until we have a rock-solid proof” is not valid. (Curious whether you disagree.) ((And to preempt the likely objection: To me, this position is perfectly compatible with initially dismissing crackpots and God’s wrath without engaging with even listening to the arguments; see below.))
Why do I believe this?
First, I don’t think we encounter so many different Pascal wagers that considering each one at least briefly will paralyze us. This is made easier because many of them are somewhat similar to each other (eg “religions”, “conspiracy theories”, “crackpot designs for perpetum mobile”, …), which allows you to only thing through a single representative of each class[1], and then use the heuristics like “if it pattern-matches on a conspiracy theory, tentatively ignore it outright”. (Keyword “tentatively”. It seems fair to initially dismiss covid lab-leak hypothesis as a conspiracy. But if you later hear that there was a bio-lab in Wuhan, you should reevaluate. Similarly, I dismiss all religions outright, but I would spend some attention on it if many people I trust suddenly started saying that they saw miracles.)
Second, many Pascal wagers are non-actionable. If you learn about meteors or super-vulcanos in 18th century, there is not much you can do. Similarly, I assign a non-zero probability to the simulation probability. And perhaps even a non-zero probability of somebody running a simulation of a world that actually works on christianity, just for the lolz of it.[2] But it is equally possible that this guy runs it on anti-christianity (saints going to hell), just for the lolz of it. And I don’t know which is more probable, and I have no way of resolving this, and the payoffs are not literally infinite (that guy would have a finite compute budget). So there is no point in me acting on this either.
Third, let me comment a bit on the meteor example [I don’t know the historical details, so I will be fabulating somewhat]: Sometime in the previous century, we learned about an asteroid hitting earth in the past and killing the dinosaurs. This seems like “there might be a giant explosion from an asteroid”. We did not force everybody to live underground, not because we didn’t take the risk seriously, but because doing so would have been stupid (we probably didn’t have an asteroid-surviving technology anyway, and the best way to get it would have been to stay on the surface). However, we did spend lots of money on the asteroid detection program to understand the risk better. As a result, we now understand this particular “Pascal’s wager” better, and it seems that in this particular case, the risk is small. That is, the risk is small enough that it is more important to focus on other things. However, I believe that we should still spend some effort on preventing asteroid risk (actually, I guess we do do that). And if we had no more-pressing issues (X-risk and otherwise), I totally would invest into some resources asteroid-explosion-survival measures.
To summarize the meteor example: We did not dismiss the risk out of hand.We investigated it seriously, and then we only dismissed it once we understood it. And it never made sense to apply extreme measures, because they wouldn’t have been effective even if the risk was real. (Though I am not sure whether that was on purpose or because we wouldn’t been able to apply extreme measures even if they were appropriate.) ((Uhm, please please please, I hope nobody starts making claims about low-level implications of the asteroid risk-analysis to AI risk-analysis. The low-level details of the two cases have nothing in common, so there are no implications to draw.))
(2) For “stop AI progress” to be rational, it is enough to be uncertain.
Sure, if you claim something like “90% that AI is going to kill us” or “we are totally sure that AI is risky”, then the burden of proof is on you.
But most people who advocate treating AI as risky, and that we should stop/slow/take extreme measures, do not claim anything like “90% that AI is going to kill us”. Most often, you see claims like (i) “at least 5% that AI could kill us”, or (ii) “we really don’t know whether AI might kill us, one way or the other”.
And if you claim non-(i) or non-(ii), the burden of proof totally is on you. And I think that literally nobody in the world has a good enough model of AI & the strategic landscape to justify non-(i) or non-(ii). (Definitely nobody has published a gears-level model like that.)
Also, I think that to make Pascal’s wager on AI rational, it is enough to believe (either (i) or (ii)) together with “we have no quick way of resolving the uncertainty”.
(3) AI risk as Pascal’s wager:
I claim that it is rational to treat AI as a Pascal’s wager, but that in this particular case, the details come out very differently (primarily because we understand AI so little).
To give an illustration (there will be many nonsense bits in the details; the goal is to give the general vibe):
If you know literally nothing about AI at all, it is fair to initially pattern-match it to “yeah yeah, random people talking about the end of the world, that’s bullshit”.
However, once you learn that AI stands for “artificial intelligence” you really should do a double-take: “Hey, intelligence is this thing that maybe makes us, and not other species, dominant on this planet. Also, humans are intelligent and they do horrible things to each other. So it is at least plausible that this could destroy us.” (That’s without knowing anything at all about how AI should work.)
At that point, you should consider whether this AI thing might be achievable. And if it is, say, year 1900, then you should go “I have literally no clue, so I should act as if the chance is non-negligible”. And then (much) later people invent computers and you hear about the substrate-independence thesis. You don’t understand any of this, but it sounds plausible. And then you should treat AI as a real possibility, until you can get a gear’s-level understanding of the details. (Which we still don’t have!)
Also, you try understand whether AI will be dangerous or not, and you will mostly come to the conclusion that you have no clue whatsoever. (And we still don’t even now. And on the current timeline, we probably won’t know until we get superintelligence.)
And from that point onward, you should be like “holy ****, this AI thing could be really dangerous! We should never build any thing like that unless we understand exactly what we are doing!”.
This does not mean you will not build a whole bunch of AI. You will build calculators, and you will build GOFAI, and tabular model-based RL.[3] And then somebody comes up with the idea of deep learning, and you are like “holy ****, we have no clue how this works!”. And you put a moratorium on it, to quickly figure out what to do. And later you probably you end up controlling it pretty tightly, because you don’t know how it scales. And you only do the research on DL slowly, such that your interpretability techniques and other theory never gets behind.
Obviously, we don’t live in a world capable of something like above. But it remains a fact that we now have strong evidence that (1) AGI probably is possible and (2) we still have no clue whether it is going to be harmful of not, despite trying. And if we could do it, the rational thing to do would be to immediatelly stop most AI work until we figure out (2).
When I was younger, I did think a lot about christianity and I have a detailed model that causes me to assign literally 0 probability to it being correct. Similarly, I did engage with some conspiracy theories before dismissing them. I believe that justifies me to tentatively immediatelly dismiss anything that pattern matches to religion and conspiracy theories. However, if I never heard about any religion, then upon encountering one for the first time, I would totally not be justified to dismiss it outright. And let me bite the bullet here: I would say that if you were a random guy in ancient Egypt, without particularly good models of the world or ways to improve them, the rational thing to do could have been to act as if the Egyptian gods were real. [low confidence though, I haven’t thought about this before]
Not to be confused with the hypothesis that the universe actually works according to any of of the religions considered on earth without assuming simulation hypothesis. To that, I assign literally 0 probability.
But for all of those things, you briefly consider whether those things could be dangerous and conclude they aren’t. (Separately from AI risks: In an ideal world, you would also make some experiments to see how these things are affecting the users. EG, to see whether kids’ brains detoriate by relying on calculators. In practice, our science methods sucked back then, so you don’t learn anything. But hey, talking about how a better world would work.)
The claim is not that all AI alignment has bad epistemics—I haven’t read all of it, and I know some at least make empirical hypotheses, and like and enjoy the parts which do—the claim is that a lot do. Including, alarmingly, the genealogically first, most influential, and most doomy.
The central example is of course MIRI, which afaict maintains that there is a > 99% chance of doom because of deep, hard-to-communicate insight into the nature of intelligence, agency, and optimization, while refraining from making any non-standard predictions about intelligence in the world prior to doom. If your causal model based on understanding intelligence gives a tight 99% probability of doom, then it should also give you some tight predictions before then, and this is an absolutely immense red flag if it does not. Point me to some predictions if I’m wrong.
(It’s an immense red flag even if Yann LeCunn or some non-doomers also have sucky epistemics, make non falsifiable claims, etc.)
I agree with Jacob that the early broad correct claims about the future seem non-specific to MIRI / early AI alignment concerns—I think it probably makes more sense to defer to Kurzweil than Yud off those (haven’t read K a ton). And many early MIRI-specific claims seem just false—I just reread IEM and it heavily thinks, for instance, that clever algorithms rather than compute is the path to intelligence, which looks pretty wrong from the current view.
I’m happy to criticize Yudkowsky, I think there’s some truth to various of your criticisms in fact. I do think you are being way too uncharitable to him though. For example, ‘clever algorithms’ are in fact responsible for approximately half of the progress towards AGI in recent decades, IIRC, and I expect them to continue to play a role, and I don’t think IEM was making any claims along the lines of ‘compute is irrelevant’ or ‘compute is fairly unimportant.’
But in general, on average, the epistemics of the doomers are better than the epistemics of the non-doomers. I claim. And moreover the epistemics of Yudkowsky are better than the epistemics of most non-doomers.
I think you are confused about how to use falsifiability. Here, perhaps this meme is a succinct way explain what I mean:
So if someone prophesied doom by asteroid, we’d be unsatisfied with someone saying “there will be a giant explosion from an asteroid.” We’d probably want them to be able to predict locations in the sky where you could see the asteroid, before we shifted society to live in underground bunkers to protect us from the blast. Why do we expect them to be able to predict the doom, but no non-doom? Why would their knowledge extend to one thing, but not the other?
Similarly, if someone prophesied doom from a supernatural occurrence, we’d also want to see some miracles, proofs, signs—or at least other supernatural predictions—before we covered ourselves in sackcloth and ashes and started flagellating ourselves to protect ourselves from God’s wrath. Why would we expect them to be able to predicted doom, but nothing else? Not adhering to this standard historically would result in you selling all your possessions quite unreasonably.
I could go on for quite some time. I think this is a super reasonable standard. I think not adhering to it, historically, makes you fall for things that are straightforward cults. It’s basically a prior against theories of doom getting to by coincidence not make any predictions prior to doom. Yeah, sure.
So, yeah, I’ve seen the above meme a lot on Twitter—I think it’s evidence that the epistemics of many doomers are just bad, inasmuch as it is shared and liked and retweeted. You can construe it either as (1) mocking people for not seeing something which is just obvious, or its (2) just asserting that you don’t need to make any other predictions, doom is plenty. (2) is just a bad epistemic standard, as history suggests.
And mocking people is, well, for sure a move you can do.
I analogized AI doom to saying-the-fire-will-kill-the-dog; you analogize AI doom to saying-supernatural-shit-will-happen. Which analogy is more apt? That’s the crux.
OK, let’s zoom in on the fire analogy from the meme. The point I was making is that the doomers have been making plenty of falsifiable predictions which have in fact been coming true—e.g. that the fire will grow and grow, that the room will get hotter and brighter, and eventually the dog will burn to death. Falsifiability is relative. The fire-doomers have been making plenty of falsifiable predictions relative to interlocutors who were saying that the fire would disappear or diminish, but not making any falsifiable predictions relative to interlocutors who were saying that the fire would burn brighter and hotter but then somehow fail to kill the dog for some reason or other.
Similarly, the AGI doomers have been making plenty of falsifiable predictions relative to most people, e.g. people who didn’t think AGI was possible in our lifetimes, or people who thought it would be tool-like rather than agent-like, or modular rather than unified. But their predictions have been unfalsifiable relative to certain non-doomers who also predicted unified agentic AGI that for one reason or another wouldn’t cause doom.
OK. Next point: Falsifiability is symmetric. If my predictions are unfalsifiable relative to yours, yours are also unfalsifiable relative to me. Fundamentally what’s going on is simply that there are two groups of people whose expectations/predictions are the same (at least for now). This sort of thing happens all the time, and in fact is inescapable due to underdetermination of theory by data.
Next point: We have more resources to draw from besides falsifiability. We can evaluate arguments, we can build models and examine their assumptions, we can compare the simplicity and elegance of different models… All the usual tools of science for adjudicating disputes between different views that are both consistent with what we’ve seen so far.
Final point: AGI doomers have been unusually good about trying to stick their necks out and find ways that their views differ from the views of others. AGI doomers have been making models and arguments, and putting them in writing, and letting others attack premises and cast doubt on assumptions and poke holes in logic and cheer when a prediction turns out false.… more so than AGI non-doomers. Imagine what it would look like if the shoe was on the other foot: there’d be an “AGI non-doom” camp that would be making big multi-premise arguments about why AGI will probably go well by default, and making various predictions, and then a dispersed, disunified gaggle of doomer skeptics poking holes in the argument and saying “we don’t have enough evidence for such a crazy view” and “show me a falsifiable claim” and cheering when some prediction turns out false.
Putting it together… IMO you are using “y’all aren’t making falsifiable predictions” as a rhetorical cudgel with which to beat your opponents, when actually they would be equally justified in using it against you. It’s a symmetric weapon, it isn’t actually a component of good epistemology (at least not when used the way you are using it.)
(0) What I hint at in (1-3) seems like a very standard thing, for which there should be a standard reference post. Does anybody know it? If not, somebody should write on it.
(1) Pascal’s wagers in general: Am I interperting you correctly as saying that AI risk, doom by asteroid, belief in Christianity, doom by supernatural occurence, etc., all belong to the same class of “Pascal’s wager-like” decision problems?
If yes, then hurray, I definitely agree. However, I believe this is where you should bite the bullet and treat every Pascal’s wager seriously, and only dismiss it once you understand it somewhat. In particular, I claim that the heuristic “somebody claims X has extremely high or low payoffs ==> dismiss X until we have a rock-solid proof” is not valid. (Curious whether you disagree.) ((And to preempt the likely objection: To me, this position is perfectly compatible with initially dismissing crackpots and God’s wrath without engaging with even listening to the arguments; see below.))
Why do I believe this?
First, I don’t think we encounter so many different Pascal wagers that considering each one at least briefly will paralyze us. This is made easier because many of them are somewhat similar to each other (eg “religions”, “conspiracy theories”, “crackpot designs for perpetum mobile”, …), which allows you to only thing through a single representative of each class[1], and then use the heuristics like “if it pattern-matches on a conspiracy theory, tentatively ignore it outright”. (Keyword “tentatively”. It seems fair to initially dismiss covid lab-leak hypothesis as a conspiracy. But if you later hear that there was a bio-lab in Wuhan, you should reevaluate. Similarly, I dismiss all religions outright, but I would spend some attention on it if many people I trust suddenly started saying that they saw miracles.)
Second, many Pascal wagers are non-actionable. If you learn about meteors or super-vulcanos in 18th century, there is not much you can do. Similarly, I assign a non-zero probability to the simulation probability. And perhaps even a non-zero probability of somebody running a simulation of a world that actually works on christianity, just for the lolz of it.[2] But it is equally possible that this guy runs it on anti-christianity (saints going to hell), just for the lolz of it. And I don’t know which is more probable, and I have no way of resolving this, and the payoffs are not literally infinite (that guy would have a finite compute budget). So there is no point in me acting on this either.
Third, let me comment a bit on the meteor example [I don’t know the historical details, so I will be fabulating somewhat]: Sometime in the previous century, we learned about an asteroid hitting earth in the past and killing the dinosaurs. This seems like “there might be a giant explosion from an asteroid”. We did not force everybody to live underground, not because we didn’t take the risk seriously, but because doing so would have been stupid (we probably didn’t have an asteroid-surviving technology anyway, and the best way to get it would have been to stay on the surface). However, we did spend lots of money on the asteroid detection program to understand the risk better. As a result, we now understand this particular “Pascal’s wager” better, and it seems that in this particular case, the risk is small. That is, the risk is small enough that it is more important to focus on other things. However, I believe that we should still spend some effort on preventing asteroid risk (actually, I guess we do do that). And if we had no more-pressing issues (X-risk and otherwise), I totally would invest into some resources asteroid-explosion-survival measures.
To summarize the meteor example: We did not dismiss the risk out of hand. We investigated it seriously, and then we only dismissed it once we understood it. And it never made sense to apply extreme measures, because they wouldn’t have been effective even if the risk was real. (Though I am not sure whether that was on purpose or because we wouldn’t been able to apply extreme measures even if they were appropriate.) ((Uhm, please please please, I hope nobody starts making claims about low-level implications of the asteroid risk-analysis to AI risk-analysis. The low-level details of the two cases have nothing in common, so there are no implications to draw.))
(2) For “stop AI progress” to be rational, it is enough to be uncertain.
Sure, if you claim something like “90% that AI is going to kill us” or “we are totally sure that AI is risky”, then the burden of proof is on you.
But most people who advocate treating AI as risky, and that we should stop/slow/take extreme measures, do not claim anything like “90% that AI is going to kill us”. Most often, you see claims like (i) “at least 5% that AI could kill us”, or (ii) “we really don’t know whether AI might kill us, one way or the other”.
And if you claim non-(i) or non-(ii), the burden of proof totally is on you. And I think that literally nobody in the world has a good enough model of AI & the strategic landscape to justify non-(i) or non-(ii). (Definitely nobody has published a gears-level model like that.)
Also, I think that to make Pascal’s wager on AI rational, it is enough to believe (either (i) or (ii)) together with “we have no quick way of resolving the uncertainty”.
(3) AI risk as Pascal’s wager:
I claim that it is rational to treat AI as a Pascal’s wager, but that in this particular case, the details come out very differently (primarily because we understand AI so little).
To give an illustration (there will be many nonsense bits in the details; the goal is to give the general vibe):
If you know literally nothing about AI at all, it is fair to initially pattern-match it to “yeah yeah, random people talking about the end of the world, that’s bullshit”.
However, once you learn that AI stands for “artificial intelligence” you really should do a double-take: “Hey, intelligence is this thing that maybe makes us, and not other species, dominant on this planet. Also, humans are intelligent and they do horrible things to each other. So it is at least plausible that this could destroy us.” (That’s without knowing anything at all about how AI should work.)
At that point, you should consider whether this AI thing might be achievable. And if it is, say, year 1900, then you should go “I have literally no clue, so I should act as if the chance is non-negligible”. And then (much) later people invent computers and you hear about the substrate-independence thesis. You don’t understand any of this, but it sounds plausible. And then you should treat AI as a real possibility, until you can get a gear’s-level understanding of the details. (Which we still don’t have!)
Also, you try understand whether AI will be dangerous or not, and you will mostly come to the conclusion that you have no clue whatsoever. (And we still don’t even now. And on the current timeline, we probably won’t know until we get superintelligence.)
And from that point onward, you should be like “holy ****, this AI thing could be really dangerous! We should never build any thing like that unless we understand exactly what we are doing!”.
This does not mean you will not build a whole bunch of AI. You will build calculators, and you will build GOFAI, and tabular model-based RL.[3] And then somebody comes up with the idea of deep learning, and you are like “holy ****, we have no clue how this works!”. And you put a moratorium on it, to quickly figure out what to do. And later you probably you end up controlling it pretty tightly, because you don’t know how it scales. And you only do the research on DL slowly, such that your interpretability techniques and other theory never gets behind.
Obviously, we don’t live in a world capable of something like above. But it remains a fact that we now have strong evidence that (1) AGI probably is possible and (2) we still have no clue whether it is going to be harmful of not, despite trying. And if we could do it, the rational thing to do would be to immediatelly stop most AI work until we figure out (2).
When I was younger, I did think a lot about christianity and I have a detailed model that causes me to assign literally 0 probability to it being correct. Similarly, I did engage with some conspiracy theories before dismissing them. I believe that justifies me to tentatively immediatelly dismiss anything that pattern matches to religion and conspiracy theories. However, if I never heard about any religion, then upon encountering one for the first time, I would totally not be justified to dismiss it outright. And let me bite the bullet here: I would say that if you were a random guy in ancient Egypt, without particularly good models of the world or ways to improve them, the rational thing to do could have been to act as if the Egyptian gods were real. [low confidence though, I haven’t thought about this before]
Not to be confused with the hypothesis that the universe actually works according to any of of the religions considered on earth without assuming simulation hypothesis. To that, I assign literally 0 probability.
But for all of those things, you briefly consider whether those things could be dangerous and conclude they aren’t. (Separately from AI risks: In an ideal world, you would also make some experiments to see how these things are affecting the users. EG, to see whether kids’ brains detoriate by relying on calculators. In practice, our science methods sucked back then, so you don’t learn anything. But hey, talking about how a better world would work.)