I think it’s not the case that “neural networks” as discussed in this post made AlphaGo. That is, almost of the difficulty in making AlphaGo happen was picking which neural network architecture would solve the problem / buying fast enough computers to train it in a reasonable amount of time. A more recent example might be something like “model-based reinforcement learning”; for many years ‘everyone knew’ that this was the next place to go, while no one could write down an algorithm that actually performed well.
I think the underlying point—if you want to think of new things, you need to think original thoughts instead of signalling “I am not a traditionalist”—is broadly correct even if the example fails.
That said, I agree with you that the example seems unfortunately timed. In 2007, some CNNs had performed well on a handful of tasks; the big wins were still ~4-5 years in the future. If the cached wisdom had been “we need faster computers,” I think the cached wisdom would have looked pretty good.
I worry that this comment dances around the basic update to be made.
Part of this post makes fun of people who were excited about neural networks. Neural network-based approaches have done extremely well. Eliezer’s example wasn’t just “unfortunately timed.” Eliezer was wrong.
I think that’s a pretty simplistic view of the post, but given that view, I agree that’s the right update to make.
Why does it seem simplistic? Like, one of the central points of the post you link is that we should think about the specific technical features of proposals, instead of focusing on marketing questions of which camp a proposal falls into. And Eliezer saying he’s “no fan of neurons” is in the context of him responding to a comment by someone with the username Marvin Minsky defending the book Perceptrons (the post is from the Overcoming Bias era, when comments did not have threading or explicit parents).
I basically read this as Eliezer making fun of low-nuance people, not people excited about NNs; in that very post he excitedly describes a NN-based robotics project!
But that robotics project was viewed by Eliezer as an example of carefully-designed biological imitation in which the mechanism of action was known by the researchers into the deep details. Across multiple posts, Eliezer’s views from this time period emphasize that he believes that AGI can only come from a well-understood AI architecture—either a detailed imitation of the brain, or a crafted logic-based approach. This robotics project was an example of the latter, despite the fact that it used neurons.
This robot ran on a “neural network” built by detailed study of biology. The network had twenty neurons or so. Each neuron had a separate name and its own equation. And believe me, the robot’s builders knew how that network worked.
Where does that fit into the grand dichotomy? Is it top-down? Is it bottom-up? Calling it “parallel” or “distributed” seems like kind of a silly waste when you’ve only got 20 neurons—who’s going to bother multithreading that?
So this would be, in my view, another clear example of Eliezer being excited about an AI paradigm that ultimately did not lead to the black-box neural network-based LLMs that actually seem to have put us on the path to AGI.
I think that’s a pretty simplistic view of the post
To clarify, I wasn’t claiming that the point of this post is to mock neural network proponents. It’s not. It’s just a few paragraphs of the post. Updated original comment to clarify.
And Eliezer saying he’s “no fan of neurons” is in the context of him responding to a comment by someone with the username Marvin Minsky defending the book Perceptrons (the post is from the Overcoming Bias era, when comments did not have threading or explicit parents).
Can you say more why you think that context is relevant? He says “this may be clearer from other posts”, which implies to me that his “not being a fan of neurons” is not specific to that specific discussion (since I imagine he wrote those other posts independently of Marvin_Minsky’s comment).
(I have more things to say in response to your comment here, but I’d like to hear your answer to the above first!)
Can you say more why you think that context is relevant?
Yeah; from my perspective the main question here is something like “how much nuance does a statement have, and what does that imply about how far you can draw inferences from it?”. I think people are often rounding Eliezer off to a simplified model and then judging the simplified model’s predictions and then attributing that judgment to Eliezer, in a way that I think is probably inaccurate.
For this particular point, there’s also the question of what a “fan of neurons” even is; the sorts you see today are pretty different from the sorts you would see back in 2010, and different from the sort that Marvin Minsky would have seen.
Not as relevant to the narrow point, but worth pointing out somewhere, is that I’m pretty sure that even if Eliezer had been aware of the potential of modern ANNs ahead of time, I think he probably would have filtered that out of his public speech because of concerns about the alignability of those architectures, in a way that makes it not obvious how to count predictions. [Of course he can’t get any points for secretly predicting it without hashed comments, but it seems less obvious that he should lose points for not predicting it.]
Thanks for the additional response. I’ve thought through the details here as well. I think that the written artifacts he left are not the kinds of writings left by someone who actually thinks neural networks will probably work, capabilities-wise.
As you read through these collected quotes, consider how strongly “he doesn’t expect ANNs to work” and “he expects ANNs to work” predict each quote:
In Artificial Intelligence, everyone outside the field has a cached result for brilliant new revolutionary AI idea—neural networks, which work just like the human brain! New AI Idea: complete the pattern: “Logical AIs, despite all the big promises, have failed to provide real intelligence for decades—what we need are neural networks!”
This cached thought has been around for three decades. Still no general intelligence. But, somehow, everyone outside the field knows that neural networks are the Dominant-Paradigm-Overthrowing New Idea, ever since backpropagation was invented in the 1970s. Talk about your aging hippies.
...
I’m no fan of neurons; this may be clearer from other posts
...
But there is just no law which says that if X has property A and Y has property A then X and Y must share any other property. “I built my network, and it’s massively parallel and interconnected and complicated, just like the human brain from which intelligence emerges! Behold, now intelligence shall emerge from this neural network as well!” And nothing happens. Why should it?
...
Wasn’t it in some sense reasonable to have high hopes of neural networks? After all, they’re just like the human brain, which is also massively parallel, distributed, asynchronous, and -
Hold on. Why not analogize to an earthworm’s brain, instead of a human’s?
A backprop network with sigmoid units… actually doesn’t much resemble biology at all. Around as much as a voodoo doll resembles its victim. The surface shape may look vaguely similar in extremely superficial aspects at a first glance. But the interiors and behaviors, and basically the whole thing apart from the surface, are nothing at all alike. All that biological neurons have in common with gradient-optimization ANNs is… the spiderwebby look.
And who says that the spiderwebby look is the important fact about biology? Maybe the performance of biological brains has nothing to do with being made out of neurons, and everything to do with the cumulative selection pressure put into the design.
Do these strike you as things which could plausibly be written by someone who actually anticipated the modern revolution?
there’s also the question of what a “fan of neurons” even is; the sorts you see today are pretty different from the sorts you would see back in 2010, and different from the sort that Marvin Minsky would have seen.
If Eliezer wasn’t a fan of those particular ANNs, in 2010, because those literal empirically tried setups hadn’t yet led to AGI… That’s an uninteresting complaint. It’s trivial. ANN proponents also wouldn’t anticipate AGI from already-tried experiments which had already failed to produce AGI.
The interesting version of the claim is the one which talks about research directions, no? About being excited about neural network research in terms of its future prospects?
I’m pretty sure that even if Eliezer had been aware of the potential of modern ANNs ahead of time, I think he probably would have filtered that out of his public speech because of concerns about the alignability of those architectures
In the world where he was secretly aware, he could have pretended to not expect much of ANNs. In that case, that’s dishonest. Also risky, it’s possibly safer to just not bring it up and not direct even more attention to the matter. If you think that X is a capabilities hazard, then I think a good rule of thumb is don’t talk about X.
So, even privileging this “he secretly knew” hypothesis by considering it explicitly, it isn’t predicting observed reality particularly strongly, since “don’t talk about it at all” is another reasonable prediction of that hypothesis, and that didn’t happen.
in a way that makes it not obvious how to count predictions.
Let’s consider what incentives we want to set up. We want people who can predict the future to be recognized and appreciated, and we want people who can’t to be taken less seriously in such domains. We do not want predictions to communicate sociohazardous content.
For sociohazards like this, hashed comments should suffice quite well, for this kind of problematic prediction. You can’t fake it if you can’t predict it in advance. If you can predict it in advance, you can still get credit, without leaking much information.
I am therefore (hopefully predictably) unimpressed by hypotheses around secret correct predictions which clash with his actual public writing, unless he had verifiably contemporary predictions which were secret but correct.
[Of course he can’t get any points for secretly predicting it without hashed comments, but it seems less obvious that he should lose points for not predicting it.]
Conservation of expected evidence. If you would have updated upwards on his predictive abilities if he had made hashed comments and then revealed them, then observing not-that makes you update downwards (eta—on average, with a few finicky details here that I think work out to the same overall conclusion; happy to discuss if you want).
Do these strike you as things which could plausibly be written by someone who actually anticipated the modern revolution?
I do not think I claimed that Eliezer anticipated the modern revolution, and I would not claim that based on those quotes.
The point that I have been attempting to make since here is that ‘neural networks_2007’, and the ‘neural networks_1970s’ Eliezer describes in the post, did not point to the modern revolution; in fact other things were necessary. I see your point that this is maybe a research taste question—even if it doesn’t point to the right idea directly, does it at least point there indirectly?--to which I think it is evidence against Eliezer’s research taste (on what will work, not necessarily on what will be alignable).
[I also have long thought Eliezer’s allergy to the word “emergence” is misplaced (and that it’s a useful word while thinking about dynamical systems modeling in a reductionistic way, which is a behavior that I think he approves of) while agreeing with him that I’m not optimistic about people whose plan for building intelligence doesn’t route thru them understanding what intelligence is and how it works in a pretty deep way.]
Conservation of expected evidence. If you would have updated upwards on his predictive abilities if he had made hashed comments and then revealed them, then observing not-that makes you update downwards (eta—on average, with a few finicky details here that I think work out to the same overall conclusion; happy to discuss if you want).
I agree with regards to Bayesian superintelligences but not bounded agents, mostly because I think this depends on how you do the accounting. Consider the difference between scheme A, where you transfer prediction points from everyone who didn’t make a correct prediction to people who did make correct predictions, and scheme B, where you transfer prediction points from people who make incorrect predictions to people who make correct predictions, leaving untouched people who didn’t make predictions. On my understanding, things like logical induction and infrabayesianism look more like scheme B.
I do not think I claimed that Eliezer anticipated the modern revolution, and I would not claim that based on those quotes.
The point that I have been attempting to make since here is that ‘neural networks_2007’, and the ‘neural networks_1970s’ Eliezer describes in the post, did not point to the modern revolution; in fact other things were necessary.
I apologize if I have misunderstood your intended point. Thanks for the clarification. I agree with this claim (insofar as I understand what the 2007 landscape looked like, which may be “not much”). I think that the claim is not that interesting, though, but this might be coming down to semantics.
The following is what I perceived us to disagree on, so I’d consider us to be in agreement on the point I originally wanted to discuss:
I see your point that this is maybe a research taste question—even if it doesn’t point to the right idea directly, does it at least point there indirectly?--to which I think it is evidence against Eliezer’s research taste (on what will work, not necessarily on what will be alignable).
I’m not optimistic about people whose plan for building intelligence doesn’t route thru them understanding what intelligence is and how it works in a pretty deep way
Yeah. I think that in a grown-up world, we would do this, and really take our time.
On my understanding, things like logical induction and infrabayesianism look more like scheme B.
Nice, I like this connection. Will think more about this, don’t want to hastily unpack my thoughts into a response which isn’t true to my intuitions here.
I was recently looking at Yudkowsky’s (2008) “Artificial Intelligence as a Positive and Negative Factor in Global Risk” and came across this passage which seems relevant here:
Friendly AI is not a module you can instantly invent at the exact moment when it is first needed, and then bolt on to an existing, polished design which is otherwise completely unchanged.
The field of AI has techniques, such as neural networks and evolutionary programming, which have grown in power with the slow tweaking of decades. But neural networks are opaque—the user has no idea how the neural net is making its decisions—and cannot easily be rendered unopaque; the people who invented and polished neural networks were not thinking about the long-term problems of Friendly AI. Evolutionary programming (EP) is stochastic, and does not precisely preserve the optimization target in the generated code; EP gives you code that does what you ask, most of the time, under the tested circumstances, but the code may also do something else on the side. EP is a powerful, still maturing technique that is intrinsically unsuited to the demands of Friendly AI. Friendly AI, as I have proposed it, requires repeated cycles of recursive self-improvement that precisely preserve a stable optimization target.
The most powerful current AI techniques, as they were developed and then polished and improved over time, have basic incompatibilities with the requirements of Friendly AI as I currently see them. The Y2K problem—which proved very expensive to fix, though not global-catastrophic—analogously arose from failing to foresee tomorrow’s design requirements. The nightmare scenario is that we find ourselves stuck with a catalog of mature, powerful, publicly available AI techniques which combine to yield non-Friendly AI, but which cannot be used to build Friendly AI without redoing the last three decades of AI work from scratch.
If the cached wisdom had been “we need faster computers,” I think the cached wisdom would have looked pretty good.
If you think neural networks are like brains, you might think that you would get human-like cognitive abilities at human-like sizes. I think this was a very common view (and it has aged quite well IMO).
I think Eliezer does disagree. I find his disagreement fairly annoying. He calls biological anchors the “trick that never works” and gives an initial example of Moravec predicting AGI in 2010 in the book Mind Children.
But as far as I can tell so far that’s just Eliezer putting words in Moravec’s mouth. Moravec doesn’t make very precise predictions in the book, but the heading of the relevant section is “human equivalence in 40 years” (i.e. 2028, the book was written in 1988). Eliezer thinks that Moravec ought to think that human-level AI and shortly thereafter a singularity will occur at the time when a giant cluster is as big as a brain, which Moravec puts in 2010. But I don’t see any evidence that Moravec agreed with that implication, and the book seems to generally talk about a timeframe like 2030-2040. Eliezer repeated this claim in our conversation but still didn’t really provide any indication Moravec held this view.
To the extent that people were imagining neural networks, I don’t think they would expect trained neural networks to be the size of a computing cluster. It’s not not the straightforward extrapolation from the kinds of neural networks people were actually computing, so someone going on vibes wouldn’t make that forecast. And if you try to actually pencil out the training cost it’s clear it won’t work, since you have to run a neural network a huge number of times during training, so someone trying to think things through on paper wouldn’t think that either. At least since the 1990 I’ve seen a lot of people making predictions along these lines, but as far as I can tell they seem to give actual predictions in the 2020s or 2030s which currently look quite good to me relative to every other forecasting methodology.
This graph nicely summarizes his timeline from Mind Children in 1988. The book itself presents his view that AI progress is primarily constrained by compute power available to most researchers, which is usually around that of a PC.
Moravec et al were correct in multiple key disagreements with EY et al:
That progress was smooth and predictable from Moore’s Law (similar to how the arrival of flight is postdictable from ICE progress)
That AGI would be based on brain-reverse engineering, and thus will be inherently anthropomorphic
That “recursive self-improvement” was mostly relevant only in the larger systemic sense (civilization level)
LLMs are far more anthropomorphic (brain-like) than the fast clean consequential reasoners EY expected:
close correspondence to linguistic cortex (internal computations and training objective)
complete with human-like cognitive biases!
unexpected human-like limitations: struggle with simple tasks like arithmetic, longer term planning, etc
AGI misalignment insights from jungian psychology more effective/useful/popular than MIRI’s core research
All of this was predicted from the systems/cybernetic framework/rubric that human minds are software constructs, brains are efficient and tractable, and thus AGI is mostly about reverse engineering the brain and then downloading/distilling human mindware into the new digital substrate.
I don’t know if the graph settles the question—is Moravec predicting AGI at “Human equivalence in a supercomputer” or “Human equivalence in a personal computer”? Hard to say from the graph.
The fact that he specifically talks about “compute power available to most researchers” makes it more clear what his predictions are. Taken literally that view would suggest something like: a trillion dollar computing budget spread across 10k researchers in 2010 would result in AGI in not-too-long, which looks a bit less plausible as a prediction but not out of the question.
I can’t believe that post is sitting at 185 karma considering how it opens with a complete blatant misquote/lie about moravec’s central prediction, and only gets worse from there.
Moravec predicted—in mind children in 1988! - AGI in 2028, based on moore’s law and the brain reverse engineering assumption. He was prescient—a true prophet/futurist. EY was wrong and his attempt to smear Moravec here is simply embarrassing.
Even with some disagreements writ how powerful AI can be, I definitely agreee that Eliezer is pretty bad epistemically speaking on anything related to AI or alignment topics, and we should stop treating him as any kind of authority.
I think it’s not the case that “neural networks” as discussed in this post made AlphaGo. That is, almost of the difficulty in making AlphaGo happen was picking which neural network architecture would solve the problem / buying fast enough computers to train it in a reasonable amount of time. A more recent example might be something like “model-based reinforcement learning”; for many years ‘everyone knew’ that this was the next place to go, while no one could write down an algorithm that actually performed well.
I think the underlying point—if you want to think of new things, you need to think original thoughts instead of signalling “I am not a traditionalist”—is broadly correct even if the example fails.
That said, I agree with you that the example seems unfortunately timed. In 2007, some CNNs had performed well on a handful of tasks; the big wins were still ~4-5 years in the future. If the cached wisdom had been “we need faster computers,” I think the cached wisdom would have looked pretty good.
I worry that this comment dances around the basic update to be made.
Part of this post makes fun of people who were excited about neural networks. Neural network-based approaches have done extremely well. Eliezer’s example wasn’t just “unfortunately timed.” Eliezer was wrong.
(Edited “This post” → “Part of this post”)
I think that’s a pretty simplistic view of the post, but given that view, I agree that’s the right update to make.
Why does it seem simplistic? Like, one of the central points of the post you link is that we should think about the specific technical features of proposals, instead of focusing on marketing questions of which camp a proposal falls into. And Eliezer saying he’s “no fan of neurons” is in the context of him responding to a comment by someone with the username Marvin Minsky defending the book Perceptrons (the post is from the Overcoming Bias era, when comments did not have threading or explicit parents).
I basically read this as Eliezer making fun of low-nuance people, not people excited about NNs; in that very post he excitedly describes a NN-based robotics project!
But that robotics project was viewed by Eliezer as an example of carefully-designed biological imitation in which the mechanism of action was known by the researchers into the deep details. Across multiple posts, Eliezer’s views from this time period emphasize that he believes that AGI can only come from a well-understood AI architecture—either a detailed imitation of the brain, or a crafted logic-based approach. This robotics project was an example of the latter, despite the fact that it used neurons.
So this would be, in my view, another clear example of Eliezer being excited about an AI paradigm that ultimately did not lead to the black-box neural network-based LLMs that actually seem to have put us on the path to AGI.
To clarify, I wasn’t claiming that the point of this post is to mock neural network proponents. It’s not. It’s just a few paragraphs of the post. Updated original comment to clarify.
Can you say more why you think that context is relevant? He says “this may be clearer from other posts”, which implies to me that his “not being a fan of neurons” is not specific to that specific discussion (since I imagine he wrote those other posts independently of Marvin_Minsky’s comment).
(I have more things to say in response to your comment here, but I’d like to hear your answer to the above first!)
Yeah; from my perspective the main question here is something like “how much nuance does a statement have, and what does that imply about how far you can draw inferences from it?”. I think people are often rounding Eliezer off to a simplified model and then judging the simplified model’s predictions and then attributing that judgment to Eliezer, in a way that I think is probably inaccurate.
For this particular point, there’s also the question of what a “fan of neurons” even is; the sorts you see today are pretty different from the sorts you would see back in 2010, and different from the sort that Marvin Minsky would have seen.
Not as relevant to the narrow point, but worth pointing out somewhere, is that I’m pretty sure that even if Eliezer had been aware of the potential of modern ANNs ahead of time, I think he probably would have filtered that out of his public speech because of concerns about the alignability of those architectures, in a way that makes it not obvious how to count predictions. [Of course he can’t get any points for secretly predicting it without hashed comments, but it seems less obvious that he should lose points for not predicting it.]
Thanks for the additional response. I’ve thought through the details here as well. I think that the written artifacts he left are not the kinds of writings left by someone who actually thinks neural networks will probably work, capabilities-wise.
As you read through these collected quotes, consider how strongly “he doesn’t expect ANNs to work” and “he expects ANNs to work” predict each quote:
Do these strike you as things which could plausibly be written by someone who actually anticipated the modern revolution?
If Eliezer wasn’t a fan of those particular ANNs, in 2010, because those literal empirically tried setups hadn’t yet led to AGI… That’s an uninteresting complaint. It’s trivial. ANN proponents also wouldn’t anticipate AGI from already-tried experiments which had already failed to produce AGI.
The interesting version of the claim is the one which talks about research directions, no? About being excited about neural network research in terms of its future prospects?
In the world where he was secretly aware, he could have pretended to not expect much of ANNs. In that case, that’s dishonest. Also risky, it’s possibly safer to just not bring it up and not direct even more attention to the matter. If you think that X is a capabilities hazard, then I think a good rule of thumb is don’t talk about X.
So, even privileging this “he secretly knew” hypothesis by considering it explicitly, it isn’t predicting observed reality particularly strongly, since “don’t talk about it at all” is another reasonable prediction of that hypothesis, and that didn’t happen.
Let’s consider what incentives we want to set up. We want people who can predict the future to be recognized and appreciated, and we want people who can’t to be taken less seriously in such domains. We do not want predictions to communicate sociohazardous content.
For sociohazards like this, hashed comments should suffice quite well, for this kind of problematic prediction. You can’t fake it if you can’t predict it in advance. If you can predict it in advance, you can still get credit, without leaking much information.
I am therefore (hopefully predictably) unimpressed by hypotheses around secret correct predictions which clash with his actual public writing, unless he had verifiably contemporary predictions which were secret but correct.
Conservation of expected evidence. If you would have updated upwards on his predictive abilities if he had made hashed comments and then revealed them, then observing not-that makes you update downwards (eta—on average, with a few finicky details here that I think work out to the same overall conclusion; happy to discuss if you want).
[EDITed out a final part for now]
I do not think I claimed that Eliezer anticipated the modern revolution, and I would not claim that based on those quotes.
The point that I have been attempting to make since here is that ‘neural networks_2007’, and the ‘neural networks_1970s’ Eliezer describes in the post, did not point to the modern revolution; in fact other things were necessary. I see your point that this is maybe a research taste question—even if it doesn’t point to the right idea directly, does it at least point there indirectly?--to which I think it is evidence against Eliezer’s research taste (on what will work, not necessarily on what will be alignable).
[I also have long thought Eliezer’s allergy to the word “emergence” is misplaced (and that it’s a useful word while thinking about dynamical systems modeling in a reductionistic way, which is a behavior that I think he approves of) while agreeing with him that I’m not optimistic about people whose plan for building intelligence doesn’t route thru them understanding what intelligence is and how it works in a pretty deep way.]
I agree with regards to Bayesian superintelligences but not bounded agents, mostly because I think this depends on how you do the accounting. Consider the difference between scheme A, where you transfer prediction points from everyone who didn’t make a correct prediction to people who did make correct predictions, and scheme B, where you transfer prediction points from people who make incorrect predictions to people who make correct predictions, leaving untouched people who didn’t make predictions. On my understanding, things like logical induction and infrabayesianism look more like scheme B.
I apologize if I have misunderstood your intended point. Thanks for the clarification. I agree with this claim (insofar as I understand what the 2007 landscape looked like, which may be “not much”). I think that the claim is not that interesting, though, but this might be coming down to semantics.
The following is what I perceived us to disagree on, so I’d consider us to be in agreement on the point I originally wanted to discuss:
Yeah. I think that in a grown-up world, we would do this, and really take our time.
Nice, I like this connection. Will think more about this, don’t want to hastily unpack my thoughts into a response which isn’t true to my intuitions here.
I was recently looking at Yudkowsky’s (2008) “Artificial Intelligence as a Positive and
Negative Factor in Global Risk” and came across this passage which seems relevant here:
If you think neural networks are like brains, you might think that you would get human-like cognitive abilities at human-like sizes. I think this was a very common view (and it has aged quite well IMO).
Agreed, tho I think Eliezer disagrees?
I think Eliezer does disagree. I find his disagreement fairly annoying. He calls biological anchors the “trick that never works” and gives an initial example of Moravec predicting AGI in 2010 in the book Mind Children.
But as far as I can tell so far that’s just Eliezer putting words in Moravec’s mouth. Moravec doesn’t make very precise predictions in the book, but the heading of the relevant section is “human equivalence in 40 years” (i.e. 2028, the book was written in 1988). Eliezer thinks that Moravec ought to think that human-level AI and shortly thereafter a singularity will occur at the time when a giant cluster is as big as a brain, which Moravec puts in 2010. But I don’t see any evidence that Moravec agreed with that implication, and the book seems to generally talk about a timeframe like 2030-2040. Eliezer repeated this claim in our conversation but still didn’t really provide any indication Moravec held this view.
To the extent that people were imagining neural networks, I don’t think they would expect trained neural networks to be the size of a computing cluster. It’s not not the straightforward extrapolation from the kinds of neural networks people were actually computing, so someone going on vibes wouldn’t make that forecast. And if you try to actually pencil out the training cost it’s clear it won’t work, since you have to run a neural network a huge number of times during training, so someone trying to think things through on paper wouldn’t think that either. At least since the 1990 I’ve seen a lot of people making predictions along these lines, but as far as I can tell they seem to give actual predictions in the 2020s or 2030s which currently look quite good to me relative to every other forecasting methodology.
This graph nicely summarizes his timeline from Mind Children in 1988. The book itself presents his view that AI progress is primarily constrained by compute power available to most researchers, which is usually around that of a PC.
Moravec et al were correct in multiple key disagreements with EY et al:
That progress was smooth and predictable from Moore’s Law (similar to how the arrival of flight is postdictable from ICE progress)
That AGI would be based on brain-reverse engineering, and thus will be inherently anthropomorphic
That “recursive self-improvement” was mostly relevant only in the larger systemic sense (civilization level)
LLMs are far more anthropomorphic (brain-like) than the fast clean consequential reasoners EY expected:
close correspondence to linguistic cortex (internal computations and training objective)
complete with human-like cognitive biases!
unexpected human-like limitations: struggle with simple tasks like arithmetic, longer term planning, etc
AGI misalignment insights from jungian psychology more effective/useful/popular than MIRI’s core research
All of this was predicted from the systems/cybernetic framework/rubric that human minds are software constructs, brains are efficient and tractable, and thus AGI is mostly about reverse engineering the brain and then downloading/distilling human mindware into the new digital substrate.
I don’t know if the graph settles the question—is Moravec predicting AGI at “Human equivalence in a supercomputer” or “Human equivalence in a personal computer”? Hard to say from the graph.
The fact that he specifically talks about “compute power available to most researchers” makes it more clear what his predictions are. Taken literally that view would suggest something like: a trillion dollar computing budget spread across 10k researchers in 2010 would result in AGI in not-too-long, which looks a bit less plausible as a prediction but not out of the question.
I can’t believe that post is sitting at 185 karma considering how it opens with a complete blatant misquote/lie about moravec’s central prediction, and only gets worse from there.
Moravec predicted—in mind children in 1988! - AGI in 2028, based on moore’s law and the brain reverse engineering assumption. He was prescient—a true prophet/futurist. EY was wrong and his attempt to smear Moravec here is simply embarrassing.
I’m reminded of this thread from 2022: https://www.lesswrong.com/posts/27EznPncmCtnpSojH/link-post-on-deference-and-yudkowsky-s-ai-risk-estimates?commentId=SLjkYtCfddvH9j38T#SLjkYtCfddvH9j38T
Even with some disagreements writ how powerful AI can be, I definitely agreee that Eliezer is pretty bad epistemically speaking on anything related to AI or alignment topics, and we should stop treating him as any kind of authority.