You’re right but I like the chef example anyway. Even if cherry picked, it does get at a core truth—this kind of intuition evolves in every field. I love the stories of old hands intuitively seeing things a mile away.
angmoh
Sutskever’s response to Dwarkesh in their interview was a convincing refutation of this argument for me:
Dwarkesh Patel
So you could argue that next-token prediction can only help us match human performance and maybe not surpass it? What would it take to surpass human performance?Ilya Sutskever
I challenge the claim that next-token prediction cannot surpass human performance. On the surface, it looks like it cannot. It looks like if you just learn to imitate, to predict what people do, it means that you can only copy people. But here is a counter argument for why it might not be quite so. If your base neural net is smart enough, you just ask it — What would a person with great insight, wisdom, and capability do? Maybe such a person doesn’t exist, but there’s a pretty good chance that the neural net will be able to extrapolate how such a person would behave. Do you see what I mean?Dwarkesh Patel
Yes, although where would it get that sort of insight about what that person would do? If not from…Ilya Sutskever
From the data of regular people. Because if you think about it, what does it mean to predict the next token well enough? It’s actually a much deeper question than it seems. Predicting the next token well means that you understand the underlying reality that led to the creation of that token. It’s not statistics. Like it is statistics but what is statistics? In order to understand those statistics to compress them, you need to understand what is it about the world that creates this set of statistics? And so then you say — Well, I have all those people. What is it about people that creates their behaviors? Well they have thoughts and their feelings, and they have ideas, and they do things in certain ways. All of those could be deduced from next-token prediction. And I’d argue that this should make it possible, not indefinitely but to a pretty decent degree to say — Well, can you guess what you’d do if you took a person with this characteristic and that characteristic? Like such a person doesn’t exist but because you’re so good at predicting the next token, you should still be able to guess what that person who would do. This hypothetical, imaginary person with far greater mental ability than the rest of us
“Dwarkesh chose excellent questions throughout, displaying an excellent sense of when to follow up and how, and when to pivot.”
This is the basic essence of why Dwarkesh does such good interviews. He does the groundwork to be able to ask relevant and interesting questions, ie. actually reading their books/works, and seems to consistently have put actual thought into analysing the worldview of his subjects.
The unambitiousness of modern geoengineering in general is dismaying.
For my Australian perspective: in the early 1900s there were people discussing how make use of the massive tracts of desert wasteland than make up most of the outback (ie: https://en.wikipedia.org/wiki/Bradfield_Scheme). None of this stuff could be considered today—one tree getting chopped down is liable to make the news: https://www.bbc.com/news/world-australia-54700074
Hard to escape the conclusion that we should all go lie in a ditch so as to guarantee that no impact to anything occurs ever.
This seems about right. Sam is a bit of a cowboy and probably doesn’t bother involving the board more than he absolutely has to.
Stefánsson’s “The Fat of the Land” is not really worth reading for any scientific insight today, but it’s entertaining early 1900s anthropology.
I don’t have much of an opinion on any specific diet approach, but I can tell you my own experience with weight loss: I’ve always been between 15-25% bodyfat, yoyoing around. This routine isn’t ideal, so I too am a ‘victim’ of the weight gain phenomenon.
I have no satisfying answers for “why are we getting fatter” or “what makes caloric deficits so hard to maintain”. I appreciate the diet blogging community that tries to tackle these questions with citizen science.
I assume you’re familiar with Vilhjálmur Stefánsson’s work if you are interested in low protein carnivore diets, but I was really was surprised to see how similar the ‘ex150’ sounds compared to the classic ~80:20 fat:protein experiments. These aren’t really new ideas—although I’m sure there’s a lot more information available on the details.
Anyway, dieting seems like something where while people on average fail, you do see some individual successes, so it’s worth poking around the edges and giving things a go. It’s always nice to see results from the coalface.
Ultimately the new GLP-1 agonist weightloss drugs seem to be awesome by both data and anecdata. So the food composition experimentation might fade away a bit over the next few years for the express purpose of weight loss.
Good post.
For Westerners looking to get a palatable foothold on the priorities and viewpoints of the CCP (and Xi), I endorse “The Avoidable War” written last year by Kevin Rudd (former Prime Minister of Australia, speaks mandarin, has worked in China, loads of diplomatic experience—imo about as good of a person that exists to interpret Chinese grand strategy and explain it from a Western viewpoint). The book is (imo, for a politician), impressively objective in its analysis.
Some good stuff in there explaining the nature of Chinese cynicism about foreign motivations that echoes some of what is written in this post, but with a bit more historical background and strategic context.
Yeah—it’s odd, but TC is a self-professed contrarian after all.
I think the question here is: why doesn’t he actually address the fundamentals of the AGI doom case? The “it’s unlikely / unknown” position is really quite a weak argument which I doubt he would make if he actually understood EY’s position.
Seeing the state of the discourse on AGI risk just makes it more and more clear that the AGI risk awareness movement has failed to express its arguments in terms that non-rationalists can understand.People like TC should the first type of public intellectual to grok it, because EY’s doom case is is highly analogous to market dynamics. And yet.
For example, a major point of disagreement between me and Eliezer is that Eliezer often dismisses plans as “too complicated to work in practice,” but that dismissal seems divorced from experience with getting things to work in practice (e.g. some of the ideas that Eliezer dismisses are not much more complex than RLHF with AI assistants helping human raters). In fact I think that you can implement complex things by taking small steps—almost all of these implementation difficulties do improve with empirical feedback.
EY’s counter to this?
@Gerald Monroe On the question of Japan’s unique lack of variation, I think it’s unlikely to be decisive here. The ‘monoculture’ argument may have some merit, but even a genetically ‘homogenous’ population has plenty of variation—especially one 125m strong like the Japanese.
Fertility related traits are just so fundamental to genetic fitness that selection is guaranteed to wring out the higher fertility alleles where the environment allows.
Imo Japan is one of the more illuminating examples on this topic:
Japan had a TFR of 5 in the 1930s. It’s been only 3 generations since Japan’s TFR began to fall, and France took 5 to stablise around the current level (1830s-1980s). I agree that the trend since 2005 is too short term to be sure, but it’s interesting to note! The above modelling suggests that a faster fertility transition should result in a faster bounceback—the lower the TFR the more adaptive high-TFR genes + cultures will be relatively.
The fertility transition hit East Asia harder and faster than it did Europe. There’s merit to the theory that it’s because Europe had a slower transition to today’s mainstream fertility-surpressing universal culture (technological advancement, enlightenment values. women’s lib etc), since much of these cultural changes were developed in the West (consider the analogy to megafauna in Africa).
It’s extremely difficult to quantify this sort of thing but it does support a model where both genes and culture are load-bearing inputs to TFR. In countries where culture propped up fertility one way or another there could be said to be a cultural fertility overhang, and when these forces were removed TFR naturally cratered in the short term. Where countries had less cultural overhang, or a slower transition from high-TFR culture to low-TFR culture, the transition was less dramatic because there was time for cultural counter-developments or genetic selection to act.
The example of Sth Korea (TFR >5 until the 60s) supports some of these theories. The timing is especially interesting—the 60s were a major leap forward in progressive cultural hegemony, and Sth Korea (an extremely poor society prior) copped that right in the face after the Korean War. The idea is that the speed of TFR-decline is related to the severity of cultural change—makes sense to me.
An optimistic Sth Korean pro-natalist could interpret this current ultra-low TFR period as evidence of an extremely effective ‘weeding’ process which is sure to be followed by a period of high fertility preferences as the most enthusiastic ‘breeders’ will be all that remain!
Note: When I refer to culture above I take the expansive definition which includes technology, wealth, social changes—ie. anything that isn’t genetic.
There’s good reason to believe the fertility transition as a general phenomenon is subject to negative feedback, thus almost certain to be self correcting. Selection self-evidently favours high fertility culture and genes.
See this study for an attempt to model the effects:
Correlations in fertility across generations: can low fertility persist?
Martin Kolk, Daniel Cownden and Magnus Enquist
Published:22 March 2014https://royalsocietypublishing.org/doi/10.1098/rspb.2013.2561#d3e788,
″… Our models suggest a mechanism in which the recent fertility decline may be reversed in the long run. Intergenerational fertility correlations create cultural and genetic selection processes that favour lifestyles with higher fertility. Only through continuous cultural change, introducing novel lifestyles associated with reduced child-bearing, can low fertility persist.”
An example in favour of this model is France. France is generally regarded as a low-fertility pioneer, but today has the highest TFR in Western Europe.
It’s still worth paying attention to policy, but worth noting the strong undercurrents at play which are likely somewhat independent of govt tweaking. I think genetics are likely a more relevant factor than has been mentioned in this thread so far.
Agree—Gell-Mann amnesia sums up my experience with trying to get ChatGPT to be useful for a professional context so far. It is weak on accuracy and details.
My questions:
Is this something that can be overcome through skilful prompting?
Is there something about LLMs that makes these issues difficult to overcome?
ChatGPT is very conservative with providing factual information even when it’s possible to tease out the relevant information (i.e. topics like law or commercial activities), is this purposeful throttling?
The speed issue is the #1 factor stopping me from trying audiobooks. A book might take me 4-8 hours to read but the internet tells me audio is 2-3x slower. I have a lot of other prejudices against audiobooks (flipping / skimming less easy, less focus on the task etc) but that’s the main one.
Seconded—I’d like to see more of this angle of analysis too. I assume the reason why the ‘soft take-off’ is underdiscussed is that tech people a) pay more attention to the news on AI, and b) see the future of this stuff viscerally and the endgame is what looms. I think that’s not wrong, because the endgame is indeed transformative. But how we get there and how quickly it happens is a completely open question.
I work in the AEC industry (Architecture, Engineering, Construction) − 90%+ of people have zero idea about recent advances in AI. But on the other hand, should they be personally worried about their employment prospects in the next decade? I feel like lots of LW-type people would say “Yes!”—I can only speak personally, but it’s really hard for me to see it happening. If only for the fact that doing anything in meatspace takes a long time. There are plenty of great ‘digital’ solutions to problems in this industry that have been around for 10+ years and have still made no headway. I know AGI is different, but it’s worth mentioning how slow things can be, and how much of a grinding bureaucracy many industries are.
The other way I think about it is that ultimately human concerns (politics, agency etc) underpin all economic activity, and there will be massive political and bureaucratic opposition to extreme levels of economic ‘disruption’ (in the negative sense of mass unemployment etc). I foresee active responses from the populace and governments shaping the path this takes to a significant degree (eg. increases in labour force protectionism). Not so much that capitalism just takes to AI like a duck to water, governments let it happen, and 99% of people end up in a terrafoam box in a few short years.
I can anecdotally report that when I started consistently getting up immediately upon my alarm going off the subjective feeling of the first 5-10 minutes was far superior and I didn’t feel much tiredness even with a relatively short night of sleep. I started doing this through setting my alarm to the maximum latest time possible and still allow me to get to work on time, and then noticed how much better I felt while doing this (previously I was a chronic snooze-button masher and felt pretty groggy waking up).
I have noticed in the WFH/office phases of the past 2 years the effect has persisted—a worse subjective waking up experience when WFH. I think it’s because without the pressure of needing to be at work, I don’t develop an automatic routine of simply getting up unthinkingly when my alarm goes off.
I think part of this is related to how good your internal biological clock is. I’m the type of person that will wake up pretty consistently a few minutes before an alarm goes off anyway (sometimes even when the alarm is not my usual time, such as for a flight). Maybe the ‘get up immediately’ approach wouldn’t work as well if this isn’t you.
So I looked into this, and the ‘refine win condition’ seems to be an actual technique that some of the best 3pt shooters do employ! I looked up some interviews with some of the best basketball shooters (Steph Curry and Ray Allen) and they mention playing little games with themselves while practice. They will do things such a define win condition as a swish instead of just making the shot, or employing some particular type of rhythm or footwork, or slightly altering the angle of their shot.
But on the other hand, they also mention plenty of repetitive drilling, and watching the available footage of them practicing, it is hard to see a ‘stop after win’ approach in action. There are videos of Steph Curry practicing with a coach who passes him the ball—it looks pretty repetitive to me (although maybe he’s playing some mental games with himself that are hard to pick up?).
Thanks for the post—skill acquisition does seem like an area lacking attention from the rationalist community in general.
I wonder if these learnings would apply to sports. “Stop after win” doesn’t seem like a productive idea if you are trying to get better at, say, shooting a 3pt shot in basketball—the traditional approach is “reps reps reps”. Thoughts?
I’ll take the bait since this is one of the cool meta aspects of poker!
There’s a saying in online poker: “move up to where they respect your raises”—it’s poking fun at the notion that it’s possible to play well without modelling your opponents. The idea is that it’s not valid to conclude that if you lose to a poor player, that you weren’t “rewarded properly”—it is in fact your fault for lacking the situational awareness to adjust your strategy.
For a good player sitting with a person who thinks ‘all reds’ is a good hand, it’ll be obvious before you ever see their cards.
Anyway your point is right about the difficulty of learning ‘organically’ where you only play bad players. A common failure mode in online poker involved players getting stuck at local maximums strategically—they’d adopt an autopilot-style strategy that did very well at lower limits surrounded by ‘all reds’ types, but get owned when they moved up to higher stakes and failed to adjust.