Student in fundamental and applied mathematics, interested in theoretical computer science and AI alignment
Matrice Jacobine
An aligned ASI, if it were possible, would be capable of a degree of perfection beyond that of human institutions.
The corollary of this is that an aligned ASI in the strong sense of “aligned” used here would have to dissolve currently existing human institutions, and the latter will obviously oppose that. As it stand, even if we solve technical alignment (which I do think is plausible at this rate), we’ll end up with either an ASI aligned to a nation-state, or a corporate ASI turning all available matter in economium, both of which are x-risks in the longtermist sense (and maybe even s-risks e.g. in the former case if Xi or Trump are bioconservative and speciesist, which I’m fairly sure they are).
As Yudkowsky wrote nearly 18 years ago in Applause Lights:
Suppose that a group of democratic republics form a consortium to develop AI, and there’s a lot of politicking during the process—some interest groups have unusually large influence, others get shafted—in other words, the result looks just like the products of modern democracies. Alternatively, suppose a group of rebel nerds develops an AI in their basement, and instructs the AI to poll everyone in the world—dropping cellphones to anyone who doesn’t have them—and do whatever the majority says. Which of these do you think is more “democratic,” and would you feel safe with either?
To re-use @Richard_Ngo’s framework of the three waves of AI safety, the first generation around SIAI/MIRI had a tendency to believe that that creating AGI was mostly an engineering problem and dismissed the lines of thought that predicted modern scaling laws. So the idea of being that “group of rebel nerds” and creating Friendly AGI in your basement (which was ostensibly SIAI/MIRI’s goal) could have seemed realistic to them back then.
Then the deep learning revolution of the 2010s happened and it turned out that the first wave of AI safety was wrong and the bottleneck to AGI really is access to large amounts of compute, which you can only get through financial backing by corporations (for DeepMind, Google; for OpenAI, Microsoft; for Anthropic, Amazon), and which is easy for the state to clamp down on.
And then the AI boom of the 2020s happened and the states themselves are now more and more conscious of the threat of AGI. Applause Lights was wrote in 2007. I would predict that by 2027 a private project with the explicit goal of developing AGI to overthrow all existing national governments would receive about the same public reaction as a private project with the explicit goal of developing nuclear weapons to overthrow all existing national governments.
For the third wave of AI safety (quoting Ngo again), there are different ways you can go from this:
Push for your preferred state or corporation to achieve aligned (in the weak sense) ASI, thus trusting them with the entire long-term future of humanity
The more realistic golden mean between those two plans: develop artificial intelligence in a decentralized way in a way that this still let us first achieve a post-scarcity economy, longtermist institutional reform, possibly a long reflection, and only then (possibly) building aligned ASI. This is the common throughline between differential technological development, d/acc, coceleration, Tool AI, organic alignment, mutual alignment, cyborgism, and other related ideas to me.
Oh wow that’s surprising, I thought Ted Chiang was an AI skeptic not too long ago?
“Long” timelines to advanced AI have gotten crazy short
Large Language Models Pass the Turing Test
I’m using the 2016 survey and counting non-binary yes.
@nostalgebraist @Mantas Mazeika “I think this conversation is taking an adversarial tone.” If this is how the conversation is going this might be the case to end it and work on a, well, adversarial collaboration outside the forum.
Would you mind to cross-post this on the EA Forum?
Demonstrating specification gaming in reasoning models
US AI Safety Institute will be ‘gutted,’ Axios reports
It does seem that the LLMs are subject to deontological constraints (Figure 19), but I think that in fact makes the paper’s framing of questions as evaluation between world-states instead of specific actions more apt at evaluating whether LLMs have utility functions over world-states behind those deontological constraints. Your reinterpretation of how those world-state descriptions are actually interpreted by LLMs is an important remark and certainly change the conclusions we can make from this article regarding to implicit bias, but (unless you debunk those results) the most important discoveries of the paper from my point of view, that LLMs have utility functions over world-states which are 1/ consistent across LLMs, 2/ more and more consistent as model size increase, and 3/ can be subject to mechanical interpretability methods, remain the same.
… I don’t agree, but would it at least be relevant that the “soft CCP-approved platitudes” are now AI-safetyist?
So that answer your question “Why does the linked article merit our attention?” right?
Why does the linked article merit our attention?
It is written by a Chinese former politician in a Chinese-owned newspaper.
?
Cooperation for AI safety must transcend geopolitical interference
The current AI strategic landscape: one bear’s perspective
I’m not convinced “almost all sentient beings on Earth” would pick out of the blue (i.e. without chain of thought) the reflectively optimal option at least 60% of the times when asked unconstrained responses (i.e. not even a MCQ).
The most important part of the experimental setup is “unconstrained text response”. If in the largest LLMs 60% of unconstrained text responses wind up being “the outcome it assigns the highest utility”, then that’s surely evidence for “utility maximization” and even “the paperclip hyper-optimization caricature”. What more do you want exactly?
This doesn’t contradict the Thurstonian model at all. This only show order effects are one of the many factors going in utility variance, one of the factors of the Thurstonian model. Why should it be considered differently than any other such factor? The calculations still show utility variance (including order effects) decrease when scaled (Figure 12), you don’t need to eyeball based on a few examples in a Twitter thread on a single factor.
If that was the case we wouldn’t expect to have those results about the VNM consistency of such preferences.
“co-ordination issues” is a major euphemism here: such a global safe AI would not just require the same kind of coordination one generally expect in relations between nation-states (even in the eyes of the most idealistic liberal-internationalists), but effectively having already achieved a world government and species-wide agreement on a same moral philosophy – which may itself require having already achieved at the very least a post-scarcity economy. This is more or less what I mean in the last bullet point by “and only then (possibly) building aligned ASI”.
Depending on whether this advice is available to everyone or only to the leadership of existing institutions, this would fall either under Tool AI (which is one of the approaches in my third bullet point) or state-aligned (but CEV-unaligned) ASI (a known x-risk and plausibly a s-risk).
If the merged AGIs are all CEV-unaligned, I don’t see why we should assume that, just because it is a merger from AGIs from across the world, the merged AGI would suddenly be CEV-aligned.