I think there is another assumption (or another way of framing your assumption, I’m not sure if it is really distinct), which is that all people throughout time are alike. It’s not necessarily groundless, but it is an assumption. There could be some symmetry breaking factor that determines your position that doesn’t simply align with “total humans that ever exist”, and thus renders using the “birth rank” inappropriate to determine what number you have. This thing doesn’t have to be related to reproduction though.
ChosunOne
It still seems slightly fuzzy in that other than check/mate situations no moves are fully mandatory and eg recaptures may occasionally turn out to be the wrong move?
Indeed it can be difficult to know when it is actually better not to continue the line vs when it is, but that is precisely what MCTS would help figure out. MCTS would do actual exploration of board states and the budget for which states it explores would be informed by the policy network. It’s usually better to continue a line vs not, so I would expect MCTS to spend most of its budget continuing the line, and the policy would be updated during training with whether or not the recommendation resulted in more wins. Ultimately though, the policy network is probably storing a fuzzy pattern matcher for good board states (perhaps encoding common lines or interpolations of lines encountered by the MCTS) that it can use to more effectively guide the search by giving it an appropriate score.
To be clear, I don’t think a transformer is completely incapable of doing any search, just that it is probably not learning to do it in this case and is probably pretty inefficient at doing it when prompted to.
In chess, a “line” is sequence of moves that are hard to interrupt. There are kind of obvious moves you have to play or else you are just losing (such as recapturing a piece, moving king out of check, performing checkmate etc). Leela uses the neural network more for policy, which means giving a score to a given board position, which then the MCTS can use to determine whether or not to prune that direction or explore that section more. So it makes sense that Leela would have an embedding of powerful lines as part of its heuristic, since it isn’t doing to main work of search. It’s more pattern recognition on the board state, so it can learn to recognize the kinds of lines that are useful and whether or not they are “present” in the current board state. It gets this information from the MCTS system as it trains, and compresses the “triggers” into the earlier evaluations, which then this paper explores.
It’s very cool work and result, but I feel it’s too strong to say that the policy network is doing search as opposed to recognizing lines from its training at earlier board states.
The cited paper in Section 5 (Conclusion-Limitations) states plainly:
(2) We focus on look-ahead along a single line of play; we do not test whether Leela compares multiple different lines of play (what one might call search). … (4) Chess as a domain might favor look-ahead to an unusually strong extent.The paper is more just looking at how Leela evaluates a given line rather than doing any kind of search. And this makes sense. Pattern recognition is an extremely important part of playing chess (as a player myself), and it is embedded in another system doing the actual search, namely Monte Carlo Tree Search. So it isn’t surprising that it has learned to look ahead in a straight line since that’s what all of its training experience is going to entail. If transformers were any good at doing the search, I would expect a chess bot without employing something like MCTS.
My crux is that LLMs are inherently bad at search tasks over a new domain. Thus, I don’t expect LLMs to scale to improve search.
Anecdotal evidence: I’ve used LLMs extensively and my experience is that LLMs are great at retrieval but terrible at suggestion when it comes to ideas. You usually get something resembling an amalgamation of Google searches vs. suggestions from some kind of insight.
To your question of what to do if you are outmatched and you only have an ASI at your disposal, I think the most logical thing to do is “do what the ASI tells you to”. The problem is that we have no way of predicting the outcomes if there is truly an ASI in the room. If it’s a superintelligence it is going to have better suggestions than anything you can come up with.
Then I wonder, at what point does that matter? Or more specifically, when does that matter in the context of ai-risk?
Clearly there is some relationship between something like “more compute” and “more intelligence” since something too simple cannot be intelligent, but I don’t know where that relationship breaks down. Evolution clearly found a path for optimizing intelligence via proxy in our brains, and I think the fear is that you may yet be able to go quite further than human-level intelligence before the extra compute fails to deliver more meaningful intelligence described in your post.
It seems premature to reject the orthogonality thesis of optimizing for things that “obviously bring more intelligence” before they start to break down.
So if I understand your point correctly, you expect something like “give me more compute” at some point fail to deliver more intelligence since intelligence isn’t just “more compute”?
I don’t think you claim has support as presented. Part of the problem surrounding the question is that we still don’t really have any way of measuring how “conscious” something is. In order to claim that something is or isn’t conscious, you should have some working definition of what conscious means, and how it can be measured. If you want to have a serious discussion instead of competing emotional positions, you need to support the claim with points that can be confirmed or denied. Why doesn’t a goldfish have consciousness, or an earthworm? How can you know? You can borrow concepts from some theoretical concepts like Integrated Information Theory or Non-Trivial Information Closure etc. but you should make some effort to ground the claim.
As for my opinion, whatever your definition of consciousness is I think a digital entity can in principle be conscious since I think consciousness is closely related to certain types of information, but beyond that without a more rigorous perspective I think the conversation isn’t going anywhere fruitful.
Well ultimately no information about the past is truly lost as far as we know. A hyper-advanced civilization could collect all the thermal radiation from earth reflected off of various celestial bodies and recover a near complete history, at least in principle. So I think the more you make it easy for yourself to be reconstructed/resurrected/what have you the sooner it would likely be, and the less alien of an environment you would find yourself in after the fact. Cryo is a good example of having a reasonable expectation of where to end up barring catastrophe since you are preserving a lot of you in good form.
An interesting consequence of your description is that resurrection is possible if you can manage to reconstruct the last brain state of someone who had died. If you go one one step further, then I think it is fairly likely that experience is eternal, since you don’t experience any of the intervening time (akin to your film reel analogy with adding extra frames in between) being dead and since there is no limit to how much intervening time can pass.
I’m curious how much space is left after learning the MSP in the network. Does representing the MSP take up the full bandwidth of the model (even if it is represented inefficiently)? Could you maintain performance of the model by subtracting out the contributions of anything else that isn’t part of the MSP?
I observe this behavior a lot when using GPT-4 to assist in code. The moment it starts spitting out code that has a bug, the likelihood of future code snippets having bugs grows very quickly.
I’ve found that using Bing/Chat-GPT has been enormously helpful in my own workflows. No need to have to carefully read documentation and tutorials just to get a starter template up and running. Sure it breaks here and there, but it seems way more efficient to look up stuff when it goes wrong vs. starting from scratch. Then, while my program is running, I can go back and try to understand what all the options do.
It’s also been very helpful for finding research on a given topic and answering basic questions about some of the main ideas.
I’m not sure how that makes the problem much easier? If you get the maligned superintelligence mask, it only needs to get out of the larger model/send instructions to the wrong people once to have game over scenario. You don’t necessarily get to change it after the fact. And changing it once doesn’t guarantee it doesn’t pop up again.
This could be true, but then you still have the issue of there being “superintelligent malign AI” as one of the masks if your pile of masks is big enough.
At a crude level, the earth represents a source of resources with which to colonize the universe shard and maximize its compute power (and thus find the optimal path to achieve its other goals). Simply utilizing all the available mass on earth to do that as quickly as possible hedges against the possibility of other entities in the universe shard from impeding progress toward its goals.
The question I’ve been contemplating is “Is it worth it to actually try to spend any resources dissassembling all matter on Earth given the cost of needing to devote resources to stop humans from interfering” when weighed against the lost time and energy from solar output given the resources needed to accomplish that. I don’t know, but it’s a reasonable way of getting “kill everyone” from an unaligned goal if it makes sense to pursue that strategy.
Thanks for the measured response.
If I understand the following correctly:
Putin made it very clear on the day of the attack that he was threatening nukes to anyone who “interfered” in Ukraine, with his infamous “the consequences will be such as you have never seen in your entire history” -speech. NATO has been helping Ukraine by training their forces and supplying materiell for years before the invasion, and vowed to keep doing so. This can be considered “calling his bluff” to an extent, or as a piecemeal maneuver in it’s own right. Yet they withdrew their personell from the country in the days and weeks leading up to the attack. Some have called Biden weak for doing that, essentially “clearing the way” for Putin by removing the tripwire force, and maybe he is. What is clear is that he didn’t want for that bluff to be called.
You interpret that as being specifically a warning against overt deployment of troops to Ukraine? I suppose my reading of it was more broad, and as such NATO already fell on the side of violating my understanding of “interfering”. While at the outset I can see that being a strong reason at the beginning of the war, i.e. “Don’t take my attempt at a quick victory away from me or else I’ll nuke you”, I don’t know how feasible that remains over time. Putin can’t think that if the war goes on for months without victory that everyone would just sit on the sidelines forever. I suppose clarity from the Russians about their commitment to the war in general would help especially regarding:
Putin’s unambiguous red line is Russia’s geographical border, and he is trying very successfully and believably to assert a red line over any direct military intervention inside Ukraine, and less successfully over less direct help.
While I agree that Russia’s border is not something NATO tanks should go rolling across, I haven’t seen as strong of a message in recent days threatening nuclear retaliation if say a THAAD battery near the Polish border with Ukraine engaged a Russian fighter jet. NATO could plausibly claim the fighter penetrated Polish airspace (even if it wasn’t actually the case).
In fact, the US and USSR engaged in direct aerial combat in the Korean War, in the infamous “MiG Alley” without escalating into a full fledged war.
That does not mean he would back down from other more direct help, with the infamous Polish fighter jets toeing the line too close for comfort, so the US backed down on that one.
Agreed, but if Russia starts bombing the supply convoys from NATO, that would almost certainly invite more direct NATO intervention. “Russia is bombing humanitarian aid convoys” etc.
All this to say I think the situation is a lot more nuanced than “If NATO fires a single bullet at a Russian it’s the end of civilization”.
Obviously Putin would be in a much stronger position if he would have been able to conquer Ukraine within a few days.
Strangely enough, I think this was the intention. I think the prospect of this war escalating could if nothing else be used to help force the Russians to reevaluate their goals and hasten the end of the war.
Yes I did, and it doesn’t follow that nuclear retaliation is immediate.
Beaufre notes that for piecemeal maneuvers to be effective, they have to be presented as fait accompli – accomplished so quickly that anything but nuclear retaliation would arrive too late to do any good and of course nuclear retaliation would be pointless
Failure to perform the fait accompli means that options other than nuclear retaliation are possible.
When Putin called that obvious bluff, it would have damaged the credibility and thus the deterrence value of that same statement when applied to NATO members or Taiwan, weakening the effect of US deterrence, and thus potentially encouraging another state (like China) to try to call an American bluff elsewhere (essentially inviting a piecemeal maneuver).
Take this statement and reverse the positions. If NATO calls Russia’s bluff that any and all military assistance to Ukraine would be met with nuclear retaliation, as they have already done, then Russia by this logic is inviting a piecemeal maneuver on the part of NATO.
Russia is fighting an aggressive war. NATO can clearly signal via way of action that it has no intention of threatening the existence of the Russian state.
Ukraine is already in an effectively total war (from their perspective; Russia is not totally mobilized) with Russia. Russian forces are already targeting Ukrainian civilian centers with the apparent aim of inflicting civilian casualties and making the refugee situation worse.
Involvement here doesn’t escalate the situation inside of Ukraine beyond its borders. I’d rather see counterpoints to my arguments than blanket assertions I didn’t read the article or that it “addresses my points”. Please point out exactly where I’m missing something.
I think the factors that determine your reference class can be related to changes over time in the environment you inhabit, not just how you are built. This is what I mean by not necessarily related to reproduction. If I cloned a billion people with the same cloning vat over 1000 years, how then would you determine reference class? But maybe something about the environment would narrow that reference class despite all the people being made by the same process (like the presence or absence of other things in the environment as they relate to the people coming out of the vat).