I wonder whether stuff like “turn off the wifi” is about costly signals? (My first-order opinion is still that it’s dumb.)
rotatingpaguro
I started reading, but I can’t understand what the parity problem is, in the section that ought to define it.
I guess, the parity problem is finding the set S given black-box access to the function, is it?
I think I prefer Claude’s attitude as assistant. The other two look too greedy to be wise.
Referring to the section “What is Intelligence Even, Anyway?”:
I think AIXI is fairly described as a search over the space of Turing machines. Why do you think otherwise? Or maybe are you making a distinction at a more granular level?
When you say “true probability”, what do you mean?
The current hypotheses I have about what you mean are (in part non-exclusive):
You think some notion of objective, non-observer dependent probability makes sense, and that’s the true probability.
You do not think “true probability” exists, you are referencing to it to say the market price is not anything like that.
You define “true probability” a probability that observers contextually agree on (like a coin flip observed by humans who don’t know the thrower).
Anton Leicht says evals are in trouble as something one could use in a regulation or law. Why? He lists four factors. Marius Hobbhahn of Apollo also has thoughts. I’m going to post a lot of disagreement and pushback, but I thank Anton for the exercise, which I believe is highly useful.
I think there’s one important factor missing: if you really used evals for regulation, then they would be gamed. I trust more the eval when the company is not actually at stake on it. If it was, there would be a natural tendence for evals to slide towards empty box-checking.
I sometimes wonder about this. This post does pose the question, but I don’t think it gives an analysis that could make me change my mind on anything, it’s too shallow and not adversarial.
I read part of the paper. That there’s a cultural difference north-south about honesty and willingness to break the rules matches my experience on the ground.
I find this intellectually stimulating, but it does not look useful in practice, because with repeated i.i.d. data the information in the data is much higher than the prior if the prior is diffuse/universal/ignorance.
Italians over time sorted themselves geographically by honesty, which is both weird and damn cool, and also makes a lot of sense. There are multiple equilibria, so let everyone find the one that suits them. We need to use this more in logic puzzles. In one Italian villa everyone tells the truth, in the other…
I can’t get access to the paper, anyone has a tip on this?
I agree with whay you say about how to maximize what you get out of an interview. I also agree about that discussion vs. debate distinction you make, and I wasn’t specifically trying to go there when I used the word “debate”, I was just sloppy with words.
I guess you agree that it is friction to create a social norm that you should do a read up of the other person material before engaging in public. I expect less discussions would happen. There is not a clear threshold at how much you should be prepared.
I guess we disagree about how much value do we lose due to eliminating discussions that could have happaned, vs. how much value we gain by eliminating some lower quality discussions.
Another angle I have in mind that sidesteps this direct compromise, is that maybe what we value out of such discussions is not just doing an optimal play in terms of information transmitted between the parties. A public discussion has many different viewers. In the case at hand, I expect many people get more out of the discussion if they can see Wolfram think through the thing for the first time in real time, rather than having two informed people start discussing finer points in medias res.
I see your proposed condition for meaningful debate as bureaucracy that adds friction rather than value.
I somewhat disagree with Tenobrus’ commentary about Wolfram.
I watched the full podcast, and my impression was that Wolfram uses a “scientific hat”, of which he is well aware of, which comes with a certain ritual and method for looking at new things and learning them. Wolfram is doing the ritual of understanding what Yudkowsky says, which involves picking at the details of everything.
Wolfram often recognizes that maybe he feels like agreeing with something, but “scientifically” he has a duty to pick it apart. I think this has to be understood as a learning process rather than as a state of belief.
So, should the restrictions on gambling be based on feedback loop length? Should sport betting be broadly legal when about the far enough future?
current inference scaling methods tend to be tied to CoT and the like, which are quite transparent
Aschenbrenner in Situational Awareness predicts illegible chains of thought are going to prevail because they are more efficient. I know of one developer claiming to do this (https://platonicresearch.com/) but I guess there must be many.
Related, I have a vague understanding on how product safety certification works in EU, and there are multiple private companies doing the certification in every state.
Half-informed take on “the SNPs explain a small part of the genetic variance”: maybe the regression methods are bad?
Not sure if I missed something because I read quickly, but: all these are purely correlational studies, without causal inference, right?
OpenAI is recklessly scaling AI. Besides accelerating “progress” toward mass extinction, it causes increasing harms. Many communities are now speaking up. In my circles only, I count seven new books critiquing AI corps. It’s what happens when you scrape everyone’s personal data to train inscrutable models (computed by polluting data centers) used to cheaply automate out professionals and spread disinformation and deepfakes.
Could you justify that it causes increasing harms? My intuition is that OpenAI is currently net-positive without taking into account future risks. It’s just an intuition, however, I have not spent time thinking about it and writing down numbers.
(I agree it’s net-negative overall.)
My current thinking is that
relying on the CoT staying legible because it’s English, and
hoping the (racing) labs do not drop human language when it becomes economically convenient to do so,
were hopes to be destroyed as quickly as possible. (This is not a confident opinion, it originates from 15 minutes of vague thoughts.)
To be clear, I don’t think that in general it is right to say “Doing the right thing is hopeless because no one else is doing it”, I typically prefer to rather “do the thing that if everyone did that, the world would be better”. My intuition is that it makes sense to try to coordinate on bottlenecks like introducing compute governance and limiting flops, but not on a specific incremental improvement of AI techniques, because I think the people thinking things like “I will restrain myself from using this specific AI sub-techinque because it increases x-risk” are not coordinated enough to self-coordinate at that level of detail, and are not powerful enough to have an influence through small changes.
(Again, I am not confident, I can imagine paths were I’m wrong, haven’t worked through them.)
(Conflict of interest disclosure: I collaborate with people who started developing this kind of stuff before Meta.)