Nothing is “mere.” I, too, can see the stars on a desert night, and feel them. But do I see less or more? The vastness of the heavens stretches my imagination—stuck on this carousel, my little eye can catch one-million-year-old light. A vast pattern—of which I am a part—perhaps my stuff was belched from some forgotten star, as one is belching there. Or see them with the greater eye of Palomar, rushing all apart from some common starting point when they were perhaps all together. What is the pattern, or the meaning, or the why? It does not do harm to the mystery to know a little about it.
- Richard P. Feynman on The Relation of Physics to Other Sciences
Dalcy
If after all that it still sounds completely wack, check the date. Anything from before like 2003 or so is me as a kid, where “kid” is defined as “didn’t find out about heuristics and biases yet”, and sure at that age I was young enough to proclaim AI timelines or whatevs.
https://twitter.com/ESYudkowsky/status/1650180666951352320
btw there’s no input box for the “How much would you pay for each of these?” question.
although I’ve practiced opening those emotional channels a bit, so this is a less uncommon experience for me than for most
i’m curious, what did you do to open those emotional channels?
Out of the set of all possible variables one might use to describe a system, most of them cannot be used on their own to reliably predict forward time evolution because they depend on the many other variables in a non-Markovian way. But hydro variables have closed equations of motion, which can be deterministic or stochastic but at the least are Markovian.
This idea sounds very similar to this—it definitely seems extendable beyond the context of physics:
We argue that they are both; more specifically, that the set of macrostates forms the unique maximal partition of phase space which 1) is consistent with our observations (a subjective fact about our ability to observe the system) and 2) obeys a Markov process (an objective fact about the system’s dynamics).
I don’t see any feasible way that gene editing or ‘mind uploading’ could work within the next few decades. Gene editing for intelligence seems unfeasible because human intelligence is a massively polygenic trait, influenced by thousands to tens of thousands of quantitative trait loci.
I think the authors in the post referenced above agree with this premise and still consider human intelligence augmentation via polygenic editing to be feasible within the next few decades! I think their technical claims hold up, so personally I’d be very excited to see MIRI pivot towards supporting their general direction. I’d be interested to hear your opinions on their post.
I am curious as to how often the asymptotic results proven using features of the problem that seem basically practically-irrelevant become relevant in practice.
Like, I understand that there are many asymptotic results (e.g., free energy principle in SLT) that are useful in practice, but i feel like there’s something sus about similar results from information theory or complexity theory where the way in which they prove certain bounds (or inclusion relationship, for complexity theory) seem totally detached from practicality?
joint source coding theorem is often stated as why we can consider the problem of compression and redundancy separately, but when you actually look at the proof it only talks about possibility (which is proven in terms of insanely long codes) and thus not-at-all trivial that this equivalence is something that holds in the context of practical code-engineering
complexity theory talks about stuff like quantifying some property over all possible boolean circuits of a given size which seems to me considering a feature of the problem just so utterly irrelevant to real programs that I’m suspicious it can say meaningful things about stuff we see in practice
as an aside, does the P vs NP distinction even matter in practice? we just … seem to have very good approximation to NP problems by algorithms that take into account the structures specific to the problem and domains where we want things to be fast; and as long as complexity methods doesn’t take into account those fine structures that are specific to a problem, i don’t see how it would characterize such well-approximated problems using complexity classes.
Wigderson’s book had a short section on average complexity which I hoped would be this kind of a result, and I’m unimpressed (the problem doesn’t sound easier—now how do you specify the natural distribution??)
- When Are Results from Computational Complexity Not Too Coarse? by Jul 3, 2024, 7:06 PM; 41 points) (
- Sep 28, 2024, 3:59 PM; 7 points) 's comment on Cole Wyeth’s Shortform by (
Found an example in the wild with Mutual information! These equivalent definitions of Mutual Information undergo concept splintering as you go beyond just 2 variables:
interpretation: common information
… become co-information, the central atom of your I-diagram
interpretation: relative entropy b/w joint and product of margin
… become total-correlation
interpretation: joint entropy minus all unshared info
… become bound information
… each with different properties (eg co-information is a bit too sensitive because just a single pair being independent reduces the whole thing to 0, total-correlation seems to overcount a bit, etc) and so with different uses (eg bound information is interesting for time-series).
The limit’s probably much higher with sub-Landauer thermodynamic efficiency.
Epistemic Motif of Abstract-Concrete Cycles & Domain Expansion
‘Symmetry’ implies ‘redundant coordinate’ implies ‘cyclic coordinates in your Lagrangian / Hamiltonian’ implies ‘conservation of conjugate momentum’
And because the action principle (where the true system trajectory extremizes your action, i.e. integral of Lagrangian) works in various dynamical systems, the above argument works in non-physical dynamical systems.
Thus conserved quantities usually exist in a given dynamical system.
mmm, but why does the action principle hold in such a wide variety of systems though? (like how you get entropy by postulating something to be maximized in an equilibrium setting)
Bella is meeting a psychotherapist, but they treat her fear as something irrational. This doesn’t help, and only makes Bella more anxious. She feels like even her therapist doesn’t understand her.
How would one find a therapist in their local area who’s aware of what’s going on in the EA/rat circles such that they wouldn’t find statements about, say, x-risks as being schizophrenic/paranoid?
I am very interested in this, especially in the context of alignment research and solving not-yet-understood problems in general. Since I have no strong commitments this month (and was going to do something similar to this anyways), I will try this every day for the next two weeks and report back on how it goes (writing this comment as a commitment mechanism!)
Have a large group of people attempt to practice problems from each domain, randomizing the order that they each tackle the problems in. (The ideal version of this takes a few months)
...
As part of each problem, they do meta-reflection on “how to think better”, aiming specifically to extract general insights and intuitions. They check what processes seemed to actually lead to the answer, even when they switch to a new domain they haven’t studied before.
Within this upper-level feedback loop (at the scale of whole problems, taking hours or days), I’m guessing a lower-level loop would involve something like cognitive strategy tuning to get real-time feedback as you’re solving the problems?
I had something like locality in mind when writing this shortform, the context being: [I’m in my room → I notice itch → I realize there’s a mosquito somewhere in my room → I deliberately pursue and kill the mosquito that I wouldn’t have known existed without the itch]
But, again, this probably wouldn’t amount to much selection pressure, partially due to the fact that the vast majority of mosquito population exists in places where such locality doesn’t hold i.e. in an open environment.
Makes sense. I think we’re using the terms differently in scope. By “DL paradigm” I meant to encompass the kind of stuff you mentioned (RL-directing-SS-target (active learning), online learning, different architecture, etc) because they really seemed like “engineering challenges” to me (despite them covering a broad space of algorithms) in the sense that capabilities researchers already seem to be working on & scaling them without facing any apparent blockers to further progress, i.e. in need of any “fundamental breakthroughs”—by which I was pointing more at paradigm shifts away from DL like, idk, symbolic learning.
But the evolutionary timescale at which mosquitos can adapt to avoid detection must be faster than that of humans adapting to find mosquitos itchy! Or so I thought—my current boring guess is that (1) mechanisms for the human body to detect foreign particles are fairly “broad”, (2) the required adaptation from the mosquitos to evade them are not-way-too-simple, and (3) we just haven’t put enough selection pressure to make such change happen.
To me, the fact that the human brain basically implements SSL+RL is very very strong evidence that the current DL paradigm (with a bit of “engineering” effort, but nothing like fundamental breakthroughs) will kinda just keep scaling until we reach point-of-no-return. Does this broadly look correct to people here? Would really appreciate other perspectives.
What are the errors in this essay? As I’m reading through the Brain-like AGI sequence I keep seeing this post being referenced (but this post says I should instead read the sequence!)
I would really like to have a single reference post of yours that contains the core ideas about phasic dopamine rather than the reference being the sequence posts (which is heavily dependent on a bunch of previous posts; also Post 5 and 6 feels more high-level than this one?)
Answering my own question, review / survey articles like https://arxiv.org/abs/1811.12560 seem like a pretty good intro.
What is the shape predicted by compmech under a generation setting, and do you expect it instead of the fractal shape to show up under, say, a GAN loss? If so, and if their shapes are sufficiently distinct from the controls that are run to make sure the fractals aren’t just a visualization artifact, that would be further evidence in favor of the applicability of compmech in this setup.