Currently a MATS scholar working with Vanessa Kosoy on the learning-theoretic agenda for AI Alignment.
Arjun Pitchanathan
Epistemic status: Quick dump of something that might be useful to someone. o3 and Opus 4 independently agree on the numerical calculations for the bolded result below, but I didn’t check the calculations myself in any detail.
When we say “roughly”, e.g. or would be fine; it may be a judgement call on our part if the bound is much larger than that.
Let . With probability , set , and otherwise draw . Let . Let and . We will investigate latents for .
Set , then note that the stochastic error ) because induces perfect conditional independence and symmetry of and . Now compute the deterministic errors of , , , which are equal to respectively.
Then it turns out that with , all of these latents have error greater than , if you believe this claude opus 4 artifact (full chat here, corroboration by o3 here). Conditional on there not being some other kind of latent that gets better deterministic error, and the calculations being correct, I would expect that a bit more fiddling around could produce much better bounds, say or more, since I think I’ve explored very little of the search space.
e.g. one could create more As and Bs by either adding more Ys, or more Xs and Zs. Or one could pick the probabilities out of some discrete set of possibilities instead of having them be fixed.
Yes, thanks!
representation of a variable for variable
Hm, I don’t understand what is supposed to be here.
Isn’t it the case that when you sing a high note, you feel something higher in your mouth/larynx/whatever , and when you sing a low note, you feel something lower? Seems difficult to tell whether I actually do need to do that or I’ve just conditioned myself to, because of the metaphor.
If you’re reading the text in a two-dimensional visual display, you are giving yourself an advantage over the LLM. You should actually be reading it in a one-dimensional format with new-line symbols.
(disclosure, I only skimmed your COT for like a few seconds)
the real odds would be less about the ELO and more on whether he was drunk while playing me
not sure if that would help :)
I don’t think this will work because we are already using subscripts to denote which environment’s list we are referring to
I’m not. I guess this is the part that makes it confusing
for readability we define and to be the accessible and outer action spaces of respectively
Do you have a suggestion for alternate notation? I use this because we often need to refer to the action space corresponding to a state. I think this would be needed even with the language framing.
(I also assigned to make it more readable)
Do you have any predictions about the first year when AI assistance will give a 2x/10x/100x factor “productivity boost” to AI research?
Hey, I’ve been reading stuff from this community since about 2017. I’m now in the SERI MATS program where I’m working with Vanessa Kosoy. Looking forward to contributing something back after lurking for so long :P
I have not read the post, but am confused as to why it is at −3 karma. Would some of the downvoters care to explain their reasoning?