Um. If you want to convince a mathematician like Terry Tao to be interested in AI alignment, you will need to present yourself as a reasonably competent mathematician or related expert and actually formulate an AI problem in such a way so that someone like Terry Tao would be interested in it. If you yourself are not interested in the problem, then Terry Tao will not be interested in it either.
Terry Tao is interested in random matrix theory (he wrote the book on it), and random matrix theory is somewhat related to my approach to AI interpretability and alignment. If you are going to send these problems to a mathematician, please inform me about this before you do so.
My approach to alignment: Given matrices A1,…,Ar;B1,…,Br, define a superoperator Γ(A1,…,Ar;B1,…,Br) by setting
Γ(A1,…,Ar;B1,…,Br)(X)=∑rk=1AkXB∗k, and define Φ(A1,…,Ar)=Γ(A1,…,Ar;A1,…,Ar). Define the L2-spectral radius of A1,…,Ar as ρ2(A1,…,Ar)=ρ(Φ(A1,…,Ar))1/2. Here, ρ(A)=limn→∞∥An∥1/n is the usual spectral radius.
Define ρK2,d(A1,…,Ar)=max{ρ(Γ(A1,…,Ar;X1,…,Xr))ρ2(X1,…,Xr):X1,…,Xr∈Md(K)}. Here, K is either the field of reals, field of complex numbers, or division ring of quaternions.
Given matrices A1,…,Ar;B1,…,Br, define
∥(A1,…,Ar)≃(B1,…,Br)∥=Γ(A1,…,Ar;B1,…,Br)ρ2(A1,…,Ar)ρ2(B1,…,Br). The value ∥(A1,…,Ar)≃(B1,…,Br)∥ is always a real number in the interval [0,1] that is a measure of how jointly similar the tuples (A1,…,Ar),(B1,…,Br) are. The motivation behind ρK2,d(A1,…,Ar) is that ρK2,d(A1,…,Ar)ρ2(A1,…,Ar) is always a real number in [0,1] (well except when the denominator is zero) that measures how well A1,…,Ar can be approximated by d×d-matrices. Informally, ρK2,d(A1,…,Ar)ρ2(A1,…,Ar) measures how random A1,…,Ar are where a lower value of ρK2,d(A1,…,Ar)ρ2(A1,…,Ar) indicates a lower degree of randomness.
A better theoretical understanding of ρK2,d(A1,…,Ar) would be great. If X1,…,Xr∈Md(K) and ρ(Γ(A1,…,Ar;X1,…,Xr))ϕ2(X1,…,Xr) is locally maximized, then we say that (X1,…,Xr) is an LSRDR of (A1,…,Ar). Said differently, (X1,…,Xr)∈Md(K) is an LSRDR of (A1,…,Ar) if the similarity ∥(A1,…,Ar)≃(X1,…,Xr)∥ is maximized.
Here, the notion of an LSRDR is a machine learning notion that seems to be much more interpretable and much less subject to noise than many other machine learning notions. But a solid mathematical theory behind LSRDRs would help us understand not just what LSRDRs do, but the mathematical theory would help us understand why they do it.
Problems in random matrix theory concerning LSRDRs:
Suppose that U1,…,Ur are random matrices (according to some distribution). Then what are some bounds for ρK2,d(U1,…,Ur).
Suppose that U1,…,Ur are random matrices and A1,…,Ar are non-random matrices. What can we say about the spectrum of Γ(A1,…,Ar;U1,…,Ur)? My computer experiments indicate that this spectrum satisfies the circular law, and the radius of the disc for this circular law is proportional to ρ2(A1,…,Ar), but a proof of this circular law would be nice.
Tensors can be naturally associated with collections of matrices. Suppose now that U1,…,Ur are the matrices associated with a random tensor. Then what are some bounds for ρK2,d(U1,…,Ur).
P.S. By massively downvoting my posts where I talk about mathematics that is clearly applicable to AI interpretability and alignment, the people on this site are simply demonstrating that they need to do a lot of soul searching before they annoy people like Terry Tao with their lack of mathematical expertise.
P.P.S. Instead of trying to get a high profile mathematician like Terry Tao to be interested in problems, it may be better to search for a specific mathematician who is an expert in a specific area related to AI alignment since it may be easier to contact a lower profile mathematician, and a lower profile mathematician may have more specific things to say and contribute. You are lucky that Terry Tao is interested in random matrix theory, but this does not mean that Terry Tao is interested in anything in the intersection between alignment and random matrix theory. It may be better to search harder for mathematicians who are interested in your specific problems.
P.P.P.S. To get more mathematicians interested in alignment, it may be a good idea to develop AI systems that behave much more mathematically. Neural networks currently do not behave very mathematically since they look like the things that engineers would come up with instead of mathematicians.
P.P.P.P.S. I have developed the notion of an LSRDR for cryptocurrency research because I am using this to evaluate the cryptographic security of cryptographic functions.
Um. If you want to convince a mathematician like Terry Tao to be interested in AI alignment, you will need to present yourself as a reasonably competent mathematician or related expert and actually formulate an AI problem in such a way so that someone like Terry Tao would be interested in it. If you yourself are not interested in the problem, then Terry Tao will not be interested in it either.
Terry Tao is interested in random matrix theory (he wrote the book on it), and random matrix theory is somewhat related to my approach to AI interpretability and alignment. If you are going to send these problems to a mathematician, please inform me about this before you do so.
My approach to alignment: Given matrices A1,…,Ar;B1,…,Br, define a superoperator Γ(A1,…,Ar;B1,…,Br) by setting
Γ(A1,…,Ar;B1,…,Br)(X)=∑rk=1AkXB∗k, and define Φ(A1,…,Ar)=Γ(A1,…,Ar;A1,…,Ar). Define the L2-spectral radius of A1,…,Ar as ρ2(A1,…,Ar)=ρ(Φ(A1,…,Ar))1/2. Here, ρ(A)=limn→∞∥An∥1/n is the usual spectral radius.
Define ρK2,d(A1,…,Ar)=max{ρ(Γ(A1,…,Ar;X1,…,Xr))ρ2(X1,…,Xr):X1,…,Xr∈Md(K)}. Here, K is either the field of reals, field of complex numbers, or division ring of quaternions.
Given matrices A1,…,Ar;B1,…,Br, define
∥(A1,…,Ar)≃(B1,…,Br)∥=Γ(A1,…,Ar;B1,…,Br)ρ2(A1,…,Ar)ρ2(B1,…,Br). The value ∥(A1,…,Ar)≃(B1,…,Br)∥ is always a real number in the interval [0,1] that is a measure of how jointly similar the tuples (A1,…,Ar),(B1,…,Br) are. The motivation behind ρK2,d(A1,…,Ar) is that ρK2,d(A1,…,Ar)ρ2(A1,…,Ar) is always a real number in [0,1] (well except when the denominator is zero) that measures how well A1,…,Ar can be approximated by d×d-matrices. Informally, ρK2,d(A1,…,Ar)ρ2(A1,…,Ar) measures how random A1,…,Ar are where a lower value of ρK2,d(A1,…,Ar)ρ2(A1,…,Ar) indicates a lower degree of randomness.
A better theoretical understanding of ρK2,d(A1,…,Ar) would be great. If X1,…,Xr∈Md(K) and ρ(Γ(A1,…,Ar;X1,…,Xr))ϕ2(X1,…,Xr) is locally maximized, then we say that (X1,…,Xr) is an LSRDR of (A1,…,Ar). Said differently, (X1,…,Xr)∈Md(K) is an LSRDR of (A1,…,Ar) if the similarity ∥(A1,…,Ar)≃(X1,…,Xr)∥ is maximized.
Here, the notion of an LSRDR is a machine learning notion that seems to be much more interpretable and much less subject to noise than many other machine learning notions. But a solid mathematical theory behind LSRDRs would help us understand not just what LSRDRs do, but the mathematical theory would help us understand why they do it.
Problems in random matrix theory concerning LSRDRs:
Suppose that U1,…,Ur are random matrices (according to some distribution). Then what are some bounds for ρK2,d(U1,…,Ur).
Suppose that U1,…,Ur are random matrices and A1,…,Ar are non-random matrices. What can we say about the spectrum of Γ(A1,…,Ar;U1,…,Ur)? My computer experiments indicate that this spectrum satisfies the circular law, and the radius of the disc for this circular law is proportional to ρ2(A1,…,Ar), but a proof of this circular law would be nice.
Tensors can be naturally associated with collections of matrices. Suppose now that U1,…,Ur are the matrices associated with a random tensor. Then what are some bounds for ρK2,d(U1,…,Ur).
P.S. By massively downvoting my posts where I talk about mathematics that is clearly applicable to AI interpretability and alignment, the people on this site are simply demonstrating that they need to do a lot of soul searching before they annoy people like Terry Tao with their lack of mathematical expertise.
P.P.S. Instead of trying to get a high profile mathematician like Terry Tao to be interested in problems, it may be better to search for a specific mathematician who is an expert in a specific area related to AI alignment since it may be easier to contact a lower profile mathematician, and a lower profile mathematician may have more specific things to say and contribute. You are lucky that Terry Tao is interested in random matrix theory, but this does not mean that Terry Tao is interested in anything in the intersection between alignment and random matrix theory. It may be better to search harder for mathematicians who are interested in your specific problems.
P.P.P.S. To get more mathematicians interested in alignment, it may be a good idea to develop AI systems that behave much more mathematically. Neural networks currently do not behave very mathematically since they look like the things that engineers would come up with instead of mathematicians.
P.P.P.P.S. I have developed the notion of an LSRDR for cryptocurrency research because I am using this to evaluate the cryptographic security of cryptographic functions.