Thanks so much for this comment (and sorry for taking ~1 year to respond!!). I really liked everything you said.
For 1 and 2, I agree with everything and don’t have anything to add. 3. I agree that there is something about the input/output mapping that is meaningful but it is not everything. Having a full theory for exactly the difference, and what the distinctions between what structure counts as interesting internal computation (not a great descriptor of what I mean but can’t think of anything better right now) vs input output computation would be great.
4. I also think a great goal would be in generalizing and formalizing what an “observer” of a computation is. I have a few ideas but they are pretty half-baked right now.
5. That is an interesting point. I think it’s fair. I do want to be careful to make sure that any “disagreements” are substantial and not just semantic squabling here. I like your distinction between representation work and computational work. The idea of using vs. performing a computation is also interesting. At the end of the day I am always left craving some formalism where you could really see the nature of these distinctions.
6. Sounds like a good idea!
7. Agreed on all counts.
8. I was trying to ask the question if there is anything that tells us that the output node is semantically meaningful without reference to e.g. the input images of cats, or even knowledge of the input data distribution. Interpretability work, both in artificial neural networks and more traditionally in neuroscience, always use knowledge of input distributions or even input identity to correlate activity of neurons to the input, and in that way assign semantics to neural activity (e.g. recently, othello board states, or in neuroscience jennifer aniston neurons or orientation tuned neurons) . But when I’m sitting down with my eyes closed and just thinking, there’s no homonculus there that has access to input distributions on my retina that can correlate some activity pattern to “cat.” So how can the neural states in my brain “represent” or embody or whatever word you want to use, the semantic information of cat, without this process of correlating to some ground truth data. WHere does “cat” come from when theres no cat there in the activity?!
Thanks so much for this comment (and sorry for taking ~1 year to respond!!). I really liked everything you said.
For 1 and 2, I agree with everything and don’t have anything to add.
3. I agree that there is something about the input/output mapping that is meaningful but it is not everything. Having a full theory for exactly the difference, and what the distinctions between what structure counts as interesting internal computation (not a great descriptor of what I mean but can’t think of anything better right now) vs input output computation would be great.
4. I also think a great goal would be in generalizing and formalizing what an “observer” of a computation is. I have a few ideas but they are pretty half-baked right now.
5. That is an interesting point. I think it’s fair. I do want to be careful to make sure that any “disagreements” are substantial and not just semantic squabling here. I like your distinction between representation work and computational work. The idea of using vs. performing a computation is also interesting. At the end of the day I am always left craving some formalism where you could really see the nature of these distinctions.
6. Sounds like a good idea!
7. Agreed on all counts.
8. I was trying to ask the question if there is anything that tells us that the output node is semantically meaningful without reference to e.g. the input images of cats, or even knowledge of the input data distribution. Interpretability work, both in artificial neural networks and more traditionally in neuroscience, always use knowledge of input distributions or even input identity to correlate activity of neurons to the input, and in that way assign semantics to neural activity (e.g. recently, othello board states, or in neuroscience jennifer aniston neurons or orientation tuned neurons) . But when I’m sitting down with my eyes closed and just thinking, there’s no homonculus there that has access to input distributions on my retina that can correlate some activity pattern to “cat.” So how can the neural states in my brain “represent” or embody or whatever word you want to use, the semantic information of cat, without this process of correlating to some ground truth data. WHere does “cat” come from when theres no cat there in the activity?!
9. SO WILD