This seems like a solid attempt at understanding the concepts, but I think the true complexity of these concepts is much greater than you’re anticipating. I simply dropped your post into GPT4′s input—this isn’t that great of a summary tbh but it’s a start:
The post discusses the idea of consciousness and goal-formation as a function of recurrence in neural networks, attributing the concept to Giulio Tononi. It also raises the possibility of using this idea to enforce alignment in artificial intelligence systems. Here are some comments on the points made in the post:
Recurrence in neural networks: Recurrent neural networks (RNNs) are a type of neural network architecture that allows for the feeding back of output to input, which can help model temporal dependencies and maintain a memory of previous inputs. In the context of human brains, cortical columns may exhibit recurrent connections, which could potentially be related to the formation of goals and consciousness. However, the exact relationship between recurrence and consciousness is not yet well-established, and more research is needed to explore this idea.
Goal-setting agency and intent: The post suggests that goal-setting agency and intent may arise from the feedback of conceptions and ideas through cortical columns, with unmotivated and near-deterministic pass-throughs serving as a snapshot of the thinking process. While this is an interesting idea, it is important to note that consciousness and goal-formation are complex phenomena that likely involve multiple interacting processes in the brain. Further research is needed to determine the extent to which recurrence might contribute to these processes.
Alignment enforcement through editorial control: The post proposes that controlling the feedback-to-prompt in an artificial intelligence system could help enforce alignment with human-controlled preferences and tasking. This is an intriguing idea, as it suggests that by pruning or modifying the outputs of a system, we could potentially guide its behavior towards desired goals. However, implementing such an editorial control mechanism in practice could be challenging, and it may not be sufficient on its own to ensure complete alignment with human values.
Relationship to Asimov’s Laws: The idea of enforcing alignment through editorial control of feedback in an AI system is reminiscent of Asimov’s Laws of Robotics, which outline basic principles for robots to follow in order to ensure they behave safely and in accordance with human preferences. However, it is important to recognize that enforcing alignment in AI systems is a complex problem that may require more nuanced approaches than simply implementing a set of rules or guidelines.
In summary, the post presents an interesting idea about the role of recurrence in consciousness and goal-formation, and the potential application of this concept to enforce alignment in AI systems. While the idea has some merits, further research is needed to better understand the relationship between recurrence and consciousness, and to explore the practical challenges of implementing alignment enforcement mechanisms in artificial intelligence systems.
This seems like a solid attempt at understanding the concepts, but I think the true complexity of these concepts is much greater than you’re anticipating. I simply dropped your post into GPT4′s input—this isn’t that great of a summary tbh but it’s a start: