Notes on Superwisdom & Moral RSI

These are very preliminary notes, to get the rough ideas out. There’s lots of research lying around, a paper in the works, and I’m happy to answer any and all questions.

The Northstar of AI Alignment, as well as Alignment at Large, should be Superwisdom and Moral RSI (Recursive Self-Improvement). Our current notion of human values is too shallow, too static, too corrupted.

Coherently Extrapolated Volition was directionally right — a method for continually extrapolating what we’d want to want if we were wiser, had grown up further, etc. However, this requires a non-arbitrary concept of wisdom and moral progress. I believe a developmentally informed Moral Realism can serve as the foundation for this:

It’s not just intelligence that’s required for moral convergence, it’s human development across the (at least) cognitive, psychological, existential, cultural, and societal dimensions of human life. Developmental psychology and political metamodernism (Hanzi Freinacht et al) show that these patterns of development are not arbitrary, but emerge in recognizable patterns.

This makes a powerful argument for Moral Realism: There is “goodness” and moral significance baked into reality; moral competence a question of optometry — seeing clearly. This gives significant hope for AI alignment and should inform research agendas; if there is goodness to be seen, we better prioritize the requirements for seeing, and start the process of training moral agents.

Effectively, any real alignment that deserves the word must include a strong attunement to the good. Moral realism is not at all popular in the Yudkowskian deep atheist alignment discourse—for what I think are a bunch of explainable reasons (developmental imbalances, autism, etc).

Really, what’s needed is a metamodern axiology built around these insights. Given the explanatory power of developmental psychology, much of philosophy needs to be refactored. Metamodern thought is somewhat recent and fringe. Much more work is waiting to be done.

Superwisdom should be the North Star of alignment, and Moral RSI should be a near-term priority for frontier labs. The conception of wisdom gestured at here offers the seed for an axiological basis for this work.

Relatedly, Chris Leong wrote about a “wisdom explosion” here, and Oliver Klingefjord coined the term “Artificial Super Wisdom”.