I was not thinking about it before reading this comment, but even partial solutions to the problem in this post would probably both advance capabilities and safety. My first impression is that it helps build capability in a way that ensures more alignment, so it might be a net positive for alignment and safety. But that wouldn’t necessarily hold if we also care about the misuse of aligned AI (which we probably should).
I was not thinking about it before reading this comment, but even partial solutions to the problem in this post would probably both advance capabilities and safety. My first impression is that it helps build capability in a way that ensures more alignment, so it might be a net positive for alignment and safety. But that wouldn’t necessarily hold if we also care about the misuse of aligned AI (which we probably should).