Sure. I don’t think you can fit the entire problem of AI alignment within CS, but I think the time is somewhat ripe for people to get more grants for better interpretability, and for progressively more ambitious attempts to make general-ish AI (right now language models because that’s where the cheap data about humans is, but we might also imagine near-future AI that fuses language models with limited action spaces amenable to RL [MineRL season 4?]).
Sure. I don’t think you can fit the entire problem of AI alignment within CS, but I think the time is somewhat ripe for people to get more grants for better interpretability, and for progressively more ambitious attempts to make general-ish AI (right now language models because that’s where the cheap data about humans is, but we might also imagine near-future AI that fuses language models with limited action spaces amenable to RL [MineRL season 4?]).