Before we even start a training run, we should try to have *actually good *abstract arguments about alignment properties of the AI. Interpretability work is easier if you’re just trying to check details relevant to those arguments, rather than trying to figure out the whole AI.
Thanks for the post! I particularly appreciated this point
Thanks for the post! I particularly appreciated this point