There’s a fairly clear venn diagram between alignment research and capabilities research.
This appears to be the crux of our disagreement. I do not think the venn diagram is clear at all. But if I had to guess, I think there is a large overlap between “make an AI that doesn’t spew out racist garbage” and “make an AI that doesn’t murder us all”.
This appears to be the crux of our disagreement. I do not think the venn diagram is clear at all. But if I had to guess, I think there is a large overlap between “make an AI that doesn’t spew out racist garbage” and “make an AI that doesn’t murder us all”.