suppose we sorted out a verbal specification of an aligned AI and had a candidate FAI coded up—could we then use Debate on the question “does this candidate match the verbal specification?”
I’m less excited about this, and more excited about candidate training processes or candidate paradigms of AI research (for example, solutions to embedded agency). I expect that there will be a large cluster of techniques which produce safe AGIs, we just need to find them—which may be difficult, but hopefully less difficult with Debate involved.
I’m less excited about this, and more excited about candidate training processes or candidate paradigms of AI research (for example, solutions to embedded agency). I expect that there will be a large cluster of techniques which produce safe AGIs, we just need to find them—which may be difficult, but hopefully less difficult with Debate involved.