I think it depends on which domain you’re delegating in. E.g. physical objects, especially complex systems like an AC unit, are plausibly much harder to validate than a mathematical proof.
In that vein, I wonder if requiring the AI to construct a validation proof would be feasible for alignment delegation? In that case, I’d expect us to find more use and safety from [ETA: delegation of] theoretical work than empirical.
I think it depends on which domain you’re delegating in. E.g. physical objects, especially complex systems like an AC unit, are plausibly much harder to validate than a mathematical proof.
In that vein, I wonder if requiring the AI to construct a validation proof would be feasible for alignment delegation? In that case, I’d expect us to find more use and safety from [ETA: delegation of] theoretical work than empirical.
That seems a lot like Davidad’s alignment research agenda.