Maybe there’s some way of doing oversight that verifies that it is either a random soup or some super sophisticated hidden attack? And maybe we can rule out the later if the model hasn’t undergone nearly enough training yet?
Maybe there’s some way of doing oversight that verifies that it is either a random soup or some super sophisticated hidden attack? And maybe we can rule out the later if the model hasn’t undergone nearly enough training yet?