It seems to me that this “counter-counterexample” also applies for complexity – if the translator is able to reuse computation from the predictor, wouldn’t that both reduce the complexity and the time?
The difference is that the complexity savings are bounded by a constant that doesn’t depend on the size of the predictor’s model (namely the complexity of the human’s model). But in the case of computation time, the savings can be as large as “compute used to do inference in the human’s model.” And that scales up with the size of our computers.
The difference is that the complexity savings are bounded by a constant that doesn’t depend on the size of the predictor’s model (namely the complexity of the human’s model). But in the case of computation time, the savings can be as large as “compute used to do inference in the human’s model.” And that scales up with the size of our computers.