kind of LLM computing the explanation in advance and weighting the score with this precomputed score
I think this would work fine for things that don’t essentially require step-by-step reasoning (“detailed reasoning” might still be accessing this capability), which is why the order from the post is not as bad as writing down the bottom line before the argument. When you ask for the argument, it’s being anticipated, as much as that’s possible to do without seeing its prefix explicitly. But this should be a problem when the outcome of the argument can’t be anticipated.
I think this would work fine for things that don’t essentially require step-by-step reasoning (“detailed reasoning” might still be accessing this capability), which is why the order from the post is not as bad as writing down the bottom line before the argument. When you ask for the argument, it’s being anticipated, as much as that’s possible to do without seeing its prefix explicitly. But this should be a problem when the outcome of the argument can’t be anticipated.
Good point, i should add this in limitation and futur directions. Do you have an example in mind?