Here’s a few more questions about the same strategy:
If I understand correctly, the IG strategy is to learn a joint model for observations and actions pθ(v,a;Z), where v, a, and Z are video, actions, and proposed change to the Bayes net, respectively. Then we do inference using pθ(v,a;Z∗), where Z∗ is optimized for predictive usefulness.
This fails because there’s no easy way to get P(diamond is in the vault) from pθ.
A simple way around this would be to learn pθ(v,a,y;Z) instead, where y=1 if the diamond is in the vault and 0 otherwise.
Is my understanding correct?
If so, I would guess that my simple workaround doesn’t count as a strategy because we can only use this to predict whether the diamond is in the vault (or some other set of questions that must be fixed at training time), as opposed to any question we want an answer to. Is this correct? Is there some other reason this wouldn’t count, or does it in fact count?
Here’s a few more questions about the same strategy:
If I understand correctly, the IG strategy is to learn a joint model for observations and actions pθ(v,a;Z), where v, a, and Z are video, actions, and proposed change to the Bayes net, respectively. Then we do inference using pθ(v,a;Z∗), where Z∗ is optimized for predictive usefulness.
This fails because there’s no easy way to get P(diamond is in the vault) from pθ.
A simple way around this would be to learn pθ(v,a,y;Z) instead, where y=1 if the diamond is in the vault and 0 otherwise.
Is my understanding correct?
If so, I would guess that my simple workaround doesn’t count as a strategy because we can only use this to predict whether the diamond is in the vault (or some other set of questions that must be fixed at training time), as opposed to any question we want an answer to. Is this correct? Is there some other reason this wouldn’t count, or does it in fact count?