We’re trying to address cases where the human isn’t actually able to update on all of D and form a posterior based on that. We’re trying to approximate ‘what the human posterior would be if they had been able to look at all of D’. So to do that, we learn the human prior, and we learn the human likelihood, then have the ML do the computationally-intensive part of looking at all of D and updating based on everything in there.
We’re trying to address cases where the human isn’t actually able to update on all of D and form a posterior based on that. We’re trying to approximate ‘what the human posterior would be if they had been able to look at all of D’. So to do that, we learn the human prior, and we learn the human likelihood, then have the ML do the computationally-intensive part of looking at all of D and updating based on everything in there.
Does that make sense?