The model of environment values observations and behaviors, not statements about “uncomputability” and such. No observation should be left out, declared impossible. If you, as a human, decide to trust in something you label “halting oracle”, that’s your decision, and this is a decision you’d want any trusted AI to carry through as well.
I suspect that the roots of this confusion are something not unlike mind projection fallacy, with magical properties attributed to models, but I’m not competent to discuss domain-specific aspects of this question.
The model of environment values observations and behaviors, not statements about “uncomputability” and such. No observation should be left out, declared impossible. If you, as a human, decide to trust in something you label “halting oracle”, that’s your decision, and this is a decision you’d want any trusted AI to carry through as well.
I suspect that the roots of this confusion are something not unlike mind projection fallacy, with magical properties attributed to models, but I’m not competent to discuss domain-specific aspects of this question.