I’m focusing on the code in Appendix B.
What happens when self.diamondShard’s assessment of whether some consequences contain diamonds differs from ours? (Assume the agent’s world model is especially good.)
self.diamondShard
The same thing which happens if the assessment isn’t different from ours—the agent is more likely to take that plan, all else equal.
I’m focusing on the code in Appendix B.
What happens when
self.diamondShard
’s assessment of whether some consequences contain diamonds differs from ours? (Assume the agent’s world model is especially good.)The same thing which happens if the assessment isn’t different from ours—the agent is more likely to take that plan, all else equal.