Ah, on the numbers thing, what I meant was only that AFAIK there always exists some formula for which higher output numbers will correspond to things any abitrary agent (at least, all the logically valid and sound ones that I’ve thought of) would prefer.
So even for a hard tier system, there’s a way to compute a number linearly representative of how happy the agent is with worldstates, where at the extreme all lower-tier values flatline into arbitrarily large negatives (or other, more creative / leakproof weighing) whenever they incur infinitesimal risk of opportunity cost towards the higher-tier values.
The reason I’m said this is because it’s often disputed and/or my audience isn’t aware of it, and I often have to prove even the most basic versions of this claim (such as “you can represent a tiered system where as soon as the higher tier is empty, the lower tier is worthless using a relatively simple mathematical formula”) by showing them the actual equations and explaining how it works.
Ah, on the numbers thing, what I meant was only that AFAIK there always exists some formula for which higher output numbers will correspond to things any abitrary agent (at least, all the logically valid and sound ones that I’ve thought of) would prefer.
So even for a hard tier system, there’s a way to compute a number linearly representative of how happy the agent is with worldstates, where at the extreme all lower-tier values flatline into arbitrarily large negatives (or other, more creative / leakproof weighing) whenever they incur infinitesimal risk of opportunity cost towards the higher-tier values.
The reason I’m said this is because it’s often disputed and/or my audience isn’t aware of it, and I often have to prove even the most basic versions of this claim (such as “you can represent a tiered system where as soon as the higher tier is empty, the lower tier is worthless using a relatively simple mathematical formula”) by showing them the actual equations and explaining how it works.