It’s this class of patterns that frequently recur as a crucial considerations in contexts re optimization, and I’ve been making too many shoddy comments about it. (Recent1[1], Recent2.) Somebody who can write ought to unify the many aspects of it give it a public name so it can enter discourse or something.
In the context of conjunctive search/optimization
The problem of fully updated deference also assumes a concave option-set. The concavity is proportional to the number of independent-ish factors in your utility function. My idionym (in my notes) for when you’re incentivized to optimize for a subset of those factors (rather than a compromise), is instrumental drive for monotely (IDMT), and it’s one aspect of Goodhart.
It’s one reason why proxy-metrics/policies often “break down under optimization pressure”.
When you decompose the proxy into its subfunctions, you often tend to find that optimizing for a subset of them is more effective.
(Another reason is just that the metric has lots of confounders which didn’t map to real value anyway; but that’s a separate matter from conjunctive optimization over multiple dimensions of value.)
You can sorta think of stuff like the Weber-Fechner Law (incl scope-insensitivity) as (among other things) an “alignment mechanism” in the brain: it enforces diminishing returns to stimuli-specificity, and this reduces your tendency to wirehead on a subset of the brain’s reward-proxies.
Watch the blue twirly thing until you forget how bored you are by this essay, then continue.
In the context of how intensity of something is inversely proportional to the number of options
Humans differentiate into specific social roles because .
If you differentiate into a less crowded category, you have fewer competitors for the type of social status associated with that category. Specializing toward a specific role makes you more likely to be top-scoring in a specific category.
Political candidates have some incentive to be extreme/polarizing.
If you try to please everybody, you spread out your appeal so it’s below everybody’s threshold, and you’re not getting anybody’s votes.
Your marginal likelihood of tipping the election is proportional to how close the candidate is to the threshold, so everybody has an incentive to vote for ~Schelling-points in what people expect other people to vote for. This has the effect of concentrating votes over the two most salient options.
You tend to feel demotivated when you have too many tasks to choose from on your todo-list.
Motivational salience is normalized across all conscious options[2], so you’d have more absolute salience for your top option if you had fewer options.
I tend to say a lot of wrong stuff, so do take my utterances with grains of salt. I don’t optimize for being safe to defer to, but it doesn’t matter if I say a bunch of wrong stuff if some of the patterns can work as gears in your own models. Screens off concerns about deference or how right or wrong I am.
I rly like the framing of concave vs convex option-set btw!
Lizka has a post abt concave option-set in forum-post writing! From my comment on it:
As you allude to by the exponential decay of the green dots in your last graph, there are exponential costs to compromising what you are optimizing for in order to appeal to a wider variety of interests. On the flip-side, how usefwl to a subgroup you can expect to be is exponentially proportional to how purely you optimize for that particular subset of people (depending on how independent the optimization criteria are). This strategy is also known as “horizontal segmentation”.
The benefits of segmentation ought to be compared against what is plausibly an exponential decay in the number of people who fit a marginally smaller subset of optimization criteria. So it’s not obvious in general whether you should on the margin try to aim more purely for a subset, or aim for broader appeal.
Normalization is an explicit step in taking the population vector of an ensemble involved in some computation. So if you imagine the vector for the ensemble(s) involved in choosing what to do next, and take the projection of that vector onto directions representing each option, the intensity of your motivation for any option is proportional to the length of that projection relative to the length of all other projections. (Although here I’m just extrapolating the formula to visualize its consequences—this step isn’t explicitly supported by anything I’ve read. E.g. I doubt cosine similarity is appropriate for it.)
Another aspect of costs of compromise is: How bad is it for altruists to have to compromise their cognitive search between [what you believe you can explain to funders] vs [what you believe is effective]? Re my recent harumph about the fact that John Wentworth must justify his research to get paid. Like what? After all this time, anybody doubts him? The insistence that he explain himself is surely more for show now, as it demonstrates the funders are doing their jobs “seriously”.
This relates to costs of compromise!
It’s this class of patterns that frequently recur as a crucial considerations in contexts re optimization, and I’ve been making too many shoddy comments about it. (Recent1[1], Recent2.) Somebody who can write ought to unify the many aspects of it give it a public name so it can enter discourse or something.
In the context of conjunctive search/optimization
The problem of fully updated deference also assumes a concave option-set. The concavity is proportional to the number of independent-ish factors in your utility function. My idionym (in my notes) for when you’re incentivized to optimize for a subset of those factors (rather than a compromise), is instrumental drive for monotely (IDMT), and it’s one aspect of Goodhart.
It’s one reason why proxy-metrics/policies often “break down under optimization pressure”.
When you decompose the proxy into its subfunctions, you often tend to find that optimizing for a subset of them is more effective.
(Another reason is just that the metric has lots of confounders which didn’t map to real value anyway; but that’s a separate matter from conjunctive optimization over multiple dimensions of value.)
You can sorta think of stuff like the Weber-Fechner Law (incl scope-insensitivity) as (among other things) an “alignment mechanism” in the brain: it enforces diminishing returns to stimuli-specificity, and this reduces your tendency to wirehead on a subset of the brain’s reward-proxies.
Pareto nonconvexity is annoying
From Wikipedia: Multi-Objective optimization:
In the context of how intensity of something is inversely proportional to the number of options
Humans differentiate into specific social roles because .
If you differentiate into a less crowded category, you have fewer competitors for the type of social status associated with that category. Specializing toward a specific role makes you more likely to be top-scoring in a specific category.
Political candidates have some incentive to be extreme/polarizing.
If you try to please everybody, you spread out your appeal so it’s below everybody’s threshold, and you’re not getting anybody’s votes.
You have a disincentive to vote for third-parties in winner-takes-all elections.
Your marginal likelihood of tipping the election is proportional to how close the candidate is to the threshold, so everybody has an incentive to vote for ~Schelling-points in what people expect other people to vote for. This has the effect of concentrating votes over the two most salient options.
You tend to feel demotivated when you have too many tasks to choose from on your todo-list.
Motivational salience is normalized across all conscious options[2], so you’d have more absolute salience for your top option if you had fewer options.
I tend to say a lot of wrong stuff, so do take my utterances with grains of salt. I don’t optimize for being safe to defer to, but it doesn’t matter if I say a bunch of wrong stuff if some of the patterns can work as gears in your own models. Screens off concerns about deference or how right or wrong I am.
I rly like the framing of concave vs convex option-set btw!
Lizka has a post abt concave option-set in forum-post writing! From my comment on it:
Normalization is an explicit step in taking the population vector of an ensemble involved in some computation. So if you imagine the vector for the ensemble(s) involved in choosing what to do next, and take the projection of that vector onto directions representing each option, the intensity of your motivation for any option is proportional to the length of that projection relative to the length of all other projections. (Although here I’m just extrapolating the formula to visualize its consequences—this step isn’t explicitly supported by anything I’ve read. E.g. I doubt cosine similarity is appropriate for it.)
I love this comment, and have strong-upvoted it.
I especially enjoy that I seem to have basically popularized a frame you’ve already been thinking about.
Another aspect of costs of compromise is: How bad is it for altruists to have to compromise their cognitive search between [what you believe you can explain to funders] vs [what you believe is effective]? Re my recent harumph about the fact that John Wentworth must justify his research to get paid. Like what? After all this time, anybody doubts him? The insistence that he explain himself is surely more for show now, as it demonstrates the funders are doing their jobs “seriously”.