Speaking as the only person so far who has chosen Option 1, I definitely don’t downvote every post that I don’t get value out of. “How much I get out of it” and “how much karma I think it should have” are two very different things. If I read something and don’t get value out of it, usually I’m agnostic as to how much other people get out of it and don’t vote. Downvoting it requires me being confident that public consensus is wrong, which is rare (but happens occasionally).
I guess this means that my algorithm isn’t ‘always assess desired karma target and move in that direction’ but that plus a whole bunch of epistemic humility. I’m also not claiming to know exactly what I’m doing, though I think it’s right to say that it’s primarily about the karma target.
I recently strong-downvoted a post that I would have weak-upvoted if it had been at a lower karma
Was that post good or bad? It sounded to me like you thought the post had value, just not as much as was currently showing. If you downvoted a post you thought had positive value (you were confident that its current karma value was too high?), why not downvote one that you don’t see any value in?
If being agnostic is a cause for not voting at all, a 50% great, 50% agnostic post would get a higher score than a 50% great, 50% slightly good post as the slightly good experiences would downvote and the agnostics wouldn’t.
I think my main concern with the “vote to try to give posts/comments the total karma they should have” rule is that I can’t see a way to operationalise it which doesn’t suffer from worse problems than the simple “I want to see more/less of this” rule.
Was that post good or bad? It sounded to me like you thought the post had value, just not as much as was currently showing. If you downvoted a post you thought had positive value (you were confident that its current karma value was too high?)
It was good—the if part of your quote is pretty accurate.
why not downvote one that you don’t see any value in?
Downvote a different post of the same author because I didn’t like that one? That doesn’t sound like a good idea.
Downvote a different post of the same author because I didn’t like that one? That doesn’t sound like a good idea.
No, I mean why wouldn’t you downvote a hypothetical post that you are agnostic about?
Imagine there are two posts, both have 50 karma.
You read one and feel confident that it is net positive but that 50 is too high.
You read the other and it is not net positive for you—you just have a meh reaction to it.
It seems very odd to me that one would downvote the former but not the latter. The net effect is to encourage people to read/write a post that is more likely to provide a meh reaction than be net positive.
No, I mean why wouldn’t you downvote a hypothetical post that you are agnostic about?
Because such a post has more probability mass on the “it’s really good” hypothesis. If I’m confident that a post is only slightly good, well then I’m also confident that it’s not very good.
I think my main concern with the “vote to try to give posts/comments the total karma they should have” rule is that I can’t see a way to operationalise it which doesn’t suffer from worse problems than the simple “I want to see more/less of this” rule.
The fundamental problem is that we’re trying to map a multidimensional thing into a single dimension. Whenever you do this you end up throwing out some information and you have to do the best you can.
As described by jimmy, with the “I want to see more/less of this” rule you lose some information on magnitude of like/dislike with the “I want to see more/less of this” rule. This is somewhat mitigated by having weak and strong votes, plus the dither factor jimmy describes (which I think for me is quite significant) so overall I’m not hugely worried about this.
(You can also get some of this information back if you’re really interested by comparing total number of votes to score although this is less obvious)
I’m not sure how a “how much total karma should this post have” rule even works in practice but a couple of options:
How much karma a post has needs to link to post value / correct amount of reward to the author.
If I judge this according to how much value I personally got out of it then the great-great grandparent comment applies and 50% awesome, 50% meh posts get 0 karma—a worse result than with the “I want to see more/less of this” rule, with all of the information from the 50% of people who found it awesome disappearing.
If instead I am trying to judge how much value I think the average LWer would get out of it then I think this gets really hard to assess. As an example, the recent 10 fun questions results showed that people weren’t very good at guessing whether others believed the Civilisational Inadequacy thesis more or less than they themselves did. Here you lose some information on people’s actual opinions in favour of information on what other people think their opinion might be, adding significant noise to the result.
Whichever option you chose you probably end up throwing out information on how many people got value from the post. You can try to get around this by each person estimating how many others would find it useful but I think this just adds more noise to the result.
You can try to make the rule some combination of rules (as it seems most people do) but then to me it seems like interpreting karma scores becomes really difficult. We also run into the problem of how much weighting to give to each sub-rule and if people give different weightings then you get a discrepancy in how effective each person’s opinion is.
I’m interested if someone can explain another way that a “how much total karma should this post have” rule would work in practice which doesn’t run into such problems.
(note: none of this is representing LW team opinion or anything)
I think karma is sufficiently noisy that trying to get some kind of “real” information out of it is already pretty intractable. People definitely vary in how they use it – some people vote liberally, some people vote rarely, some people use it as a “yay/boo” thing for things they disagree with, some people use it in some principled fashion. I’m betting that almost nobody consciously considers whether to upvote every single comment they read.
(Note that we deliberately don’t show number of upvotes/downvotes on a post (mostly for a reason unrelated to this, but it means that “how many people downvoted this” is not public information that is supposed to communicate anything))
In my mind, the algorithm I’m implementing is still the “do you want to see more/less of this?”, I just also have a vague term for “more/less relative to what?”. If I see a bunch of people have upvoted a thing that I want to see more of, enough that I feel like the system has already given it enough reinforcement, I feel less inclined to upvote it myself, because the system already was giving it the amount of “more of this” that I wanted.
This just seems like the usual system with a bit more information, rather than less, to me.
Karma has a few concrete effects on the site: how long posts stay on the home page, whether comments get collapsed, and how comments and sometimes posts get sorted on a page. I think is basically fine to use voting to deliberately manipulate those things, based on your judgment of how relatively important a post/comment is. That essentially is the operationalization of what it means to want “more or less of this”, and it seems super reasonable to have that be encouraged rather than treated as an abuse of the system (otherwise, you have a system that relies on people avoiding being strategic, which is an asshole filter)
Downvoting a comment because it’s getting sorted to the top of a thread, when you think another comment is relatively more important, seems fine to me.
(This is in addition to “karma has some vague psychological effect on how high status / community-endorsed a post feels”, which I think is also fine to vote on as another knob on the “less/more of this” thing)
Some notes on my own usage:
I very rarely downvote things because they are “too high.” When I do that, it’s in particular cases where I honestly think it was important that something was said, but there were some aspects of it that I definitely don’t want too much of on the site. The prototypical example is a post that makes some good points, but also some bad points (or uses bad arguments). If it were to get 100+ karma, I’d feel pretty bad about a high profile post making bad arguments. But, it still contributed enough value that I’d also feel bad if it didn’t get at least some reinforcement.
Usually, I make the decision once based how many people have already voted on a thing, and it’s just a matter of “not bothering to upvote.” If someone makes a clever joke, and it already has a few upvotes, I usually don’t upvote it further. I like having clever jokes on LessWrong, but I think it’s somewhat bad if they end up getting more karma than the more substantive comments.
I sometimes withdraw upvotes (on, say, a clever joke that ended up getting 70 karma, where when I first voted on it it only had 5). I do this more commonly than downvoting things that are “too high.” Obviously I don’t do this reliably, but I also wasn’t reliably checking each comment and deciding how hard to upvote it and I don’t expect most people were either. This doesn’t seem like it adds any noise to the system to me beyond what was there already, and meanwhile I think sends a more accurate signal of “Rays preferences about what there should be more/less-of” than not doing that when I happen to notice.
I think the asshole filter is a good point and to be honest its possibly enough to get me to change my mind about this subject. There should be some mitigation in the karma weighting system but even long term members might be assholes.
Does it prove too much? Should the current karma, then, actually be the main consideration in deciding how to vote? Few people on the site seem willing to bite that bullet. Should I almost always use my strong votes as if I don’t then an asshole might do that and therefore have an oversized effect?
Count me confused.
On the other points I won’t go through point by point. The main thing I think is that what you’re describing is in conflict with the explicit phrasing of the reasons for voting. Compare:
What should my votes mean?
We encourage people to vote such that upvote means “I want to see more of this” and downvote means “I want to see less of this.”
with
What should karma indicate?
The karma on a post is intended to indicate whether, and by how much, members of the site would like to see more of the posts/comments in question. We encourage people to vote accordingly.
The former is the FAQ but I think the latter is what you’re describing. If this is the case then I think this ends up being an asshole filter in itself and the phrasing in the FAQ should be corrected.
(I realise as Ray qua user this has nothing to do with you but if you can pass this along to Ray qua admin that would be great!)
I actually think it’s fairly bad that Strong Upvotes aren’t asshole-filter-proof, and in this case I bite the bullet in the direction of “we should limit the power of strong upvotes somehow so they can’t be abused.” (I’ve thought this for awhile, it just hasn’t been top priority, and/or the team couldn’t come up with an improvement that seemed better to everyone)
That said, I think you did just remind me that there a ton of vulnerabilities in the karma system that absolutely rely on people not abusing them most of the time, and yeah I just actually retract that part of my argument. I do think we should eventually someday have a karma system that’s more resilient, and the only reason it’s not a higher priority is that in fact people are mostly good people, and the system just actually mostly works, so it’s not as high priority as other site changes.
But, I do still stand by “manipulating the position/weighting/visibility of posts is basically what “see more / less of this” is actually supposed to mean, and is basically in the spirit of it.
We encourage people to vote such that upvote means “I want to see more of this” and downvote means “I want to see less of this.”
I basically always interpreted this to mean “I want to see more/less of this, and among the things that factor into what I want to see more/less of are subtle things about site norms that impact other people.”
I realize it’s ambiguously worded.
My overall biggest crux here is “Karma is so far away from being a robust system that means concrete things that you just shouldn’t worry too much about what exactly it means. You should know that other people are using it differently from you [for any value of ‘you’]. It’s a vague, kludgy approximation that seems to mostly output reasonable things, and that’s basically fine for now.”
The points I was trying to make were (kinda scattered across the comments here!):
1. It is advantageous if people have a shared understanding of the system
2. Voting your own belief actually should work pretty well
3. There is a written norm in favour of voting your own belief
I think we disagree on all 3 to some extent, at least in how important they are. I think if we lose the disagreement on number 3 then disagreements on 1&2 are less important.
I’m ok with a norm of voting based somewhat on target karma (making it overly strong an effect I think would be detrimental), especially as this is now common knowledge and seems to be most people’s preference.
This whole thing has resolved some of my confusion as to why karma scores end up the way they do.
I want to note that the I see the “vote towards the ideal karma” as completely compatible with “vote your belief.”
I think there are two fairly different questions here:
Should your vote include your beliefs about how much you want other people to see a given post, or what you think is best for others?
Should you vote based on your ideal total-karma for a post
It so happens I think we might disagree about both of them (and disagree about what the best interpretation of the current rules are about them). But, those are quite different questions, and you can do the second based entirely on your own preferences/beliefs.
When a post is at 50, I can think that is a bit too high just from my general sense of what I want to see more of on the site. And it’s be throwing away information about my own beliefs to not give me the fine-gradation of “I want to see these posts on the site about as often as I would if they got 50 karma, not the amount that I would if they got 200 karma.”
When a post is at 50, I can think that is a bit too high just from my general sense of what I want to see more of on the site. And it’s be throwing away information about my own beliefs to not give me the fine-gradation of “I want to see these posts on the site about as often as I would if they got 50 karma, not the amount that I would if they got 200 karma.”
This is true when the equilibrium position of the karma system is set to Total Karma Voting.
I think that Blind voting would move the karma system to a new equilibrium. I’m not convinced we should do so as I think it would be a fairly unstable equilibrium but I think it would work if everyone did it and would allow for fine grained expressions of your belief.
The equilibrium I envisage would be that the current amount of something that LW has is taken into account when people blind vote their opinion.
As an example, I think the reason that joke comments can get fairly high karma is that they’re rare. If more people start writing joke comments as a result then that’s fine for as long as people are upvoting.
At some point the people who value the jokes least stop upvoting them or start downvoting them. This continues until the reward experienced by the jokers roughly matches the effort taken or some other balancing factor.
In the case of low positive value posts, some people have a higher threshold for what they will give an upvote for and the more low positive value posts there are the fewer people will upvote them.
(I think its important to note here that we are not really that homogenous in our opinions and weightings of different sources of value. Alot of the worries about Blind voting seem to assume that we’re all going to vote the same way about the same posts which I think is highly unrealistic. There also seems to be the assumption that everything fractionally above 0 value will get an upvote which again seems unrealistic. Frankly I think that anyone who can write a post which is good enough that it persuades 100 different people with different standards to click the upvote button then they deserve to get 150 karma!)
The key then is that in order to get an oversized reward for the amount of effort put in, you have to do better than average at providing value.
In Blind Voting, accounting-for-how-much-of-a-certain-thing-there-currently-is-on-LW is doing the same thing as considering-what-message-the-total-karma-sends does with Total Karma Voting. The former seems to have a lag in the message getting out but I think when you’re in a rough equilibrium the lag is relatively short.
So this brings me onto what I think the main cost of Total karma voting is. If an author looks at a post which has 25 karma from 10 votes, what does it mean? Roughly speaking it means that it was considered about as valuable as another 25 karma post. The 10 votes tells the author how efficient the karma market was for the post and possibly gives limited information on how varied the opinions were.
With Blind voting the author sees that and knows that 10 people had an opinion that this post was wanted more or less and that their average strength of opinion was 2.5 karma points in favour. This probably consists of something like 3 people who want alot more like it and 7 people who want a little more like it (or possibly some who wish there was less like it or were just yay/booing).
I agree that karma is a kludge and the true meaning isn’t necessarily clear but with Blind voting it seems importantly less of a kludge and some extra information can be extracted.
I want to note that the I see the “vote towards the ideal karma” as completely compatible with “vote your belief.”
Agreed. I was looking for a shorthand way of referring to the different voting policies but am yet to find one which is satisfactory—you’ve (rightly) shot down a couple of my ideas! Total Karma voting seems fine for one policy, maybe direct opinion voting for the other? If you shoot that one down too you can come up with your own!
I think “blind voting” captures the distinction better – the key difference is whether you’re supposed to look at or model the outcome.
Btw another reason I think “take total karma into account” is important is because of how big a slap downvotes feel like. Blind voting both means that “mildly good comments” will get like 80 karma, but also means that mildly bad comments will get like −80 karma, which would make the site feel very punishing.
I do think that it would be very bad if this happened. However I don’t think this is likely. Quoting my other comment:
I think its important to note here that we are not really that homogenous in our opinions and weightings of different sources of value. Alot of the worries about Blind voting seem to assume that we’re all going to vote the same way about the same posts which I think is highly unrealistic. There also seems to be the assumption that everything fractionally above 0 value will get an upvote which again seems unrealistic.
This seems even more true for downvotes—I think people realise that downvotes feel extra bad and only use them sparingly. For instance, I only really downvote when I think something has been a definite breaking of a conversational norm or if someone is doubling down on an argument which has been convincingly refuted.
I think a spread of opinions on what constitutes a downvote (and a general feeling that comments get less votes in general) would make the −80 only happen to super egregiously bad comments.
It seems the definition of abuse of Strong Upvotes is about a person using them all the time. You could say that if a person uses Strong votes more then X% of the time they vote the impact of their Strong votes gets reduced.
Adjusting in the other direction seems useful as well. If someone Strong Upvotes ten times less frequently than average I would want to see their strong upvote as worth somewhat more.
There’s a hypothetical direction we could go where voting-weight is determined based on your vote frequency. The main disadvantage of this is that it becomes a lot harder to predict and conceptualize what voting does.
One hesitation habryka had about penalizing excessive strong downvotes is people would end up trying to conserve them as a resource, like a videogame where you end up hoarding all your potions because you “might need them some day” and never actually use them.
Speaking as the only person so far who has chosen Option 1, I definitely don’t downvote every post that I don’t get value out of. “How much I get out of it” and “how much karma I think it should have” are two very different things. If I read something and don’t get value out of it, usually I’m agnostic as to how much other people get out of it and don’t vote. Downvoting it requires me being confident that public consensus is wrong, which is rare (but happens occasionally).
I guess this means that my algorithm isn’t ‘always assess desired karma target and move in that direction’ but that plus a whole bunch of epistemic humility. I’m also not claiming to know exactly what I’m doing, though I think it’s right to say that it’s primarily about the karma target.
Hmm, interesting—I’m now slightly confused by:
Was that post good or bad? It sounded to me like you thought the post had value, just not as much as was currently showing. If you downvoted a post you thought had positive value (you were confident that its current karma value was too high?), why not downvote one that you don’t see any value in?
If being agnostic is a cause for not voting at all, a 50% great, 50% agnostic post would get a higher score than a 50% great, 50% slightly good post as the slightly good experiences would downvote and the agnostics wouldn’t.
I think my main concern with the “vote to try to give posts/comments the total karma they should have” rule is that I can’t see a way to operationalise it which doesn’t suffer from worse problems than the simple “I want to see more/less of this” rule.
It was good—the if part of your quote is pretty accurate.
Downvote a different post of the same author because I didn’t like that one? That doesn’t sound like a good idea.
No, I mean why wouldn’t you downvote a hypothetical post that you are agnostic about?
Imagine there are two posts, both have 50 karma.
You read one and feel confident that it is net positive but that 50 is too high.
You read the other and it is not net positive for you—you just have a meh reaction to it.
It seems very odd to me that one would downvote the former but not the latter. The net effect is to encourage people to read/write a post that is more likely to provide a meh reaction than be net positive.
Because such a post has more probability mass on the “it’s really good” hypothesis. If I’m confident that a post is only slightly good, well then I’m also confident that it’s not very good.
What’s the worse problems you’re seeing?
The fundamental problem is that we’re trying to map a multidimensional thing into a single dimension. Whenever you do this you end up throwing out some information and you have to do the best you can.
As described by jimmy, with the “I want to see more/less of this” rule you lose some information on magnitude of like/dislike with the “I want to see more/less of this” rule. This is somewhat mitigated by having weak and strong votes, plus the dither factor jimmy describes (which I think for me is quite significant) so overall I’m not hugely worried about this.
(You can also get some of this information back if you’re really interested by comparing total number of votes to score although this is less obvious)
I’m not sure how a “how much total karma should this post have” rule even works in practice but a couple of options:
How much karma a post has needs to link to post value / correct amount of reward to the author.
If I judge this according to how much value I personally got out of it then the great-great grandparent comment applies and 50% awesome, 50% meh posts get 0 karma—a worse result than with the “I want to see more/less of this” rule, with all of the information from the 50% of people who found it awesome disappearing.
If instead I am trying to judge how much value I think the average LWer would get out of it then I think this gets really hard to assess. As an example, the recent 10 fun questions results showed that people weren’t very good at guessing whether others believed the Civilisational Inadequacy thesis more or less than they themselves did. Here you lose some information on people’s actual opinions in favour of information on what other people think their opinion might be, adding significant noise to the result.
Whichever option you chose you probably end up throwing out information on how many people got value from the post. You can try to get around this by each person estimating how many others would find it useful but I think this just adds more noise to the result.
You can try to make the rule some combination of rules (as it seems most people do) but then to me it seems like interpreting karma scores becomes really difficult. We also run into the problem of how much weighting to give to each sub-rule and if people give different weightings then you get a discrepancy in how effective each person’s opinion is.
I’m interested if someone can explain another way that a “how much total karma should this post have” rule would work in practice which doesn’t run into such problems.
(note: none of this is representing LW team opinion or anything)
I think karma is sufficiently noisy that trying to get some kind of “real” information out of it is already pretty intractable. People definitely vary in how they use it – some people vote liberally, some people vote rarely, some people use it as a “yay/boo” thing for things they disagree with, some people use it in some principled fashion. I’m betting that almost nobody consciously considers whether to upvote every single comment they read.
(Note that we deliberately don’t show number of upvotes/downvotes on a post (mostly for a reason unrelated to this, but it means that “how many people downvoted this” is not public information that is supposed to communicate anything))
In my mind, the algorithm I’m implementing is still the “do you want to see more/less of this?”, I just also have a vague term for “more/less relative to what?”. If I see a bunch of people have upvoted a thing that I want to see more of, enough that I feel like the system has already given it enough reinforcement, I feel less inclined to upvote it myself, because the system already was giving it the amount of “more of this” that I wanted.
This just seems like the usual system with a bit more information, rather than less, to me.
Karma has a few concrete effects on the site: how long posts stay on the home page, whether comments get collapsed, and how comments and sometimes posts get sorted on a page. I think is basically fine to use voting to deliberately manipulate those things, based on your judgment of how relatively important a post/comment is. That essentially is the operationalization of what it means to want “more or less of this”, and it seems super reasonable to have that be encouraged rather than treated as an abuse of the system (otherwise, you have a system that relies on people avoiding being strategic, which is an asshole filter)
Downvoting a comment because it’s getting sorted to the top of a thread, when you think another comment is relatively more important, seems fine to me.
(This is in addition to “karma has some vague psychological effect on how high status / community-endorsed a post feels”, which I think is also fine to vote on as another knob on the “less/more of this” thing)
Some notes on my own usage:
I very rarely downvote things because they are “too high.” When I do that, it’s in particular cases where I honestly think it was important that something was said, but there were some aspects of it that I definitely don’t want too much of on the site. The prototypical example is a post that makes some good points, but also some bad points (or uses bad arguments). If it were to get 100+ karma, I’d feel pretty bad about a high profile post making bad arguments. But, it still contributed enough value that I’d also feel bad if it didn’t get at least some reinforcement.
Usually, I make the decision once based how many people have already voted on a thing, and it’s just a matter of “not bothering to upvote.” If someone makes a clever joke, and it already has a few upvotes, I usually don’t upvote it further. I like having clever jokes on LessWrong, but I think it’s somewhat bad if they end up getting more karma than the more substantive comments.
I sometimes withdraw upvotes (on, say, a clever joke that ended up getting 70 karma, where when I first voted on it it only had 5). I do this more commonly than downvoting things that are “too high.” Obviously I don’t do this reliably, but I also wasn’t reliably checking each comment and deciding how hard to upvote it and I don’t expect most people were either. This doesn’t seem like it adds any noise to the system to me beyond what was there already, and meanwhile I think sends a more accurate signal of “Rays preferences about what there should be more/less-of” than not doing that when I happen to notice.
I think the asshole filter is a good point and to be honest its possibly enough to get me to change my mind about this subject. There should be some mitigation in the karma weighting system but even long term members might be assholes.
Does it prove too much? Should the current karma, then, actually be the main consideration in deciding how to vote? Few people on the site seem willing to bite that bullet. Should I almost always use my strong votes as if I don’t then an asshole might do that and therefore have an oversized effect?
Count me confused.
On the other points I won’t go through point by point. The main thing I think is that what you’re describing is in conflict with the explicit phrasing of the reasons for voting. Compare:
with
The former is the FAQ but I think the latter is what you’re describing. If this is the case then I think this ends up being an asshole filter in itself and the phrasing in the FAQ should be corrected.
(I realise as Ray qua user this has nothing to do with you but if you can pass this along to Ray qua admin that would be great!)
I actually think it’s fairly bad that Strong Upvotes aren’t asshole-filter-proof, and in this case I bite the bullet in the direction of “we should limit the power of strong upvotes somehow so they can’t be abused.” (I’ve thought this for awhile, it just hasn’t been top priority, and/or the team couldn’t come up with an improvement that seemed better to everyone)
That said, I think you did just remind me that there a ton of vulnerabilities in the karma system that absolutely rely on people not abusing them most of the time, and yeah I just actually retract that part of my argument. I do think we should eventually someday have a karma system that’s more resilient, and the only reason it’s not a higher priority is that in fact people are mostly good people, and the system just actually mostly works, so it’s not as high priority as other site changes.
But, I do still stand by “manipulating the position/weighting/visibility of posts is basically what “see more / less of this” is actually supposed to mean, and is basically in the spirit of it.
I basically always interpreted this to mean “I want to see more/less of this, and among the things that factor into what I want to see more/less of are subtle things about site norms that impact other people.”
I realize it’s ambiguously worded.
My overall biggest crux here is “Karma is so far away from being a robust system that means concrete things that you just shouldn’t worry too much about what exactly it means. You should know that other people are using it differently from you [for any value of ‘you’]. It’s a vague, kludgy approximation that seems to mostly output reasonable things, and that’s basically fine for now.”
Ok, I think I actually agree with your crux.
The points I was trying to make were (kinda scattered across the comments here!):
1. It is advantageous if people have a shared understanding of the system
2. Voting your own belief actually should work pretty well
3. There is a written norm in favour of voting your own belief
I think we disagree on all 3 to some extent, at least in how important they are. I think if we lose the disagreement on number 3 then disagreements on 1&2 are less important.
I’m ok with a norm of voting based somewhat on target karma (making it overly strong an effect I think would be detrimental), especially as this is now common knowledge and seems to be most people’s preference.
This whole thing has resolved some of my confusion as to why karma scores end up the way they do.
I want to note that the I see the “vote towards the ideal karma” as completely compatible with “vote your belief.”
I think there are two fairly different questions here:
Should your vote include your beliefs about how much you want other people to see a given post, or what you think is best for others?
Should you vote based on your ideal total-karma for a post
It so happens I think we might disagree about both of them (and disagree about what the best interpretation of the current rules are about them). But, those are quite different questions, and you can do the second based entirely on your own preferences/beliefs.
When a post is at 50, I can think that is a bit too high just from my general sense of what I want to see more of on the site. And it’s be throwing away information about my own beliefs to not give me the fine-gradation of “I want to see these posts on the site about as often as I would if they got 50 karma, not the amount that I would if they got 200 karma.”
This is true when the equilibrium position of the karma system is set to Total Karma Voting.
I think that Blind voting would move the karma system to a new equilibrium. I’m not convinced we should do so as I think it would be a fairly unstable equilibrium but I think it would work if everyone did it and would allow for fine grained expressions of your belief.
The equilibrium I envisage would be that the current amount of something that LW has is taken into account when people blind vote their opinion.
As an example, I think the reason that joke comments can get fairly high karma is that they’re rare. If more people start writing joke comments as a result then that’s fine for as long as people are upvoting.
At some point the people who value the jokes least stop upvoting them or start downvoting them. This continues until the reward experienced by the jokers roughly matches the effort taken or some other balancing factor.
In the case of low positive value posts, some people have a higher threshold for what they will give an upvote for and the more low positive value posts there are the fewer people will upvote them.
(I think its important to note here that we are not really that homogenous in our opinions and weightings of different sources of value. Alot of the worries about Blind voting seem to assume that we’re all going to vote the same way about the same posts which I think is highly unrealistic. There also seems to be the assumption that everything fractionally above 0 value will get an upvote which again seems unrealistic. Frankly I think that anyone who can write a post which is good enough that it persuades 100 different people with different standards to click the upvote button then they deserve to get 150 karma!)
The key then is that in order to get an oversized reward for the amount of effort put in, you have to do better than average at providing value.
In Blind Voting, accounting-for-how-much-of-a-certain-thing-there-currently-is-on-LW is doing the same thing as considering-what-message-the-total-karma-sends does with Total Karma Voting. The former seems to have a lag in the message getting out but I think when you’re in a rough equilibrium the lag is relatively short.
So this brings me onto what I think the main cost of Total karma voting is. If an author looks at a post which has 25 karma from 10 votes, what does it mean? Roughly speaking it means that it was considered about as valuable as another 25 karma post. The 10 votes tells the author how efficient the karma market was for the post and possibly gives limited information on how varied the opinions were.
With Blind voting the author sees that and knows that 10 people had an opinion that this post was wanted more or less and that their average strength of opinion was 2.5 karma points in favour. This probably consists of something like 3 people who want alot more like it and 7 people who want a little more like it (or possibly some who wish there was less like it or were just yay/booing).
I agree that karma is a kludge and the true meaning isn’t necessarily clear but with Blind voting it seems importantly less of a kludge and some extra information can be extracted.
Agreed. I was looking for a shorthand way of referring to the different voting policies but am yet to find one which is satisfactory—you’ve (rightly) shot down a couple of my ideas! Total Karma voting seems fine for one policy, maybe direct opinion voting for the other? If you shoot that one down too you can come up with your own!
I think “blind voting” captures the distinction better – the key difference is whether you’re supposed to look at or model the outcome.
Btw another reason I think “take total karma into account” is important is because of how big a slap downvotes feel like. Blind voting both means that “mildly good comments” will get like 80 karma, but also means that mildly bad comments will get like −80 karma, which would make the site feel very punishing.
I do think that it would be very bad if this happened. However I don’t think this is likely. Quoting my other comment:
This seems even more true for downvotes—I think people realise that downvotes feel extra bad and only use them sparingly. For instance, I only really downvote when I think something has been a definite breaking of a conversational norm or if someone is doubling down on an argument which has been convincingly refuted.
I think a spread of opinions on what constitutes a downvote (and a general feeling that comments get less votes in general) would make the −80 only happen to super egregiously bad comments.
It seems the definition of abuse of Strong Upvotes is about a person using them all the time. You could say that if a person uses Strong votes more then X% of the time they vote the impact of their Strong votes gets reduced.
Adjusting in the other direction seems useful as well. If someone Strong Upvotes ten times less frequently than average I would want to see their strong upvote as worth somewhat more.
There’s a hypothetical direction we could go where voting-weight is determined based on your vote frequency. The main disadvantage of this is that it becomes a lot harder to predict and conceptualize what voting does.
One hesitation habryka had about penalizing excessive strong downvotes is people would end up trying to conserve them as a resource, like a videogame where you end up hoarding all your potions because you “might need them some day” and never actually use them.
Also note:
I think this sort of problem is still there if you’re not trying to “move posts towards the ‘correct’ karma.”
“I want to see less/more of this” still depends on “how good do I think this is for LW as a whole?”