Rationality Quotes September 2012
Here’s the new thread for posting quotes, with the usual rules:
Please post all quotes separately, so that they can be voted up/down separately. (If they are strongly related, reply to your own comments. If strongly ordered, then go ahead and post them together.)
Do not quote yourself
Do not quote comments/posts on LW/OB
No more than 5 quotes per person per monthly thread, please.
SMBC Comics #2719
Well, his point only makes any sense when applied to the metaphor since a better answer to the question
is:
“where would Sisyphus get a robot in the middle of Hades?”
Edit: come to think of it, this also works with the metaphor for human struggle.
I thought the correct answer would be, “No time for programming, too busy pushing a boulder.”
Though, since the whole thing was a punishment, I have no idea what the punishment for not doing his punishment would be. Can’t find it specified anywhere.
I don’t think he’s punished for disobeying, I think he’s compelled to act. He can think about doing something else, he can want to do something else, he can decide to do something else … but what he does is push the boulder.
The version I like the best is that Sisyphus keeps pushing the boulder voluntarily, because he’s too proud to admit that, despite all his cleverness, there’s something he can’t do. (Specifically, get the boulder to stay at the top of the mountain).
My favorite version is similar. Each day he tries to push the boulder a little higher, and as the boulder starts to slide back, he mentally notes his improvement before racing the boulder down to the bottom with a smile on his face.
Because he gets a little stronger and a little more skilled every day, and he knows that one day he’ll succeed.
In the M. Night version: his improvements are an asymptote—and Sisyphus didn’t pay enough attention in calculus class to realize that the limit is just below the peak.
Or maybe the limit is the peak. He still won’t reach it.
In some versions he’s harassed by harpies until he gets back to boulder-pushing. But RobinZ’s version is better.
Borrowing one of Hephaestus’, perhaps?
Now someone just has to write a book entitled “The Rationality of Sisyphus”, give it a really pretentious-sounding philosophical blurb, and then fill it with Grand Theft Robot.
He can build it. It would be pretty hard to do while pushing a boulder up a hill, but he has all the time in the world!
Does he have any suitable raw materials?
It’s a hill, not a mountain. It presumably has plants, which could be burned for fuel. It’s possible that there’s no metal in the hill, but he could make the robot out of stone. Once he gets the prototype running, he can search for better materials as it slowly pushes the rock up the hill. He just has to get back before the fuel runs out.
Of course he could.
Answer: Because the Greek gods are vindictive as fuck, and will fuck you over twice as hard when they find out that you wriggled out of it the first time.
Who was the guy who tried to bargain the gods into giving him immortality, only to get screwed because he hadn’t thought to ask for youth and health as well? He ended up being a shriveled crab like thing in a jar.
My highschool english teacher thought this fable showed that you should be careful what you wished for. I thought it showed that trying to compel those with great power through contract was a great way to get yourself fucked good an hard. Don’t think you can fuck with people a lot more powerful than you are and get away with it.
EDIT: The myth was of Tithonus. A goddess Eos was keeping him as a lover, and tried to bargain with Zeus for his immortality, without asking for eternal youth too. Ooops.
I’m no expert, but that seems to be the moral of a lot of Greek myths.
King Midas, too.
I’d say this captures the spirit of Less Wrong perfectly.
-- Linus Pauling
Citation for this was hard; the closest I got was Etzioni’s 1962 The Hard Way to Peace, pg 110. There’s also a version in the 1998 Linus Pauling on peace: a scientist speaks out on humanism and world survival : writings and talks by Linus Pauling; this version goes
Did you take “expect” to mean as in prediction, or as in what you would have them do, like the Jesus version?
How about doing unto others what maximizes total happiness, regardless of what they’d do unto you?
The former is computationally far more feasible.
By acting in a way that discourages them from hurting you, and encouraging them to help you, you are playing your part in maximizing total happiness.
Yeah, but it’s not necessarily the ideal way to act. Perhaps you should act generally better than that, or perhaps you should try to amplify it more. Do what you can to find out the optimal way to act. At least pay attention if you find new information. Don’t just make a guess and assume you’re correct.
You don’t think you should discourage others from hurting you? I think that seems sort of obvious. Now, if you could somehow give a person a strong incentive to help you/ not hurt, while simultaneously granting them a shitload of happiness, that seems ideal. This doesn’t really exclude that, it’s just on the positive side of doing/ being done unto.
You should probably discourage others from hurting you. It’s just not clear how much.
As much as possible for the least amount of harm possible and the least amount of wasted time and resources, obviously. Which varies on a case by case basis.
I mean if it was practical, you’d give your friends 2 billion units of happiness, and then after turning the cheek to your enemies, grant them 1.9 billion units of happiness, but living on planet earth, giving you 80% of the crap you gave me seems about right.
Consider the consequences if everyone follows your rule. Assume someone gives you one unit of crap, possibly accidentally. You respond with 0.8 units. (It’s hard to measure this precisely, but for the sake of argument let’s assume that both of you manage to get it exactly right). He, in turn, responds with a further 0.64 units of crap. You respond to this with 0.512 units.
This is, of course, an infinite geometric series. The end result (over an infinite time period) is that you recieve 2 and 7⁄9 units of crap, while the other person recieves 2 and 2⁄9 units of crap. He recieves exactly 80% of the amount that you recieved, but you recieved over twice as much as you started out recieving.
If you return x% of the crap you get (for 0<x<100), and everyone else follows the same rule, then the total crap you recieve for every starting unit of crap is:
%5E2%20}%0A)This is clearly minimized at x=0.
Alternatively: he could notice that he gave you 1 unit of crap and assume the 0.8 units of crap you gave him is an equal penalty.
If someone yells at you, you’re likely to respond—but if someone yells at you because you just pushed them, you’re less likely to respond.
Or he could know I was going to give him the .512 units, from prior experience, and not give .64, which is the whole point.
That assumes that he is following a different rule from the rule that you are following. Does knowing that he will give you the 0.64 units prevent you from giving him the 0.8 units?
Yes. Depending on the circumstance, I might give him much less or much more and/ or choose a different course of action entirely.
Not necessarily. If I horribly torture Jim because Jim stepped on my toes, then I am not maximizing total happiness; the unhappiness given to Jim by the torture outwieghs the unhappiness in me that is prevented by having no-one step on my toes.
That’s a lot of effort and pain to prevent someone stepping on your toes.
Also, I’m not sure that’d be a terribly effective way to prevent harm to yourself. I mean, to the extent possible, once everyone knows you tortured Jim, people will be scared shitless to step on your toes, but Jim and Jim’s family are very likely to murder you, or at least sue you for all your money and put you in jail for a long time.
You are correct; it is not terribly effective. However, any disproportionate response to a minor, or even an imagined, slight will reduce total unhappiness while discouraging others from hurting me.
No. I just told you. Sometimes a disproportionate response encourages other people to hurt you. That’s actually part of the rule.
Doing unto others that which causes maximum total happiness leaves you vulnerable to Newcomb problems. You want to do unto others that which logically entails maximum total happiness. Under certain conditions, this is the same as Pauling’s recommendation.
I never mentioned causation. If you find a way to maximize it acausally, do that.
It has a tendency to go horribly wrong.
It’s impossible to find a strategy that produces happiness better than trying to produce happiness, since if you knew of one, you’d try to produce happiness by following that strategy. If this method is what works best, then in doing what works best, you’d follow this method.
Also, linking to TVTropes tends to fall under generalizing from fictional evidence.
Art imitates life. ;)
And it’s not hard to think of real life examples of atrocities “justified” on utilitarian grounds that the rest of the world thinks are anything but justifiable. The Reign of Terror during the French Revolution, for example, is generally regarded as having gone too far.
Would it help if the link were aimed at the real life section?
It has been deleted to prevent edit war.
It’s a nice sentiment, but the optimization problem you suggest is usually intractable.
It’s better to at least attempt it than just find an easier problem and do that. You might have to rely on intuition and such to get any answer, but you’re not going to do well if you just find something easier to optimize.
Yes, but there’s no way a pithy quote is going to solve the problem for you. It might, however, contain a useful heuristic.
You may do that if you must, I recommend against it.
Why do you recommend against it? Do you have a more complicated utility function?
Most human utility functions give their own happiness more weight than other’s. If you take into account that humans increase the happiness of others because it makes themself happy, you could even say that human utility functions only care about the happiness of their corresponding humans—but that is close to a tautology (“the utility function cares about the utility of the agent only”).
That quote is really annoying because Jesus says the same thing way better, repeatedly, in the Sermon on the Mount.
Jesus used a clever quip to point out the importance of self-monitoring for illusory superiority?
Just read the Sermon on the Mount.
What’s the Jesus quote? (Or, I guess, one instance of it.)
On being nicer than you think you should be:
(Matthew 5:38-41)
On self-superiority bias:
(Matthew 7:3)
I disagree with Will’s interpretation that the former follows from the latter.
Sorry, yeah, I was being a little too trollish. What I meant was that it was a single step implication from combining a few parts of the Sermon on the Mount; the examples you gave are likely indeed the two most representative ones for reaching that conclusion. Out of context I agree the mote and beam exhortation isn’t enough.
(Also, I disagree with your choice of translation, but that of course doesn’t matter in the scheme of things. Just felt that there was a 20% chance you’d care what I thought about that matter.)
No he didn’t. You are wrong about either the religious teaching you advocate or the thing that is being advocated in the grandparent.
Eh. What irritates me is his implicit claim that the ideas there are original or exclusive to Jesus.
With the amount of censorship, deliberate credit-stealing, and other failures of memetic replication in the ancient world, the chance is pretty slim that the earliest instance you’ve heard of of a moral or religious idea is actually its earliest invention. We should expect that the earliest extant sources for an idea do not correctly attribute its origin: there are so many more ways to be wrong than right, and the correct attribution of an idea (or an entertaining story, for that matter) is not under selection pressure to stay correct as the idea itself is.
(Consider that until the 1853 discovery of the Epic of Gilgamesh, a European scholar might well have believed that the story of Noah’s Flood originated with the Bible. We similarly know that much of the mythos of Jesus echoes earlier salvific gods and demigods — Mithras, Dionysus, Osiris, etc. — whose cults were later suppressed as pagan.)
So thinking of the moral teachings of Jesus as originally Christian seems problematic. For instance, given the extensive contact between the Near East and India since the time of Alexander, it’s reasonable to consider some contact with ideas from Buddhism, Jainism, etc. — as well as the Greek (or Greco-Egyptian) philosophy more readily recognized by Christian sources.
My point here isn’t to say that Jesus was a Buddhist, of course — but rather that if we happen to observe what look like moral truths (or just moral good ideas) in one particular tradition, we shouldn’t take that tradition seriously when it claims to have discovered them or to possess unique access to them.
I don’t really care about credit for originality, just beautifulness and deepness of message. Linus’ take is ugly, whereas the Sermon is beautifully constructed. Just seems a shame not to go for the latter whenever possible.
Linus’s take fits my aesthetic better, and “beautiful” language is often unclear.
mote beam single step implications
What?
(Upvoted so someone can explain it without Karma cost.)
Downvoted because feeding Will when he is speaking this kind of pretentious drivel is precisely the kind of thing that the cost is intended to penalize. It is an example of the system working as it should!
(Note that my own earlier reply would be penalized if I made it now and that too would be a desirable outcome. If I was confident that Will’s claim about the Sermon on the Mount would be dismissed and downvoted as it has been then I would not have made a response.)
Really, it’s an example of the system backfiring,causing someone to upvote a comment that deserved a downvoting it would probably otherwise have received.
That was my point.
What probability do you assign to someone with total karma less than 5 coming and translating this specific Will_Newsome’s comment into intelligible speach? My estimate is: epsilon.
Breaking a rule, and explaining that it has to be done to provide an opportunity for something with epsilon probability and a very low value even if it happened… that’s just an example of a person deliberately breaking a rule, and signalling dissatisfaction with the rule.
People respond to incentives. Especially loss-related incentives. I do not give homeless people nickels even though I can afford to give a nearly arbitrary number of homeless people nickels. The set of people with karma less than five will be outright unable to reply—the set of people with karma greater than five will just be disincentivized, and that’s still something.
The prior probability of someone being able to explain negative-value Will_Newsome’s comments in a way that provides value for LW readers is already epsilon. Even without the disincentives.
I think that people less responding to intentionally meaningless comments is a good thing. Therefore, trivial disincentives for doing this are a good thing. Therefore, removing them in this specific situation is a bad thing.
Will is referring to Matthew 7:1-5.
This claims that people underestimate their flaws relative to others’. Will claims that the obvious implication is that one must judge others more leniently to compensate, rather than refraining from judgement entirely as said two sentences earlier.
There’s also a suggestion of projection there. Having discovered that I have some flaw (say, anger; or baseless faith), I may go about finding the same fault in others — but if I correct the flaw in myself first, the world may look different. The one with road rage drives on a highway populated by assholes and maniacs; the creationist accuses Darwinism of “being a religion”.
It is, the way they’re trying to use that word. Also is intelligent design a type of creationism? ’Cuz I think I like ID, at least more than the standard model. I’d like to think of myself as a human in the reference class “creationist who accuses Darwinism of being a religion”.
Someone who claims that faith is a good thing should not also use it as an accusation of impropriety.
The creationist does not claim — before cowans, gentiles, and the unwashed — that Darwinism is the wrong religion; rather, he claims that it is “a religion” as if to say that this is condemnation enough. To fellow creationists he may well say that Darwinism is Satanism, or a rival tribe to be vanquished by force or deception. But he does not expect that argument to fly with outsiders. With them he merely asserts that the (straw-)Darwinist is a hypocrite, a know-it-all elitist nerd who commits the grave faux-pas of mistaking his religion for science.
Meanwhile the sociologist of religion wonders where the temples of Darwin are. The strong-programme sociologist of science (who uses the methodological assumption that science doesn’t work, even as he posts on the Internet!) can mistake a laboratory for a center of ritual, but one who has studied comparative religion does not see worship happening in the microscope, the genomics software, or the fMRI.
I get the impression that that argument is used more to undermine claims that darwinism is a science than anything else.
Physics is a clear science; you can use the right equations and predict the motion of the Earth about the Sun, or the time a barometer will take to fall from a given height. This gives it a certain degree of credibility. The theory of evolution (and how the creationists love to remind everyone of that word, ‘theory’!) is also science; but they would deny it, on the basis that accepting it suggests that it is as credible as physics or mathematics. If they insist that darwinism is a religion, then both alternatives start from the same basis of credibility; the creationists can then point out, quite accurately, that their version is older and has been around for longer, and therefore at least claim seniority.
There’s a short story by Asimov that gives a very nice view of the whole argument.
That is a quintessentially Asimovian story. +1.
Remember that Darwinism is a lot more than biology. Sure, a computer isn’t exactly an altar. That doesn’t change that most of what universities are famous for in the wider world is their ideology.
Eh that’s sort of a less charitable reading of me than you could have given. But I suppose you’ve already walked with me 1.2 miles, and it’d be a stretch for me to ask for .8 more. ;)
One way we say it here is to be cautious of other-optimizing …
Though sadly sometimes the only alternative is no optimization at all.
Yes, but more frequently than that, the only alternative appears to be no optimization at all. Hence the heuristic.
“A writer who says that there are no truths, or that all truth is ‘merely relative,’ is asking you not to believe him. So don’t.” ― Roger Scruton, Modern Philosophy: An Introduction and Survey
I am sympathetic to this line, but Scruton’s dismissal seems a little facile. If somebody says the truth is relative, then they can bite the bullet if they wish and say that THAT truth is also relative, thus avoiding the trap of self-contradiction. It might still be unwise to close your ears to them.
Consider a case where we DO agree that a given subject matter is relative; e.g., taste in ice-cream. Suppose Rosie the relativist tells you: “This ice-cream vendor’s vanilla is absolutely horrible, but that’s just my opinion and obviously it’s relative to my own tastes.” You would probably agree that Rosie’s opinion is indeed “just relative”… and still give the vanilla a miss this time.
If “this vanilla ice cream is horrible” is relatively true, then “Rosie’s opinion is that this vanilla ice cream is horrible” is absolutely true.
Related:
He’s just showing that those people don’t give infinite value, not that it’s nonsense. It’s nonsense because, even if you consider life infinitely more intrinsically valuable than a green piece of paper, you’d still trade a life for green pieces of paper, so long as you could trade them back for more lives.
If life were of infinite value, trading a life for two new lives would be a meaningless operation—infinity times two is equal to infinity. Not unless by “life has infinite value” you actually mean “everything else is worthless”.
Not quite so! We could presume that value isn’t restricted to the reals + infinity, but say that something’s value is a value among the ordinals. Then, you could totally say that life has infinite value, but two lives have twice that value.
But this gives non-commutativity of value. Saving a life and then getting $100 is better than getting $100 and saving a life, which I admit seems really screwy. This also violates the Von Neumann-Morgenstern axioms.
In fact, if we claim that a slice of bread is of finite value, and, say, a human life is of infinite value in any definition, then we violate the continuity axiom… which is probably a stronger counterargument, and tightly related to the point DanielLC makes above.
If we want to assign infinite value to lives compared to slices of bread, we don’t need exotic ideas like transfinite ordinals. We can just define value as an ordered pair (# of lives, # of slices of bread). When comparing values we first compare # of lives, and only use # of slices of bread as a tiebreaker. This conforms to the intuition of “life has infinite value” and still lets you care about bread without any weird order-dependence.
This still violates the continuity axiom, but that, of itself, is not an argument against a set of preferences. As I read it, claiming “life has infinite value” is an explicit rejection of the continuity axiom.
Of course, Kaj Sotala’s point in the original comment was that in practice people demonstrate by their actions that they do accept the continuity axiom; that is, they are willing to trade a small risk of death in exchange for mundane benefits.
You could use hyperreal numbers. They behave pretty similarly to reals, and have reals as a subset. Also, if you multiply any hyperreal number besides zero by a real number, you get something isomorphic to the reals, so you can multiply by infinity and it still will work the same.
I’m not a big fan of the continuity axiom. Also, if you allow for hyperreal probabilities, you can still get it to work.
True
Only if you have a way to describe infinity in terms of a real number.
You just pick some infinite hyper real number and multiply all the real numbers by that. What’s the problem?
Oh, you’re saying assign a hyperreal infinite numbers to the value of individual lives. That works, but be very careful how you value life. Contradictions and absurdities are trivial to develop when one aspect is permitted to override every other one.
At which point why not just re-normalize everything so that you’re only dealing with reals?
You could have something have infinite value and something else have finite value. Since this has an infinitesimal chance of actually mattering, it’s a silly thing to do. I was just pointing out that you could assign something infinite utility and have it make sense.
Nitpick, I think you mean non-commutativity, the ordinals are associative. The rest of your post agrees with this interpretation.
Oops, yes. Edited in original; thanks!
John Perry, introduction to Identity, Personal Identity, and the Self
He bought the present ox along with the future ox. He could have just bought the present ox, or at least a shorter interval of one. This is known as “renting”.
Which future ox did he buy?
Sorry. The future oxen.
Of the many oxen at a given point in time in the future, which one did he buy?
Oh. I see what you mean.
He bought the ones on the future side of that worldline.
It’s convenient that way, and humans are good at keeping track. He could have bought any combination of future oxen the guy owns. This has the advantage of later oxen being in the same area as earlier oxen, simplifying transportation.
I’m not making myself clear. It’s clear from what you say that if Heraclites bought an ox yesterday, he owns an ox today. But in order to say that he owns this particular ox, he needs a better system of identity than “you never step into the same river twice”.
It’s a sensible system for deciding how to buy and sell oxen because it minimizes shipping costs. It’s a less sensible way to, for example, judge the value of a person. Should I choose Alice of Bob just because Bob is at the beginning of a world-line and Alice is not?
This does kind of come back to arguing definitions. The common idea of identity is really useful. If a philosopher thinks otherwise, he’s overthinking it. “Identity” refers to something. I just don’t think it’s anything beyond that. You in principle could base your ethics on it, but I see no reason to. It’s not as if it’s something anybody can experience. If you base your anthropics on it, you’ll only end up confusing yourself.
Alternatively, he purchased the present ox using the ox-an-hour-ago as payment.
No. There is nobody to make that transaction with, and his past self still used the past ox, so he can’t sell it.
What those eggheads need is a good ass-kicking!
A praiseworthy sentiment.
Solzhenitsyn
If only it were a line. Or even a vague boundary between clearly defined good and clearly defined evil. Or if good and evil were objectively verifiable notions.
You don’t think even a vague boundary can be found? To me it seems pretty self-evident by looking at extremes; e.g., torturing puppies all day is obviously worse than playing with puppies all day.
By no means am I secure in my metaethics (i.e., I may not be able to tell you in exquisite detail WHY the former is wrong). But even if you reduced my metaethics down to “whatever simplicio likes or doesn’t like,” I’d still be happy to persecute the puppy-torturers and happy to call them evil.
Animal testing.
And even enjoying torturing puppies all day is merely considered “more evil” because it’s a predictor of psychopathy.
So I think maybe I leapt into this exchange uncarefully, without being clear about what I was defending. I am defending the meaningfulness & utility of a distinction between good & evil actions (not states of affairs). Note that a distinction does not require a sharp dividing line (yellow is not the same as red, but the transition is not sudden).
I also foresee a potential disagreement about meta-ethics, but that is just me “extrapolating the trend of the conversation.”
Anyway, getting back to good vs evil: I am not especially strict about my use of the word “evil” but I generally use it to describe actions that (a) do a lot of harm without any comparably large benefit, AND (b) proceed from a desire to harm sentient creatures.
Seen in this light it is obvious why torturing puppies is evil, playing with them is good, and testing products on them is ethically debatable (but not evil, because of the lack of desire to harm). None of this is particularly earth-shattering as philosophical doctrine.
Not if you think animals’ interests count morally, which I do explicitly, and virtually everybody does implicitly.
I think your philosophy is probably fairly normal, it’s just any attempt to simplify such things looks like an open challenge to point out corner cases. Don’t take it too seriously.
Also I’m not fully convinced on whether animals’ interests count morally, even though they do practically by virtue of triggering my empathy. Aside from spiders. Those can just burn. (Which is an indicator that animals only count to me because they trigger my empathy, not because I care)
But.. but.. they just want to give you a hug.
You point out that there are acts easily agreed to be evil and acts easily agreed to be good, but that doesn’t imply a definable boundary between good and evil. First postulate a boundary between good and evil. Now, what is necessary to refute that boundary? A clearly defined boundary would require actions that fall near the boundary to always fall to on side or the other without fail. Easily, that is not the case. Stealing food is clearly evil if you have no need but the victim has need for the food. If the needs are opposite, then it is not clearly evil. So there is no clear boundary, but what would a vague boundary require? A think a vague boundary requires that actions can be ranked in a vague progression from “certainly good” through “overall good, slightly evil” and descend through progressively less good zones as they approach from one side, then crossing a “evil=~good” area, into a progressively more evil side. I do not see that is necessarily the case.
You are pointing to different actions labeled stealing and saying “one is good and the other is evil.” Yeah, obviously, but that is no contradiction—they are different actions! One is the action of stealing in dire need, the other is the action of stealing without need.
This is a very common confusion. Good and evil (and ethics) are situation-dependent, even according to the sternest, most thundering of moralists. That does not tell us anything one way or the other about objectivity. The same action in the same situation with the same motives is ethically the same.
Thank you for pointing out my confusion. I’ve lost confidence that I have any idea what I’m talking about on this issue.
I think the intermediate value theorem covers this. Meaning if a function has positive and negative values (good and evil) and it is continuous (I would assume a “vague boundary” or “grey area” or “goodness spectrum” to be continuous) then there must be at least one zero value. That zero value is the boundary.
It would indeed cover this if goodness spectrum was a regular function, not a set-valued map. Unfortunately, the same thoughts and actions can correspond to different shades of good and evil, even in the mind of the same person, let alone of different people. Often at the same time, too.
This shows that there is disagreement & confusion about what is good & what is evil. That no more proves good & evil are meaningless, than disagreement about physics shows that physics is meaningless.
Actually, disagreement tends to support the opposite conclusion. If I say fox-hunting is good and you say it’s evil, although we disagree on fox-hunting, we seem to agree that only one of us can possibly be right. At the very least, we agree that only one of us can win.
— Steven Kaas
Duplicate.
Clearly, the good piece of each heart is willing to destroy what it thinks of as the evil piece. It’s just a question of what piece we choose to identify with as representing the “real person”.
Charles Kettering
A problem sufficiently well-stated is a problem fully solved.
Wow, I didn’t even know that’s a quote from someone! I had inferred that (mini)lesson from a lecture I heard, but it wasn’t stated in those terms, and I never checked if someone was already known for that.
Ken Wilber
--G.K. Chesterton, “The Duel of Dr. Hirsch”
Reversed malevolence is intelligence?
Inverted information is not random noise.
...unless you’re reversing noise which is why Reverse Stupidity is not Intelligence.
If someone tells you the opposite of the truth in order to deceive you, and you believe the opposite of what they say because you know they are deceitful, then you believe the truth. (A knave is as good as a knight to a blind bat.) The problem is, a clever liar doesn’t lie all the time, but only when it matters.
It’s more likely that they’re a stupid liar than that they got it all wrong by chance.
Another problem is that for many interesting assertions X, opposite(opposite(X)) does not necessarily equal X. Indeed, opposite(opposite(X)) frequently implies NOT X.
Could you give an example? I would have thought this happens with Not(opposite(X)); for example, “I don’t hate you” is different than “I love you”, and in fact implies that I don’t. But I would have thought “opposite” was symmetric, so opposite(opposite(X)) = X.
Well, OK. So suppose (to stick with your example) I love you, and I want to deceive you about it by expressing the opposite of what I feel. So what do I say?
You seem to take for granted that opposite(“I love you”) = “I hate you.” And not, for example, “I am indifferent to you.” Or “You disgust me.” Or various other assertions. And, sure, if “I love you” has a single, unambiguous opposite, and the opposite also has a single, unambiguous opposite, then my statement is false. But it’s not clear to me that this is true.
If I end up saying “I’m indifferent to you” and you decide to believe the opposite of that… well, what do you believe?
Of course, simply negating the truth (“I don’t love you”) is unambiguously arrived at, and can be thought of as an opposite… though in practice, that’s often not what I actually do when I want to deceive someone, unless I’ve been specifically accused of the truth. (“We’re not giant purple tubes from outer space!”)
Lol, my professor would give a 100% to anyone who answered every exam question wrong. There were a couple people who pulled it off, but most scored 0<10.
I’m assuming a multiple-choice exam, and invalid answers don’t count as ‘wrong’ for that purpose?
Otherwise I can easily miss the entire exam with “Tau is exactly six.” or “The battle of Thermopylae” repeated for every answer. Even if the valid answers are [A;B;C;D].
Unless it really was the battle of Thermopylae. Not having studied, you wont know.
“The Battle of Thermopylae” is intended as the alternate for questions which might have “Tau is exactly six” as the answer.
For example: “What would be one consequence of a new state law which defines the ratio of a circle’s circumference to diameter as exactly three?”
I bet that you can’t write a question for which “Tau is exactly six.” and “The battle of Thermopylae” are both answers which gain any credit...
“Write a four word phrase or sentence.”
You win.
Judging by this and your previous evil genie comments, you’d make a lovely UFAI.
I hate to break up the fun, and I’m sure we could keep going on about this, but Decius’s original point was just that giving a wrong answer to an open-ended question is trivially easy. We can play word games and come up with elaborate counter-factuals, but the substance of that point is clearly correct, so maybe we should just move on.
That was exactly the challenge I issued. Granted, it’s trivial to write an answer which is wrong for that question, but it shows that I can’t find a wrong answer for an arbitrary question as easily as I thought I could.
“The duel between the current King of France and the former Emperor of Britain”. There, the answer won’t ever be that phrase.
What if you are in a literature class and you’re taking a test about a fiction book you didn’t read?
Care to bet on that?
An interesting corollary of the efficient market hypothesis is that, neglecting overhead due to things like brokerage fees and assuming trades are not large enough to move the market, it should be just as difficult to lose money trading securities as it is to make money.
No, not really. In an efficient marked risks uncorrelated with those of other securities shouldn’t be compensated, so you should easily be able to screw yourself over by not diversifying.
But isn’t the risk of diversifying compensated by a corresponding possibility of large reward if the sector outperforms? I wouldn’t consider a strategy that produces modest losses with high probability but large gains with low probability sufficient to disprove my claim.
Let’s go one step back on this, because I think our point of disagreement is earlier than I thought in that last comment.
The efficient market hypothesis does not claim that the profit on all securities has the same expectation value. EMH-believers don’t deny, for example, the empirically obvious fact that this expectation value is higher for insurances than for more predictable businesses. Also, you can always increase your risk and expected profit by leverage, i.e. by investing borrowed money.
This is because markets are risk-averse, so that on the same expectation value you get payed extra to except a higher standard deviation. Out- or underperforming the market is really easy by excepting more or less risk than it does on average. The claim is not that the expectation value will be the same for every security, only that the price of every security will be consistent with the same prices for risk and expected profit.
So if the EMH is true, you can not get a better deal on expected profit without also accepting higher risk and you can not get a higher risk premium than other people. But you still can get lots of different trade-offs between expected profit and risk.
Now can you do worse? Yes, because you can separate two types of risk.
Some risks are highly specific to individual companies. For example, a company may be in trouble if a key employee gets hit by a beer truck. That’s uncorrelated risk. Other risks affect the whole economy, like revolutions, asteroids or the boom-bust-cycle. That’s correlated risk.
Diversification can insure you against uncorrelated risk, because, by definition, it’s independent from the risk of other parts of your portfolio, so it’s extremely unlikely for many of your diverse investments to be affected at the same time. So if everyone is properly diversified, no one actually needs to bear uncorrelated risk. In an efficient market that means it doesn’t earn any compensation.
Correlated risk is not eliminated by diversification, because it is by definition the risk that affects all your diversified investments simultaneously.
So if you don’t diversify you are taking on uncorrelated risk without getting paid for it. If you do that you could get a strictly better deal by taking on a correlated risk of the same magnitude which you would get payed for. And since that is what the marked is doing on average, you can get a worse deal than it does.
Unless you’re a fictional character. Or possibly Mike “Bad Player” Flores:
I thought your first link would be Bloody Stupid Johnson.
This reminds me of an episode of QI, in which Johnny Vegas, who usually throws out random answers for the humor, actually managed to get a question (essentially) right.
Anonymous
Not always, since:
Des McHale
In other words, the average of a distribution is not necessarily the most probable value.
In other words: expect Lady Mode), not Lady Mean.
Don’t expect her, either. In Russian Roulette, the mode is that you don’t die, and indeed that’s the outcome for most people who play it. You should, however, expect that there’s a very large chance of instadeath, and if you were to play a bunch of games in a row, that (relatively uncommon) outcome would almost certainly kill you.
(A similar principle applies to things like stock market index funds: the mode doesn’t matter when all you care about is the sum of the stocks.)
The real lesson is this: always expect Lady PDF.
Not to be a bore but it does say “Lady Average” not “Sir or Madam Average”.
In my high school health class, for weeks the teacher touted the upcoming event: “Breast and Testicle Day!”
When the anticipated day came, it was of course the day when all the boys go off to one room to learn about testicular self-examination, and all the girls go off to another to learn about breast self-examination. So, in fact, no student actually experienced Breast and Testicle Day.
Much to their chagrin, I’m assuming.
Rather: chagrin and relief.
Not to be a bore but it does say “Lady Average” not “Sir or Madam Average”.
Lady Main Mode? Does not sound that good. Lady Median?
If you’re asking who comes around most often, Lady Mode it is—we can’t help how it sounds.
Lady Mode is the most fashionable.
-Robin Hanson, Human Enhancement
I feel like Hanson’s admittedly insightful “signaling” hammer has him treating everything as a nail.
Your contrarian stance against a high-status member of this community makes you seem formidable and savvy. Would you like to be allies with me? If yes, then the next time I go foraging I will bring you back extra fruit.
I agree in principle but I think this particular topic is fairly nailoid in nature.
I’d say it’s such a broad subject that there have to be some screws in there as well. I think Hanson has too much faith in the ability of evolved systems to function in a radically changed environment. Even if signaling dominates the evolutionary origins of our brain, it’s not advisable to just label everything we do now as directed towards signaling, any more than sex is always directed towards reproduction. You have to get into the nitty gritty of how our minds carry out the signaling. Conspiracy theorists don’t signal effectively, though you can probably relate their behavior back to mechanisms originally directed towards, or at least compatible with, signaling.
Also, an ability to switch between clear “near” thinking and fluffy “far” thinking presupposes a rational decision maker to implement the switch. I’m not sure Hanson pays enough attention to how, when, and for what reasons we do this.
Same here.
I think he’s mischaracterizing the issue.
Beliefs serve multiple functions. One is modeling accuracy, another is signaling. It’s not whether the environment is harsh or easy, it’s which function you need. There are many harsh environments where what you need is the signaling function, and not the modeling function.
I think the quote reflects reality (humans aren’t naturally rational so their beliefs are conditioned by circumstance), but is better seen as an observation than a recommendation. The best approach should always be to hold maximally accurate beliefs yourself, even if you choose to signal different ones as the situation demands. That way you can gain the social benefits of professing a false belief without letting it warp or distort your predictions.
No, that wouldn’t necessarily be the case. We should expect a cost in effort and effectiveness to try to switch on the fly between the two types of truths. Lots of far truths have little direct predictive value, but lots of signaling value. Why bear the cost for a useless bit of predictive truth, particularly if it is worse than useless and hampers signaling?
That’s part of the magic of magisteria—segregation of modes of truth by topic reduces that cost.
Hmm, maybe I shouldn’t have said “always” given that acting ability is required to signal a belief you don’t hold, but I do think what I suggest is the ideal. I think someone who trained themselves to do what I suggest, by studying people skills and so forth, would do better as they’d get the social benefits of conformity and without the disadvantages of false beliefs clouding predictions (though admittedly the time investment of learning these skills would have to be considered).
Short version: I think this is possible with training and would make you “win” more often, and thus it’s what a rationalist would do (unless the cost of training proved prohibitive, of which I’m doubtful since these skills are very transferable).
I’m not sure what you meant by the magisteria remark, but I get the impression that advocating spiritual/long-term beliefs to less stringent standards than short term ones isn’t generally seen as a good thing (see Eliezer’s “Outside the Laboratory” post among others).
Clothes serve multiple functions. One is keeping warm, another is signalling.
-L. A. Rollins, Lucifer’s Lexicon: An Updated Abridgment
“He who receives an idea from me, receives instruction himself without lessening mine; as he who lights his candle at mine, receives light without darkening me. No one possesses the less of an idea, because every other possesses the whole of it.”—Jefferson
But many people do benefit greatly from hoarding or controlling the distribution of scarce information. If you make your living off slavery instead, then of course you can be generous with knowledge.
Or if, say, you run a university.
If you do not hoard your ideas, and neither do I, then we can both benefit from the ideas of the other. If I can access the ideas of a hundred other people at the cost of sharing my own ideas, then I profit; no matter how smart I am, a hundred other people working the same problem are going to be able to produce at least some ideas that I did not think of. (This is a benefit of free/open source software; it has been shown experimentally to work pretty well in the right circumstances).
-- Tim Kreider
The interesting part is the phrase “which sounds like the kind of lunatic notion that’ll be considered a basic human right in about a century, like abolition, universal suffrage and eight-hour workdays.” If we can anticipate what the morality of the future would be, should we try to live by it now?
Not if it’s actually the same morality, but depends on technology. For example, strong prohibitions on promiscuity are very sensible in a world without cheap and effective contraceptives. Anyone who tried to live by 2012 sexual standards in 1912 would soon find they couldn’t feed their large horde of kids. Likewise, if robots are doing all the work, fine; but right now if you just redistribute all money, no work gets done.
Lack of technology was not the reason condoms weren’t as widely available in 1912.
Right idea, not a great example. People used to have lots more kids then now, most dying in childhood. Majority of women of childbearing age (gay or straight) were married and having children as often as their body allowed, so promiscuity would not have changed much. Maybe a minor correction for male infertility and sexual boredom in a standard marriage.
You seem to have rather a different idea of what I meant by “2012 standards”. Even now we do not really approve of married people sleeping around. We do, however, approve of people not getting married until age 25 or 30 or so, but sleeping with whoever they like before that. Try that pattern without contraception.
You might. I don’t. This is most probably a cultural difference. There are people in the world to day who see nothing wrong with having multiple wives, given the ability to support them (example: Jacob Zuma)
Strong norms against promiscuity out of wedlock still made sense though, since having lots of children without a committed partner to help care for them would usually have been impractical.
Not if they were gay.
Then they’d just be dead, or imprisoned.
We’re talking about morality that is based around technology. There is no technological advance that allows us to not criminalize homosexuality now where we couldn’t have in the past.
Naming three:
Condoms.
Widespread circumcision.
Antibiotics.
What?
Didn’t the Jews have that back in the years BC? It’s sort of cultural, but it’s been around for a while in some cultures...
Condoms may be older than you think.
I didn’t specify promiscuous homosexuality. Monogamously inclined gay people are as protected from STDs as anyone else at a comparable tech level—maybe more so among lesbians.
Neither did I, but would rather refrain from explaining in detail why I didn’t assume promiscuity.
It’s really annoying that you jumped to that conclusion, though. Further, I’m confused why the existence of some minority of a minority of the population that doesn’t satisfy the ancestor’s hypothetical matters.
Homosexuality was common/accepted/expected in many societies without leading to any negative consequences, so technology is not an enabler of morality here.
Homosexuality has certainly been present in many societies.
However, your link does not state, nor even suggest, that it did not lead to any negative consequences.
How do you envision living by this model now working?
That is, suppose I were to embrace the notion that having enough resources to live a comfortable life (where money can stand in as a proxy for other resources) is something everyone ought to be guaranteed.
What ought I do differently than I’m currently doing?
I would like to staple that question to the forehead of every political commentator who makes a living writing columns in far mode. What is it you would like us to do? If you don’t have a good answer, why are you talking? Argh.
Not if the morality you anticipate coming into favour is something you disagree with. If it’s something you agree with, it’s already yours, and predicting it is just a way of avoiding arguing for it.
If you are a consequentialist, you should think about the consequences of such decision.
For example, imagine a civilization where an average person has to work nine hours to produce enough food to survive. Now the pharaoh makes a new law saying that (a) all produced food has to be distribute equally among all citizens, and (b) no one can be compelled to work more than eight hours; you can work as a volunteer, but all your produced food is redistributed equally.
What would happen is such situation? In my opinion, this would be a mass Prisoners’ Dilemma where people would gradually stop cooperating (because the additional hour of work gives them epsilon benefits) and start being hungry. There would be no legal solution; people would try to make some food in their free time illegally, but the unlucky ones would simply starve and die.
The law would seem great in far mode, but its near mode consequences would be horrible. Of course, if the pharaoh is not completely insane, he would revoke the law; but there would be a lot of suffering meanwhile.
If people had “a basic human right to have enough money without having to work”, situation could progress similarly. It depends on many things—for example how much of the working people’s money would you have to redistribute to non-working ones, and how much could they keep. Assuming that one’s basic human right is to have $500 a month, but if you work, you can keep $3000 a month, some people could still prefer to work. But there is no guarantee it would work long-term. For example there would be a positive feedback loop—the more people are non-working, the more votes politicians can gain by promising to increase their “basic human right income”, the higher are taxes, and the smaller incentives to work. Also, it could work for the starting generation, but corrupt the next generation… imagine yourself as a high school student knowing that you will never ever have to work; how much effort would an average student give to studying, instead of e.g. internet browsing, Playstation gaming, or disco and sex? Years later, the same student will be unable to keep a job that requires education.
Also, if less people have to work, the more work is not done. For example, it will take more time to find a cure for cancer. How would you like a society where no one has to work, but if you become sick, you can’t find a doctor? Yes, there would be some doctors, but not enough for the whole population, and most of them would have less education and less experience than today. You would have to pay them a lot of money, because they would be rare, and because most of the money you pay them would be paid back to state as tax, so even everything you have could be not enough motivating for them.
Systems that don’t require people to work are only beneficial if non-human work (or human work not motivated by need) is still producing enough goods that the humans are better off not working and being able to spend their time in other ways. I don’t think we’re even close to that point. I can imagine societies in a hundred years that are at that point (I have no idea whether they’ll happen or not), but it would be foolish for them to condemn our lack of such a system now since we don’t have the ability to support it, just as it would be foolish for us to condemn people in earlier and less well-off times for not having welfare systems as encompassing as ours.
I’d also note that issues like abolition and universal suffrage are qualitatively distinct from the issue of a minimum guaranteed income (what the quote addresses). Even the poorest of societies can avoid holding slaves or placing women or men in legally inferior roles. The poorest societies cannot afford the “full unemployment” discussed in the quote, and neither can even the richest of modern societies right now (they could certainly come closer than the present, but I don’t think any modern economy could survive the implementation of such a system in the present).
I do agree, however, about it being a solid goal, at least for basic amenities.
To avoid having slaves, the poorest society could decide to kill all war captives, and to let starve to death all people unable to pay their debts. Yes, this would avoid legal discrimination. Is it therefore a morally preferable solution?
Since when has the institution of slavery been a charitable one? Historically, slave-owners have payed immense costs, directly and indirectly, for the privilege of owning slaves, and done so knowingly and willingly. It is human nature to derive pleasure from holding power over others.
I’m not sure about those direct costs. According to my references, a male slave in 10th-century Scandinavia cost about as much as a horse, a female slave about two-thirds as much; that’s a pretty good chunk of change but it doesn’t seem obviously out of line with the value of labor after externalities. I don’t have figures offhand for any other slaveholding cultures, but the impression I get is that the pure exercise of power was not the main determinant of value in most, if not all, of them,
I seem to recall someone arguing that, in combat between iron age tribes, it was basically a choice between massacre and slavery—if you did neither, they would wreck revenge upon your tribe further down the line.
This wouldn’t be charity, as I guess the winners did benefit from having a source of labour that didn’t need to be compensated at the market rate, but it would be a case where slavery was beneficial to the victims.
(I think Carlyle was wrong about other supposed cases of slavery proving beneficial for the victims)
One could argue that happened in Ancient Rome, with prisoners fo war as the main source of slaves. Also they/their descendants arguably benefited in the long term from being part of the larger more sophisticated culture, if they survived that long.
In poor societies that permit slavery, a man might be willing to sell himself into slavery. He gets food and lodging, possibly for his family as well as himself; his new purchaser gets a whole lot of labour. There’s a certain loss of status, but a person might well be willing to live with that in order to avoid starvation.
Elections can take quite a bit of resources to run when you have a large voting population...
No, politicians can afford to spend lots of money on them. The actual mechanism of elections have never, so far as I know, been all that expensive pre-computation.
IAWYC, but the claims that most of the economic costs of elections are in political spending, and most of the costs of actually running elections are in voting machines are both probably wrong. (Public data is terrible, so I’m crudely extrapolating all of this from local to national levels.)
The opportunity costs of voting alone dwarf spending on election campaigns. Assuming that all states have the same share of GDP, that people who don’t a full-state holiday to vote take an hour off to vote, that people work 260 days a year and 8 hours a day, and that nobody in the holiday states do work, then we get:
Political spending: 5.3 billion USD Opportunity costs of elections: 15 trillion USD (US GDP) (9/50 (states with voting holidays) 1⁄260 (percentage of work-time lost) + 41⁄50 (states without holidays) 1⁄60 1⁄8 (percentage of work-time lost)) ≈ 16 billion USD
Extrapolating from New York City figures, election machines cost ~1.9 billion nationwide. (50 million for a population ~38 times smaller than the total US population.) and extrapolating Oakland County’s 650,000 USD cost across the US’s 3143 counties, direct costs are just over 2 billion USD. (This is for a single election; however, some states have had as many as 5 elections in a single year. The cost of the voting machines can be amortized over multiple elections in multiple years.)
(If you add together the opportunity costs for holding one general and one non-general election a year (no state holidays; around ~7 billion USD), plus the costs of actually running them, plus half the cost of the campaign money, the total cost/election seems to be around 30 billion USD, or ~0.002% of the US’s GDP.)
Correction accepted. Still seems like something a poor society could afford, though, since labor and opportunity would also cost less. I understand that lots of poor societies do.
What? If anything I’d assume them to be more expensive before computers were introduced. In Italy where they are still paper based they have to hire people to count the ballots (and they have to pay them a lot, given that they select people at random and you’re not allowed to refuse unless you are ill or something).
According to Wikipedia, the 2005 elections in germany did cost 63 million euros, with a population of 81 million people. 0,78 eurocent per person or the 0,00000281st part of the GDP. Does not seem much, in the grander scheme of things. And since the german constitutional court prohibited the use of most types of voting machines, that figure does include the cost to the helpers; 13 million, again, not a prohibitive expenditure.
http://aceproject.org/ace-en/focus/core/crb/crb03
“Low electoral costs, approximately $1 to $3 per elector, tend to manifest in countries with longer electoral experience”
That’s a somewhat confusing comment. If they’re effectively conscripted (them not being allowed to refuse), not really “hired”—that would imply they don’t need to be paid a lot...
Is that that little? I think many fewer people would vote if they had to pay $3 out of their own pocket in order to do so.
A law compelling people to do stuff would be very unpopular, unless they get adequate compensation. Not paying them much would just mean they would feign illness or something. (If they didn’t select people by lot, the people doing that would be the ones applying for that job, who would presumably like it more than the rest of the population and hence be willing to do that for less.)
Well perhaps fewer people would vote if they had to pay a single cent out of their own pocket—would that mean that 0.01$ isn’t little either?
How much are these Italian ballot-counters being paid? Can we quantify this?
IIRC, something like €150 per election. I’ll look for the actual figure.
Why so? Usually when people can’t refuse to do a job, they’re paid little, not a lot.
Like jury duty. Yeah. Why would it be different in Greece?
In the UK, the counters are volunteers.
Well, yes. Almost tautologically so, I should think.
The tricky part is working out when humans are better off.
If you are a bayesian, you should think about how much evidence your imagination constitutes.
For example, imagine a civilization where an average person gains little or no total productivity by working over 8 hour per day. Imagine, moreover, that in this civilization, working 10 hours a day doubles your risk of coronary heart disease, the leading cause of death in this civilization. Finally, imagine that, in this civilization, a common way for workers to signal their dedication to their jobs is by staying at work long hours, regardless of the harm it does both to their company and themselves.
In this civilization, a law preventing individuals from working over 8 hours per day is a tremendous social good.
Work hour skepticism leaves out the question of the cost of mistakes. It’s one thing to have a higher proportion of defective widgets on an assembly line (though even that can matter, especially if you want a reputation for high quality products), another if the serious injury rate goes up, and a third if you end up with the Exxon Valdez.
You mean “incentives to fully report your income”, right? ;-) (There are countries where a sizeable fraction of the economy is underground. I come from one.)
The same they give today. Students not interested in studying mostly just cheat.
Well, if your society isn’t rich enough, you just do what you can. (And a lot of work really isn’t all that important; would it be that big of a disaster if your local store carried fewer kinds of cosmetics, or if your local restaurant had trouble hiring waiters?)
See also.
It is true that in the long run, things could work out worse with a guarantee of sufficient food/supplies for everyone. I think, though, that this post answers the wrong question; the question to answer in order to compare consequences is how probable it is to be better or worse, and by what amounts. Showing that it “could” be worse merely answers the question “can I justify holding this belief” rather than the question “what belief should I hold”. The potential benefits of a world where people are guaranteed food seem quite high on the face of it, so it is a question well worth asking seriously… or would be if one were in a position to actually do anything about it, anyway.
Prisoners’ dilemmas amongst humans with reputation and social pressure effects do not reliably work out with consistent defection, and models of societies (and students*) can easily predict almost any result by varying the factors they model and how they do so, and so contribute very little evidence in the absence of other evidence that they generate accurate predictions.
The only reliable information that I am aware of is that we know that states making such guarantees can exist for multiple generations with no obvious signs of failure, at least with the right starting conditions, because we have such states existing in the world today. The welfare systems of some European countries have worked this way for quite a long time, and while some are doing poorly economically, others are doing comparably well.
I think that it is worth assessing the consequences of deciding to live by the idea of universal availability of supplies, but they are not so straightforwardly likely to be dire as this post suggests, requiring a longer analysis.
As I wrote, it depends on many things. I can imagine a situation where this would work; I can also imagine a situation where it would not. As I also wrote, I can imagine such system functioning well if people who don’t work get enough money to survive, but people who do work get significantly more.
Data point: In Slovakia many uneducated people don’t work, because it wouldn’t make economical sense for them. Their wage, minus traveling expenses, would be only a little more, in some cases even less than their welfare. What’s the point of spending 8 hours in work if in result you have less money? They cannot get higher wages, because they are uneducated and unskilled; and in Slovakia even educated people get relatively little money. The welfare cannot be lowered, because the voters on the left would not allow it. The social pressure stops working if too many people in the same town are doing this; they provide moral support for each other. We have villages where unemployment is over 80% and people have already accommodated to this; after a decade of such life, even if you offer them a work with a decent wage, they will not take it, because it would mean walking away from their social circle.
This would not happen in a sane society, but it does happen in the real life. Other European countries seem to fare better in this aspect, but I can imagine the same thing happening there in a generation or two. A generation ago most people would probably not predict this situation in Slovakia.
I also can’t imagine the social pressure to work on the “generation Facebook”. If someone spends most of their day on Facebook or playing online multiplayer games, who exactly is going to socially press them? Their friends? Most of them live the same way. Their parents? The conflict between generations is not the same thing as peer pressure. And the “money without work is a basic human right” meme also does not help.
It could work in a country where the difference between average wage (even for a less educated and less skilled people) is much more than one needs to survive. But it can work long-term only if the amount of “basic human right money” does not grow faster than the average wage. -- OK, finally here is something that can be measured: what is the relative increase in wages vs welfares in western European countries in recent decades; optionally, extrapolate these numbers to estimate how long the system can survive.
This is interesting, particularly the idea of comparing wage growth against welfare growth predicting success of “free money” welfare. I agree that it seems reasonably unlikely that a welfare system paying more than typical wages, without restrictions conflicting with the “detached from work” principle, would be sustainable, and identifying unsustainable trends in such systems seems like an interesting way to recognise where something is going to have to change, long-term.
I appreciate the clarification; it provides what I was missing in terms of evidence or reasoned probability estimates over narrative/untested model. I’m taking a hint from feedback that I likely still communicated this poorly, and will revise my approach in future.
Back on the topic of taking these ideas as principles, perhaps more practical near-term goals which provide a subset of the guarantee, like detaching availability of resources basic survival from the availability of work, might be more probably achievable. There are a wider range of options available for implementing these ideas, and of incentives/disincentives to avoid long-term use. An example which comes to mind is providing users with credit usable only to order basic supplies and basic food. My rough estimate is that it seems likely that something in this space could be designed to operate sustainably with only the technology we have now.
On the side, relating to generation Facebook, my model of the typical 16-22 year old today would predict that they’d like to be able to buy an iPad, go to movies, afford alcohol, drive a nice car, go on holidays, and eventually get most of the same goals previous generations sought, and that their friends will also want these things. At younger ages, I agree that parental pressure wouldn’t be typically classified as “peer pressure”, but I still think it likely to provide significant incentive to do school work; the parents can punish them by taking away their toys if they don’t, as effectively as for earlier generations. My model is only based on my personal experience, so mostly this is an example of anecdotal data leading to different untested models.
I have heard this idea proposed, and many people object against it saying that it would take away the dignity of those people. In other words, some people seem to think that “basic human rights” include not just things necessary for survival, but also some luxury and perhaps some status items (which then obviously stop being status items, if everyone has them).
In theory, yes. However, as a former teacher I have seen parents completely fail at this.
Data point: A mother came to school and asked me to tell her 16 year old daughter, my student, to not spend all her free time at internet. I did not understand WTF she wanted. She explained to me that as a computer science teacher her daughter will probably regard me an authority about computers, so if I ask her to not use the computer all day long, she wil respect me. This was her last hope, because as a mother she could not convince her daughter to go away from the computer.
To me this seemed completely insane. First, the teachers in given school were never treated as authorities on anything; they were usually treated like shit both by students and school administration (a month later I left that school). Second, as a teacher I have zero influence on what my students do outside school, she as a mother is there; she has many possible ways to stop her daughter from interneting… for instance to forcibly turn off the computer, or just hide the computer somewhere while her daughter is at school. But she should have started doing something before her daughter turned 16. If she does not know that, she is clearly unqualified to have children; but there is no law against that.
OK, this was an extreme example, but during my 4-years teaching carreer I have seen or heard from colleagues about many really fucked up parents; and those people were middle and higher social class. This leads me to very pesimistic views, not shared by people who don’t have the same experience and are more free to rationalize this away. I think that if you need parents to do something non-trivial, you should expect at least 20% of population to fail at that. Let’s suppose that in each generation 20% of parents fail to pressure their children to work, in a world where work is not necessary for decent living. What happens in 5 generations?
Personal experience can give different amounts of evidence. If the experience is “me, my family, and people I willingly associate with”, that one is most biased. When you have some kind of job where you interact with people you didn’t choose, that one is a bit better—it is still biased by your personal evaluation, geography, perhaps social class… but at least you are more exposed to people you would otherwise avoid. (For example I usually avoid impolite people from dysfunctional families. As a teacher, they are just there and I have to deal with them. Then I notice that they exist and are actually pretty frequent. When I go outside of the school, they again disappear somehow.)
Of course, I would still prefer having a statistic; I am just not sure if there are available statistics on how many parents completely fail at different tasks, such as not letting their children spend all free time with a computer.
One of these things is not like the others.
Yes, no state has ever implemented truly universal suffrage (among minors).
Or non-humans.
That’s more technically problematic; how could non-human animals vote in the existing kinds of elections? Human intermediaries would have to decide what was best for the non-humans they represented. Different human political factions would support different positions as being best for the non-humans, and fight over that.
(This of course doesn’t apply to future possible non-human sentients like AI, uploads, uplifted animals, modified humans, etc.)
Lead them into the voting booth, see which lever they press.
The same is true of some minors. (Though, of course, not all.)
In Jasay’s terminology, the first is a liberty (a relation between a person and an act) and the rest are rights {relations between two or more persons (at least one rightholder and one obligor) and an act}. I find this distnction useful for thinking more clearly about these kinds of topics. Your mileage may vary.
I was actually referring to the the third being what I might call an anti-liberty, i.e., you aren’t allowed to work more than eight-hours a day, and the fact that is most definitely not enforced nor widely considered a human right.
How is that different from pointing out that you’re not allowed to sell yourself into slavery (not even partially, as in signing a contract to work for ten years and not being able to legally break it), or that you’re not allowed to sell your vote?
I’d say each of the three can be said to be unlike the others:
abolition falls under Liberty
universal suffrage falls under Equality
eight-hour workdays falls under Solidarity
So “all of these things are not like the others”.
I thought eight-hours workdays were about employers not being allowed to demand that employees work more than eight hours a day; I didn’t know you weren’t technically allowed to do that at all even if you’re OK with it.
You are allowed to work more than eight hours per day. It’s just that in many industries, employers must pay you overtime if you do so.
Even if employers were prohibited from using “willingness to work more than 8 hours per day” as a condition for employment, long workdays would probably soon become the norm.
Thus a more feasible way to limit workdays is to constrain employees rather than employers.
To see why, assume that without any restrictions on workday length, workers supply more than 8 hours. Let’s say, without loss of generality, that they supply 10. (In other words, the equilibrium quantity supplied is ten.)
If employers can’t demand the equilibrium quantity, but they’re still willing to pay to get it, then employees will have the incentive to supply it. In their competition for jobs (finding them and keeping them), employees will be supply labor up until the equilibrium quantity, regardless of whether the bosses demand it.
Working more looks good. Everyone knows that; you don’t need your boss to tell you. So if there’s competition for your spot or for a spot that you want, it would serve you well to work more.
So if your goal is to prevent ten-hour days, you’d better stop people from supplying them.
At least, this makes sense to me. But I’m no microeconomist. Perhaps we have one on LW who can state this more clearly (or who can correct any mistakes I’ve made).
See Lochner v. New York. Within the last five years there was a French strike (riot? don’t remember exactly) over a law that would limit the workweek of bakers, which would have the impact of driving small bakeries out of business, since they would need to employ (and pay benefits on) 2 bakers rather than just 1. Perhaps a French LWer remembers more details?
It would be very hard to distinguish when people were doing it because they wanted to, and when employers were demanding it. Maybe some employees are working that extra time, but one isn’t. The one that isn’t happens to be fired later on, for unrelated reasons. How do you determine that worker’s unwillingness to work extra hours is not one of the reasons they were fired? Whether it is or not, that happening will likely encourage workers to go beyond the eight hours, because the last one that didn’t got fired, and a relationship will be drawn whether there is one or not.
It’s not like you can fire employees on a whim: the “unrelated reasons” have to be substantial ones, and it’s not clear you can find ones for any employee you want to fire. (Otherwise, you could use such a mechanism to de facto compel your employees to do pretty much anything you want.)
Also, even if you somehow did manage to de facto demand workers to work ten hours a day, if you have to pay hours beyond the eighth as overtime (with a hourly wage substantially higher than the regular one), then it’s cheaper for you to hire ten people eight hours a day each than eight people ten hours a day.
Under American law, you basically can fire an employee “on a whim” as long as it isn’t a prohibited reason.
Only if they can’t get another job.
That assumption isn’t that far-fetched. Also, the same applies to doing that to compel them to work extra time (or am I missing something?).
If we can afford it.
Moral progress proceeds from economic progress.
Morality is contextual.
If we have four people on a life boat and food for three, morality must provide a mechanism for deciding who gets the food. Suppose that decision is made, then Omega magically provides sufficient food for all—morality hasn’t changed, only the decision that morality calls for.
Technological advancement has certainly caused moral change (consider society after introduction of the Pill). But having more resources does not, in itself, change what we think is right, only what we can actually achieve.
That’s an interesting claim. Are you saying that true moral dilemmas (i.e. a situation where there is no right answer) are impossible? If so, how would you argue for that?
I think they are impossible. Morality can say “no option is right” all it wants, but we still must pick an option, unless the universe segfaults and time freezes upon encountering a dilemma. Whichever decision procedure we use to make that choice (flip a coin?) can count as part of morality.
I take it for granted that faced with a dilemma we must do something, so long as doing nothing counts as doing something. But the question is whether or not there is always a morally right answer. In cases where there isn’t, I suppose we can just pick randomly, but that doesn’t mean we’ve therefore made the right moral decision.
Are we ever damned if we do, and damned if we don’t?
When someone is in a situation like that, they lower their standard for “morally right” and try again. Functional societies avoid putting people in those situations because it’s hard to raise that standard back to it’s previous level.
Well, if all available options are indeed morally wrong, we can still try to see if any are less wrong than others.
Right, but choosing the lesser of two evils is simple enough. That’s not the kind of dilemma I’m talking about. I’m asking whether or not there are wholly undecidable moral problems. Choosing between one evil and a lesser evil is no more difficult than choosing between an evil and a good.
But if you’re saying that in any hypothetical choice, we could always find something significant and decisive, then this is good evidence for the impossibility of moral dilemmas.
It’s hard to say, really.
Suppose we define a “moral dilemma for system X” as a situation in which, under system X, all possible actions are forbidden.
Consider the systems that say “Actions that maximize this (unbounded) utility function are permissible, all others are forbidden.” Then the situation “Name a positive integer, and you get that much utility” is a moral dilemma for those systems; there is no utility maximizing action, so all actions are forbidden and the system cracks. It doesn’t help much if we require the utility function to be bounded; it’s still vulnerable to situations like “Name a real number less than 30, and you get that much utility” because there isn’t a largest real number less than 30. The only way to get around this kind of attack by restricting the utility function is by requiring the range of the function to be a finite set. For example, if you’re a C++ program, your utility might be represented by a 32 bit unsigned integer, so when asked “How much utility do you want” you just answer “2^32 − 1” and when asked “How much utility less than 30.5 do you want” you just answer “30”.
(Ugh, that paragraph was a mess...)
That is an awesome example. I’m absolutely serious about stealing that from you (with your permission).
Do you think this presents a serious problem for utilitarian ethics? It seems like it should, though I guess this situation doesn’t come up all that often.
ETA: Here’s a thought on a reply. Given restrictions like time and knowledge of the names of large numbers, isn’t there in fact a largest number you can name? Something like Graham’s number won’t work (way too small) because you can always add one to it. But trans-finite numbers aren’t made larger by adding one. And likewise with the largest real number under thirty, maybe you can use a function to specify the number? Or if not, just say ’29.999....′ and just say nine as many times as you can before the time runs out (or until you calculate that the utility benefit reaches equilibrium with the costs of saying ‘nine’ over and over for a long time).
Transfinite cardinals aren’t, but transfinite ordinals are. And anyway transfinite cardinals can be made larger by exponentiating them.
Good point. What do you think of Chrono’s dilemma?
“Twenty-nine point nine nine nine nine …” until the effort of saying “nine” again becomes less than the corresponding utility difference. ;-)
Sure, be my guest.
Honestly, I don’t know. Infinities are already a problem, anyway.
My view is that a more meaningful question than ‘is this choice good or bad’ is ‘is this choice better or worse than other choices I could make’.
Would you say that there are true practical dilemmas? Is there ever a situation where, knowing everything you could know about a decision, there isn’t a better choice?
If I know there isn’t a better choice, I just follow my decision. Duh. (Having to choose between losing $500 and losing $490 is equivalent to losing $500 and then having to choose between gaining nothing and gaining $10: yes, the loss will sadden me, but that had better have no effect on my decision, and if it does it’s because of emotional hang-ups I’d rather not have. And replacing dollars with utilons wouldn’t change much.)
So you’re saying that there are no true moral dilemmas (no undecidable moral problems)?
Depends on what you mean by “undecidable”. There may be situations in which it’s hard in practice to decide whether it’s better to do A or to do B, sure, but in principle either A is better, B is better, or the choice doesn’t matter.
So, for example, suppose a situation where a (true) moral system demands both A and B, yet in this situation A and B are incomposssible. Or it forbids both A and B, yet in this situation doing neither is impossible. Those examples have a pretty deontological air to them...could we come up with examples of such dilemmas within consequentialism?
Well, the consequentialist version of a situation that demands A and B is one in which A and B provide equally positive expected consequences and no other option provides consequences that are as good. If A and B are incompossible, I suppose we can call this a moral dilemma if we like.
And, sure, consequentialism provides no tools for choosing between A and B, it merely endorses (A OR B). Which makes it undecidable using just consequentialism.
There are a number of mechanisms for resolving the dilemma that are compatible with a consequentialist perspective, though (e.g., picking one at random).
Thanks, that was helpful. I’d been having a hard time coming up with a consequentialist example.
Then, either the demand/forbiddance is not absolute or the moral system is broken.
How are you defining morality? If we use a shorthand definition that morality is a system that guides proper human action, then any “true moral dilemmas” would be a critique of whatever moral system failed to provide an answer, not proof that “true moral dilemmas” existed.
We have to make some choice. If a moral system stops giving us any useful guidance when faced with sufficiently difficult problems, that simply indicates a problem with the moral system.
ETA: For example, if I have completely strict sense of ethics based upon deontology, I may feel an absolute prohibition on lying and an absolute prohibition on allowing humans to die. That would create an moral dilemma for that system in the classical case of Nazis seeking Jews that I’m hiding in my house. So I’d have to switch to a different ethical system. If I switched to a system of deontology with a value hierarchy, I could conclude that human life has a higher value than telling the truth to governmental authorities under the circumstances and then decide to lie, solving the dilemma.
I strongly suspect that all true moral dilemmas are artifacts of the limitations of distinct moral systems, not morality per se. Since I am skeptical of moral realism, that is all the more the case; if morality can’t tell us how to act, it’s literally useless. We have to have some process for deciding on our actions.
I’m not: I anticipate that your answer to my question will vary on the basis of what you understand morality to be.
Would it? It doesn’t follow from that definition that dilemmas are impossible. This:
Is the claim I’m asking for an argument for.
I’m really confused about the point of this discussion.
The simple answer is: either a moral system cares whether you do action A or action B, preferring one to the other, or it doesn’t. If it does, then the answer to the dilemma is that you should do the action your moral system prefers. If it doesn’t, then you can do either one.
Obviously this simple answer isn’t good enough for you, but why not?
The tricky task is to distinguish between those 3 cases—and to find general rules which can do this in every situation in a unique way, and represent your concept of morality at the same time.
If you can do this, publish it.
Well, yes, finding a simple description of morality is hard. But you seem to be asking if there’s a possibility that it’s in principle impossible to distinguish between these 3 cases for some situation—and this is what you call a “true moral dilemma”—and I don’t see how the idea of that is coherent.
I did not call anything “true moral dilemma”.
Most dilemmas are situations where similar-looking moral guidelines lead to different decisions, or situations where common moral rules are inconsistent or not well-defined. In those cases, it is hard to decide whether the moral system prefers one action or the other, or does not care.
It seems to me to omit a (maybe impossible?) possibility: for example that a moral system cares about whether you do A or B in the sense that it forbids both A and B, and yet ~(A v B) is impossible. My question was just whether or not cases like these were possible, and why or why not.
I admit that I hadn’t thought of moral systems as forbidding options, only as ranking them, in which case that doesn’t come up.
If your morality does have absolute rules like that, there isn’t any reason why those rules wouldn’t come in conflict. But even then, I wouldn’t say “this is a true moral dilemma” so much as “the moral system is self-contradictory”. Not that this is a great help to someone who does discover this about themselves.
Ideally, though, you’d only have one truly absolute rule, and a ranking between the rules, Laws of Robotics style.
So, Kant for example thought that such moral conflicts were impossible, and he would have agreed with you that no moral theory can be both true, and allow for moral conflicts. But it’s not obvious to me that the inference from ‘allows for moral conflict’ to ‘is a false moral theory’ is valid. I don’t have some axe to grind here, I was just curious if anyone had an argument defending that move (or attacking it for that matter).
I don’t think that it means it’s a false moral theory, just an incompletely defined one. In cases where it doesn’t tell you what to do (or, equivalently, tells you that both options are wrong), it’s useless, and a moral theory that did tell you what to do in those cases would be better.
That one thing a couple years ago qualifies.
But unless you get into self-referencing moral problems, no. I can’t think of one off the top of my head, but I suspect that you can find ones among decisions that affect your decision algorithm and decisions where your decision-making algorithm affects the possible outcomes. Probably like Newcomb’s problem, only twistier.
(Warning: this may be basilisk territory.)
(Double-post, sorry)
There are plenty of situations where two choices are equally good or equally bad. This is called “indifference”, not “dilemma”.
Those aren’t the situations I’m talking about.
I would make the more limited claim that the existence of irreconcilable moral conflicts is evidence for moral anti-realism.
In short, if you have a decision process (aka moral system) that can’t resolve a particular problem that is strictly within its scope, you don’t really have a moral system.
Which makes figuring out what we mean by moral change / moral progress incredibly difficult.
This seems to be to be a rephrasing and clarifying of your original claim, which I read as saying something like ‘no true moral theory can allow moral conflicts’. But it’s not yet an argument for this claim.
I’m suddenly concerned that we’re arguing over a definition. It’s very possible to construct a decision procedure that tells one how to decide some, but not all moral questions. It might be that this is the best a moral decision procedure can do. Is it clearer to avoid using the label “moral system” for such a decision procedure?
This is a distraction from my main point, which was that asserting our morality changes when our economic resources change is an atypical way of using the label “morality.”
No, but if I understand what you’ve said, a true moral theory can allow for moral conflict, just because there are moral questions it cannot decide (the fact that you called them ‘moral questions’ leads me to think you think that these questions are moral ones even if a true moral theory can’t decide them).
You’re certainly right, this isn’t relevant to your main point. I was just interested in what I took to be the claim that moral conflicts (i.e. moral problems that are undecidable in a true moral theory) are impossible:
This is a distraction from you main point in at least one other sense: this claim is orthogonal to the claim that morality is not relative to economic conditions.
Yes, you correct that this was not an argument, simply my attempt to gesture at what I meant by the label “morality.” The general issue is that human societies are not rigorous about the use of the label morality. I like my usage because I think it is neutral and specific in meta-ethical disputes like the one we are having. For example, moral realists must determine whether they think “incomplete” moral systems can exist.
But beyond that, I should bow out, because I’m an anti-realist and this debate is between schools of moral realists.
Rephrasing the original question: if we can anticipate the guiding principles underlying the morality of the future, ought we apply those principles to our current circumstances to make decisions, supposing they are different?
Though you seem to be implicitly assuming that the guiding principles don’t change, merely the decisions, and those changed decisions are due to the closest implementable approximation of our guiding principles varying over time based on economic change. (Did I understand that right?)
Pretty much. Though it feels totally different from the inside. Athens could not have thrived without slave labor, and so you find folks arguing that slavery is moral, not just necessary. Since you can’t say “Action A is immoral but economically necessary, so we shall A” you instead say “Action A is moral, here are some great arguments to that effect!”
And when we have enough money, we can even invent new things to be upset about, like vegetable rights.
(nods) Got it.
On your view, is there any attempt at internal coherence?
For example, given an X such that X is equally practical (economically) in an Athenian and post-Athenian economy, and where both Athenians and moderns would agree that X is more “consistent with” slavery than non-slavery, would you expect Athenians to endorse X and moderns to reject it, or would you expect other (non-economic) factors, perhaps random noise, to predominate? (Or some third option?)
Or is such an X incoherent in the first place?
Can you give a more concrete example? I don’t understand your question.
I can’t think of a concrete example that doesn’t introduce derailing specifics.
Let me try a different question that gets at something similar: do you think that all choices a society makes that it describes as “moral” are economic choices in the sense you describe here, or just that some of them are?
Edit: whoops! got TimS and thomblake confused. Um. Unfortunately, that changes nothing of consequence: I still can’t think of a concrete example that doesn’t derail. But my followup question is not actually directed to Tim. Or, rather, ought not have been.
Probably a good counterexample would be the right for certain groups to work any job they’re qualified for, for example women or people with disabilities. Generally, those changes were profitable and would have been at any time society accepted it.
I don’t understand the position you are arguing and I really want to. Either illusion of transparency or I’m an idiot. And TheOtherDave appears to understand you. :(
I’m not really arguing for a position—the grandparent was a counterexample to the general principle I had proposed upthread, since the change was both good and an immediate economic benefit, and it took a very long time to be adopted.
(nods) Yup, that’s one example I was considering, but discarded as too potentially noisy.
But, OK, now that we’re here… if we can agree for the sake of comity that giving women the civil right to work any job would have been economically practical for Athenians, and that they nevertheless didn’t do so, presumably due to some other non-economic factors… I guess my question is, would you find it inconsistent, in that case, to find Athenians arguing that doing so would be immoral?
I don’t think so. I’m pretty sure lots of things can stand in the way of moral progress.
What is progress with respect to either? Could you possibly mean that moral states—the moral conditions of a society—follow from the economic state—the condition and system of economy. I do find it hard to see a clear, unbiased definition of moral or economic progress.
Moral progress is a trend or change for the better in the morality of members of a society. For example, when the United States went from widespread acceptance of slavery to widespread rejection of slavery, that was moral progress on most views of morality.
Economic progress is a trend or change that results in increased wealth for a society.
In general, widespread acceptance of a moral principle, like our views on slavery, animal rights, vegetable rights, and universal minimum income, only comes after we can afford it.
I think he’s trying to say that having resources is a prerequisite to spending them on moral things like universal pay, so we need to pursue wealth if we want to pursue morality. Technically, economic progress is more of a prerequisite to moral progress than a sufficient cause though, as economic progress can also result in bad moral outcomes depending on what we do with our wealth.
What is moral progress? - Is having a society with a vast disparity between rich and poor where the poor support the rich through the resource of their labor considered morally progressed from a more egalitarian tribal state? Is the progress of the empire to a point of collapse and the start of some new empire considered moral progress?
What is economic progress? - Is having a society with a vast disparity between rich and poor where the poor support the rich through the resource of their labor considered morally progressed from the primitive hunter-gatherer society where everyone had more free time considered economic progress? Is the progress of the empire to a point where the disparity in wealth incites revolution or causes collapse considered economic progress?
You’re not making arguments.
The points you raise are not responsive to the points that either he or I made.
If it increases total aggregate utility. Tribes were small, there weren’t very many people. I’m also not sure how happy most tribes were. Additionally, bad moral societies might be necessary to transition to awesome ones.
You conflate moral and economic progress in your second paragraph.
A financial system which collapses probably isn’t too healthy. It still might have improved things overall through its pre-collapse operations though.
Universal pay does not even seem possible now.
You do not answer the question and conflate the questions
How is economic progress measured—if you say the aggraegate utility please explain how that is measured.?
How is moral progress measured?
My argument is simple—the measure of either of these is based on poor heuristics.
My first reaction is to want to say that economic progress is an increase in purchasing power. However, purchasing power is measured with reference to the utility of goods. That would be fine as a solution, except that those definitions would mean that it would be literally impossible for an increase in economic progress to be bad on utilitarian grounds. That’s not what “economic progress” is generally taken to mean, so I won’t use that definition.
Instead, I’ll say that economic progress is an increase in the ability to produce goods, whether those goods are good or bad. This increase can be either numerical or qualitative, I don’t care. Now, it might not be possible to quantify this precisely, but that’s not necessary to determine that economic progress occurs. Clearly, we are now farther economically progressed than we were in the Dark Ages.
Moral progress would be measured depending on the moral theory you’re utilizing. I would use a broad sort of egoism, personally, but most people here would use utilitarianism.
With an egoist framework, you could keep track of how happy or sad you were directly. You could also measure the prevalence of factors that tend to make you happy and then subtract the prevalence of factors that tend to make you sad (while weighting for relative amounts of happiness and sadness, of course), in order to get a more objective account of your own happiness.
With a utilitarian framework, you would measure the prevalence of things that tend to make all people happy, and then subtract the prevalence of things that tend to make all people sad. If there was an increase in the number of happy people, then that would mean moral progress in the eyes of a utilitarian.
You make no argument. You merely ask a question. If you have a general counterargument or want to refute the specifics of any of my points, feel free. So far, you haven’t done anything like that. Also, although it might not be possible to quantify economic or moral progress precisely, we can probably do it well enough for most practical purposes. I don’t understand the purpose of the points you’re trying to raise here.
My original post refuted the statement:
You interjected:
You do not like the questions, the Socratic? Ok, I asserted the basis of the argument and the point of the questions:
A clear, unbiased definition of moral or economic progress does not exist.
You present models for deciding both. There exists models where economic progress varies inversely with moral progress, such as possible outcomes from the utilitarian perspective that are covered in ethics 101 at most colleges, and the manifest reality of a system where economic progress has been used for justifying an abundance of atrocities. There also exist models in either category which define progress in entirely different directions and so any statement of progress is inherently biased.
There is a link between economic states/systems and moral conditions, and it appeared that the author of the statement: “Moral progress proceeds from economic progress.” may have been oversimplifying the issue to a point of of making it unintelligible.
You mentioned wealth which implies an inherent bias also. I can personally assert a different version of wealth which excludes much of what most people consider wealth. If most people think wealth includes assets like cash or gold which I see as having an immoral nature and so their idea of accumulating wealth is immoral in my pov. (I do not include a lengthy moral case, but rather assert such a case exists). So if you see progress and wealth as interrelated then I would ask for a definition of wealth?
You also assert that economic progress is an increased ability to produce goods. I assert that there are many modes of production of which the current industrial mode finds value in quantity, as you state is the measure. Two biases arise:
1 - The bias inherent to the mode: quantity is not the only measure of progress. Competing values include quality in aesthetics, ergonomics, environmental impact, functionality, modular in use (consider open source values). I do not think having more stuff is a sign of economic progress and I am not alone in finding that the measure you have asserted says nothing of “progress”—you of course argue differently and thus we can say one measure or another of progress may differ and are thus inherently biased.
2- What mode of production is more progressed? I do not think industrialization is progress. I see many flaws in the results. Too much damage from that mode imho. I am not here to argue that position but rather to assert it exists.
Is my point about the bias inherent in describing progress clear, or do you think that there exists some definition we all agree upon as to what progress in any area is?
You say that economic production and moral progress aren’t the same. I have already said the same thing; I have already said that increased economic production might lead to morally wrong outcomes depending on how those products end up being used.
You can assert a different definition of wealth if you want, sure. I don’t understand what argument this is supposed to be responsive to. There’s a common understanding of wealth and just because different people define wealth differently, that wouldn’t invalidate my point. Having resources is key to investing them, investing resources is key to doing moral things.
You say that quantity isn’t the sole realm of value. I think that’s true. But if you take the quantity of goods and multiply them by the quality of goods (that is, the utility of the goods, like I mentioned before) then that is a sufficient definition of total economic value.
The mode of production that is most progressed is the one which produces the most.
If we had eight-hour workdays a century ago, we wouldn’t have been able to support the standard of living expected a century ago. I’m not sure we could have even supported living. The same applies to full unemployment. We may someday reach a point where we are productive enough that we can accomplish all we need when we just do it for fun, but if we try that now, we’ll all starve.
Is that true? (Technically, a century ago was 1912.)
Wikipedia on the eight-hour day:
The quote seemed to imply we didn’t have them a century ago. Just use two centuries or however long.
My point is that we didn’t stop working as long because we realized it was a good idea. We did because it became a good idea. What we consider normal now is something we could not have instituted a century ago, and attempting to institute now what what will be normal a century from now would be a bad idea.
So, accepting the premise that the ability to support “full unemployment” (aka, people working for reasons other than money) is something that increases over time, and it can’t be supported until the point is reached where it can be supported… how would we recognize when that point has been reached?
The question is, can we? Does anyone happen to have any empirical data about how good, for example, Greco-Romans were at predicting the moral views of the Middle Ages?
Additionally, is merely sounding “like the kind of lunatic notion that’ll be considered a basic human right in about a century” really a strong enough justification for us to radically alter our political and economic systems? If I had to guess, I’d predict that Kreider already believes divorcing income from work to be a good idea, for reasons that may or may not be rational, and is merely appealing to futurism to justify his bottom line.
Are you sure you can. It’s remarkably easy to make retroactive “predictions”, much harder to make actual predictions.
The way to divorce work from income is the ownership. Be an owner!
One way to divorce work from income is to own stuff.
A more popular way is to find someone else who owns stuff, then take their stuff.
That counts as work.
Not if it’s in the form of “Be poor in a country that taxes the rich to give to the poor”.
No, you just need to own enough stuff to pay workers to take even more stuff from others.
The way to divorce income from work is to pay others to do the work for you? Yes, that works.
Timothy Gowers, Vividness in Mathematics and Narrative, in Circles Disturbed: The Interplay of Mathematics and Narrative
The two examples are not contradictory, but analogous to one another. The correct conclusion in both is the same, and both are equally serious or ironic.
Suppose x² −2=0 has a solution that is rational. That leads to a contradiction. So any solution must be irrational.
Suppose x² +1=0 has a solution that is a number. That leads to a contradiction. So any solution must not be a number. Now what is a “number” in this context? From the text, something that is either positive, negative, or zero; i.e. something with a total ordering. And indeed we know (ETA: this is wrong, see below) that such solutions, the complex numbers, have no total ordering.
I see no relevant difference between the two cases.
You can work the language a little to make them analogous, but that’s not the point Gowers is making. Consider this instead:
“There are those who would believe that all equations have solutions, a view that leads to some intriguing new ideas. Consider the equation x + 1 = x. Inspecting the equation, we see that its solution must be a number which is equal to its successor. Numbers with this remarkable property are quite unlike the numbers we are familiar with. As such, they are surely worthy of further study.”
I imagine Gowers’s point to be that sometimes a contradiction does point to a way in which you can revise your assumptions to gain access to “intriguing new ideas”, but sometimes it just indicates that your assumptions are wrong.
Yes, yes they are.
(Edited again: this example is wrong, and thanks to Kindly for pointing out why. CronoDAS gives a much better answer.)
Curiously enough, the Peano axioms don’t seem to say that S(n)!=n. Lo, a finite model of Peano:
X = {0, 1} Where: 0+0=0; 0+1=1+0=1+1=1 And the usual equality operation.
In this model, x+1=1 has a solution, namely x=1. Not a very interesting model, but it serves to illustrate my point below.
Contradiction in conclusions always indicates a contradiction in assumptions. And you can always use different assumptions to get different, and perhaps non contradictory, conclusions. The usefulness and interest of this varies, of course. But proof by contradiction remains valid even if it gives you an idea about other interesting assumptions you could explore.
And that’s why I feel it’s confusing and counterproductive to use ironic language in one example, and serious proof by contradiction in another, completely analogous example, to indicate that in one case you just said “meh, a contradiction, I was wrong” while in the other you invented a cool new theory with new assumptions. The essence of math is formal language and it doesn’t mix well with irony, the best of which is the kind that not all readers notice.
But that’s the entire point of the quote! That mathematicians cannot afford the use of irony!
Yes. My goal wasn’t to argue with the quote but to improve its argument. The quote said:
And I said, it’s not just superficially similar, it’s exactly the same and there’s no relevant difference between the two that would guide us to use irony in one case and not in the other (or as readers, to perceive irony in one case and serious proof by contradiction in the other).
Your model violates the property that if S(m) = S(n), then m=n, because S(1) = S(0) yet 1 != 0. You might try to patch this by changing the model so it only has 0 as an element, but there is a further axiom that says that 0 is not the successor of any number.
Together, the two axioms used above can be used to show that the natural numbers 0, S(0), S(S(0)), etc. are all distinct. The axiom of induction can be used to show that these are all the natural numbers, so that we can’t have some extra “floating” integer x such that S(x) = x.
Right. Thanks.
The only relevant difference that I can see is that, in the first paragraph, the solutions are explicitly limited to the rational numbers; in the second case, the solutions are not explicitly limited to the reals.
There are lots of total orderings on the complex numbers. For example:
a + bi [>] c + di iff a >= c or (a = c and b >= d).
In fact, if you believe the axiom of choice there are “nice total orders” for any set at all.
Importantly, however, the complex numbers have no total ordering that respects addition and multiplication. In other words, there’s no large set of “positive complex numbers” closed under both operations.
This is also the reason why the math in this XKCD strip doesn’t actually work.
You can still find divisors for Gaussian integers. If x, y, and xy are all Gaussian integers, which will be trivially fulfilled for any x when y=1, then x, y both divide xy.
You can then generalize the \sigma function by summing over all the divisors of z and dividing by |z|.
The resulting number \sigma(z) lies in C (or maybe Q + iQ), not just Q, but it’s perfectly well defined.
If you sum over all the divisors of z, the result is perfectly well defined; however, it’s 0. Whenever x divides z, so does -x.
Over the integers, this is solved by summing over all positive divisors. However, there’s no canonical choice of what divisors to consider positive in the case of Gaussian integers, and making various arbitrary choices (like summing over all divisors in the upper half-plane) leads to unsatisfying results.
That’s like saying the standard choice of branch cut for the complex logarithm is arbitrary.
And?
When you complexify, things get messier. My point is that making a generalization is possible (though it’s probably best to sum over integers with 0 \leq arg(z) < \pi, as you pointed out), which is the only claim I’m interested in disputing. Whether it’s nice to look at is irrelevant to whether it’s functional enough to be punnable.
You’re right—the generalization works.
Mainly what I don’t like about it is that \sigma(z) no longer has the nice properties it had over the integers: for example, it’s no longer multiplicative. This doesn’t stop Gaussian integers from being friendly, though, and the rest is a matter of aesthetics.
The well-ordering principle doesn’t really have any effect on canonical orderings, like that induced by the traditional less-than relation on the real numbers.
This doesn’t affect the truth of your claim, but I do think that DanArmak’s point was quite separate from the language he chose. He might instead have worded it as having no real solution, so that any solution must be not-real.
Gah. You’re quite right. I should refrain from making rash mathematical statements. Thank you.
The first one shows that assuming that there’s a rational solution leads to contradiction, then drops the subject. The second one shows that assuming that there’s a real solution leads to a contradiction, then suggests to investigate the non-reals. How are you supposed to tell which drops the subject and which suggests investigation?
Isn’t that the entire point? I see this as a mathematical version of the modus tollens/ponens point made elsewhere in this page.
The quote says,
This seems to me to mean: the two cases are different; the first is appropriately handled by serious proof-by-contradiction, while the second is appropriately handled by irony. But readers may not be able to tell the difference, because the two texts are similar and irony is hard to identify reliably. So mathematicians should not use irony.
Whereas I would say: the two cases are the same, and irony or seriousness are equally appropriate to both. If readers could reliably identify irony, they would correctly deduce that the author treated the two cases differently, which is in fact a wrong approach. So readers are better served by treating both texts as serious.
I’m not saying mathematicians should / can effectively use irony; I’m saying the example is flawed so that it doesn’t demonstrate the problems with irony.
The difference is that mathematicians apply modus tollens and reject sqrt2 being rational, but apply modus ponens and accept the existence of i; why? Because apparently the resultant extensions of theories justify this choice—and this is the irony, the reason one’s beliefs are in discordance with one’s words/proof and the reader is expected to appreciate this discrepancy.
But what one regards as a useful enough extension to justify a modus tollens move is something other may not appreciate or will differ from field to field, and this is a barrier to understanding.
I hadn’t considered that irony. I was thinking about the explicit irony of the text itself in its proof of sqrt(2) being irrational. The reader is expected to know the punchline, that sqrt(2) is irrational but that irrational numbers are important and useful. So the text that (ironically) appears to dismiss the concept of irrational numbers is in fact wrong in its dismissal, and that is a meta-irony.
...I feel confused by the meta levels of irony. Which strengthens my belief that mathematical proofs should not be ironical if undertaken seriously.
Yes, I feel similarly about this modus stuff; it seems simple and trivial, but the applications become increasingly subtle and challenging, especially when people aren’t being explicit about the exact reasoning.
If mathematicians behaved simply as you describe, then those resultant extension theories would never have been developed, because everyone would have applied modus tollens regarding in a not-yet-proven-useful case. (Disclaimer: I know nothing about the actual historical reasons for the first explorations of complex numbers.)
Therefore, it’s best for mathematicians to always keep the M-T and M-P cases in mind when using a proof by contradiction. Of course, a lot of time the contradiction arises due to theorems already proven from axioms, and what happens if any one of the axioms in a theory is removed is usually well explored.
You’re drawing the parallel differently from the quote’s author. The second example requires assuming the existence of complex numbers to resolve the contradiction. The first example requires assuming, not the existence of irrational numbers (we already know about those, or we wouldn’t be asking the question!), but the existence of integers which are both even and odd. As far as I know, there are no completely satisfactory ways of resolving the latter situation.
In the first case, starting with p such that the highest power of 2 that divides p is an integer power of 2 (2^k for some integer k); then the highest power of 2 that divides p² is 2^2k; then the highest power of 2 that divides 2q² is also 2^2k; then the highest power of 2 that divides q is 2^(2k-1); therefore q must be a multiple of 2^(k-0.5); a noninteger power of 2.
This implies that there is a number 2^(0.5). It makes no claims as to whether or not this number is rational, or integer; it merely claims that such a number must exist. (Consider: if I had started instead with the equation x²-4=0, I would have ended up showing that a number of the form 4^(0.5) must exist—that number is rational, is indeed an integer).
Now, I think I can prove that an integer q which is a multiple of 2^(k-0.5) but which is not a multiple of 2^k, for integer k, does not exist; but I can only complete that proof by knowing in advance that 2^0.5 is irrational, so I can’t use it to prove the irrationality of 2^0.5. I can easily prove that a rational number of the form 4^(k-0.5) for integer k does exist; indeed, an infinite number of such numbers exist (examples include 2, 8, 32).
No matter how forcefully that first passage conveys the irrationality of √2, it does not prove it.
The paragraph, of course, was talking about integer powers of 2 that divide p. As in, the largest number 2^k such that 2^k divides p and k is an integer.
The largest real power of 2 that divides p is, of course, p itself, as 2^log_2(p) = p.
Looking over my post again, after a good night’s sleep, I see that it wasn’t as coherent as it appeared to me yesterday. Let me see if I can put my point a little more clearly.
The paragraph centers its claim of the irrationality of √2 on the idea that p² contains exactly twice as many powers of 2 as p does. But that is only true because √2 is irrational, making the demonstration a circular proof.
Consider. If √2 were rational, in the form of z/y for some coprime integers z and y, then it would be easy to find an integer that is not itself an integer power of 2, but whose square is an integer power of 2; z would be such a number.
The proof assumes unique prime factorization. Once we factor p as 2^a 3^b 5^c etc., then we know that p^2 factors as 2^(2a) 3^(2b) 5^(2c) etc. This is (implicitly) how we know that p^2 contains exactly twice as many powers of 2 as p.
If √2 is rational, then √2 can be written as z/y for some integers z and y, where z and y are coprime. Then, 2=z²/y².
Consider the hypothetical integer z. It is equal to √2*y. Since y and z are coprime, y cannot contain a factor of √2. Thus, z does not contain a factor of 2; the highest integer power of 2 that is a factor of z is 2^0.
On the other hand, z² does have a factor of 2; it is equal to 2*y² (since y has no factor of √2, y² therefore has no factor of 2).
Therefore, to claim that p² contains exactly twice as many powers of 2 as p is exactly equivalent to claiming that √2 is irrational.
Even if √2 is rational, it is not an integer, and (especially in your second paragraph) you are trying to do things with it that only make sense with integers.
I don’t think that continuing this conversation can be productive, since you seem to be objecting to standard techniques. A rigorous textbook in number theory can probably explain the proof that √2 is irrational more thoroughly, and you can follow back the lemmas and see exactly where your confusion lies.
Isn’t the point of math that all mathematical truths are logically equivalent? (In beore Gödel.)
Depends on the axioms you’re using.
Well, there’s an informal notion that if two theorems T1 and T2 are both true, yet T1 ⇔ T2 is much easier to prove than either T1 and T2, then the two are equivalent. (There’s also the formal notion that two axioms are equivalent if assuming either one lets you prove the other, but I don’t think that’s especially relevant here.)
That’s not close to being the most confused part of the comment you’re replying to.
Notice that you are claiming that all possible proofs of the statement “p^2 contains twice as many powers of 2 as p” require asserting without proof that sqrt(2) is irrational.
Why does the prime factorization of integers depend upon something that is, if not irrational, at least certainly not an integer? (Proof: 1^2 = 1, 2^2 = 4, and x ⇐ x^2 by induction.)
--Game of Thrones, Season 2.
Reminds me of Patton:
I especially like the way he calls the enemy “the other poor bastard”. And not, say, “the bastard”.
Also effort, expertise, and insider information on one of the most powerful Houses around. And magic powers.
He has magic powers?
Rot13′d for minor spoiling potential: Ur’f n jnet / fxvapunatre.
-- Aristosophy (again)
Charles Handy describing the Vietnam-era measurement policies of Secretary of Defense Robert McNamara
The following quotes were heavily upvoted, but then turned out to be made by a Will Newsome sockpuppet who edited the quote afterward. The original comments have been banned. The quotes are as follows:
— Aristosophy
— Aristosophy
If anyone objects to this policy response, please PM me so as to not feed the troll.
Defection too far. Ban Will.
Will is a cute troll.
Hmm, after observing it a few times on various forums I’m starting to consider that having a known, benign resident troll might keep away more destructive ones. No idea how it works but it doesn’t seem that far-fetched given all the strange territoriality-like phenomena occasionally encountered in the oddest places.
I’ve heard this claimed.
This behavior isn’t cute.
This would be somewhat in fitting with findings in Cialdini. One defector kept around and visibly punished or otherwise looking low status is effective at preventing that kind of behavior. (If not Cialdini, then Greene. Probably both.)
Yes it is, and not just a little bit.
The deliberate sabotage of threads? How cute will it be if he destroys the whole forum?
It’s really mean to say someone isn’t cute and although this entire thread isn’t very productive I find it mean that my comment rejecting the meanness to WN was selectively deleted.
Alternately, it is toxic to describe trolling behavior as ‘cute’ when it isn’t, and hasn’t been either cute or particularly witty or intelligent in a long time. This. Behavior. Is. Not. Cute.. It is lame.
I’d rather live in a world where even if we disagree with each other, annoy each other, or waste each other’s time we still don’t say anybody isn’t cute.
The opposite of cute is disgusting and is not a concept that should be applied to humans.
Example 54084954 of that fact that true-seeking and politeness are not correlated.
Also, a little fallacy of gray. Someone could be zero on the cute/disgusting scale, if even if it were so awful to label them disgusting.
Cuteness is a subjective evaluation, a way to interpret reality, not a fact.
There is a difference between rejecting a “Will is a cute troll” meme being used to justify sock-puppet bait-and-switch abuse—by specifically referring to the behavior being not-cute—and simply saying that someone is not cute apropos of nothing. Your equivocation is either disingenuous or just silly.
I have begun to suspect that Incorrect is a Will sockpuppet. Please cease to feed.
The thought crossed my mind when the edit to the ancestor made it clear that Incorrect was trolling rather than well meaning yet confused. I clicked “close” rather than “comment” on the “It looks like Incorrect is Will, I’m not going to feed him here or elsewhere”, for obvious reasons.
Please exterminate.
Here’s a conversation I had with Will a while back:
http://lesswrong.com/lw/cw1/open_problems_related_to_solomonoff_induction/6rlr?context=1#6rlr
But surely you agree that tricking people into saying “I think Will is Incorrect” is exactly the sort of thing that would amuse him?
This had better not start a trend of suspecting people with adjectival usernames to be sockpuppets.
Yours could also be interpreted as an adverb.
Even calling someone’s behavior non-cute is mean. Even meanness is cute. Once you start calling humans or the things they do non-cute you open the door to finding humans disgusting.
Even if we were to assume his behavior was trollish, damaging to lesswrong, and/or unproductive, that shouldn’t make it non-cute.
By that same argument murder is cute, rape is cute, arson is cute, genocide is cute—and you prefer to live in a world where people call these things cute than in a world where they call them non-cute.
You’re using the word “cute” wrongly.
Edited how?
If I remember correctly the second quote was edited to be something along the lines of “will_newsome is awesome.”
That is cute.. no? More childish than evil. He should just be warned that’s trolling.
There really should be a comment edit history feature. Maybe it only activates once a comment reaches +10 karma.
It was edited to add something like “Will Newsome is such a badass”—Socrates
I do find some of Will Newsome’s contributions interesting. OTOH, this behaviour is pretty fucked up. (I was wondering how hard it would be to implement a software feature to show the edit history of comments.)
If only the converse were true...
″...if you lack a thousand-year-old brain that can make trillion-year plans, dying after a billion years doesn’t sound sad to you”?
I’m confused as to what you’re trying to say. Are you saying that dying after a billion years sounds sad to you?
“If you lack a thousand-year-old brain that can make trillion-year plans, it’s because dying after a billion years doesn’t sound sad to you.”
I think meaning it’s unfortunate that thinking that dying after a billion years is sad doesn’t by itself give you the power to live that long. Maybe.
I was never one for formal logic, but isn’t that the contrapositive? I was under the impression that the converse of p then q was q then p.
Yes and that’s what nshepperd wrote.
Oh wow, never mind. My brain was temporarily broken. Is it considered bad etiquette here to retract incorrect comments?
When you retract the comment is simply struck-through not deleted, so no.
And therefore you would have a thousand-year-old brain that can make trillion-year plans.
Seems legit.
Louis C.K., on Reddit
Unfortunately, doing bad shows is not only a route to doing good shows.
True, and I hope no one thinks it is. So we can conclude that doing bad shows at first is not a strong indicator of whether you have a future as a showman.
I guess I see the quote as being directed at people who are so afraid of doing a bad show that they’ll never get in enough practice to do a good show. Or they practice by, say, filming themselves telling jokes in their basement and getting critiques from their friends who will not be too mean to them. In either case, they never get the amount of feedback they would need to become good. For such a person to hear “Yes, you will fail” can be oddly liberating, since it turns failure into something accounted for in their longer-term plans.
--George R. R. Martin, A Game of Thrones
I’m surprised at how often I have to inform people of this… I have mild scoliosis, and so I usually prefer sitting down and kicking up my feet, usually with my work in hand. Coming from a family who appreciates backbreaking work is rough when the hard work is even harder and the pain longer-lasting… which would be slightly more bearable if the aforementioned family did not see reading MYSTERIOUS TEXTS on a Kindle and using computers for MYSTERIOUS PURPOSES as signs of laziness and devotion to silly frivolities.
I have a sneaking suspicion that this is not a very new situation.
I think the quote could be trimmed to its last couple sentences and still maintain the relevant point..
I disagree, in fact. That books strengthen the mind is baldly asserted, not supported, by this quote—the rationality point I see in it is related to comparative advantage.
Oh, totally. But I prefer the full version; it’s really a beautifully written passage.
Alexander Grothendieck
...screw it, I’m not growing up.
I remember being very much afraid of all those things as a child. I’m getting better now.
(Simon Blackburn, Truth)
The pithiest definition of Blackburn’s minimalism I’ve read is in his review of Nagel’s The Last Word:
It is followed by an even pithier response to how Nagel refutes relativism (pointing that our first-order conviction that 2+2=4 or that murder is wrong is more certain than any relativist doubts) and thinks that this establishes a quasi-Platonic absolutism as the only alternative:
“What is truth” is a pretty good question, though a better one is “what do we do with truths?”
We do a lot of things with truths, it can serve a lot of different functions. The problem comes where people doing different things with their truths talk to each other.
“Nontrivial measure or it didn’t happen.” —Aristosophy
(Who’s Kate Evans? Do we know her? Aristosophy seems to have rather a lot of good quotes.)
*cough*
“I made my walled garden safe against intruders and now it’s just a walled wall.” -- Aristosophy
Is that you? That’s ingenious.
For more rational flavor:
This should be the summary for entangled truths:
how to seem and be deep:
Dark Arts:
More Dark arts:
Luminosity:
No, I’m not her. I don’t know who she is, but her Twitter is indeed glorious. (And Google Reader won’t let me subscribe to it the way I’m subscribed to other Twitters, rar.)
She’s got to be from here, here’s learning biases can hurt people:
Cryonics:
I’m starting to think this is someone I used to know from tvtropes.
Twitter RSS feed
I can confirm this works for me, in Google Reader (and my thanks to you also).
I found that, but it won’t let me subscribe to it with Google Reader, only with other things I don’t use.
This worked for me in Google Reader: http://api.twitter.com/1/statuses/user_timeline.rss?screen_name=aristosophy
I get the same results as the other one above.
Weird. You can try the last solution here: http://saravananthirumuruganathan.wordpress.com/2011/05/29/how-to-follow-a-twitter-account-feed-using-rss-reader-in-new-twitter/. I didn’t need it this time, but it worked the last time I was having Twitter + Reader problems.
I found that by Googling before mentioning it here and tried all three things. They don’t work. I’m super-confused.
That’s odd—I did subscribe to it with Google Reader, right before I posted the link.
My bookmarklet says “can’t find a feed”, and the dropdown menu doesn’t offer Google Reader as far as I can tell. How did you do it?
“Google” was auto-selected in my dropdown menu, so it was straightforward for me, same as always. Two clicks, one on Subscribe, the second to indicate Google Reader rather than Google Homepage.
Not sure how much troubleshooting help I can give you. Does that page at least show the recent tweets? Are you logged into Google? Maybe try going to your Google Reader page and entering the url there in the subscribe-to-a-new-feed place?
Yes.
Yes.
Doesn’t work, says it can’t find it. I don’t know why; this is how I’ve subscribed to Twitters in the past.
hey guys fun fact vivid is me
Leo Tolstoy, “War and Peace”, trans. Pevear and Volokhonsky
“Possibly the best statistical graph ever drawn” http://www.edwardtufte.com/tufte/posters
“Junior”, FIRE JOE MORGAN
Subway ad: “146 people were hit by trains in 2011. 47 were killed.”
Guy on Subway: “That tells me getting hit by a train ain’t that dangerous.”
Nate Silver, on his Twitter feed @fivethirtyeight
This reminds me of how I felt when I learned that a third of the passengers of the Hindenburg survived. Went something like this, if I recall:
Actually, according to Wikipedia, only 35 out of the 97 people aboard were killed. Not enough to kill even 50% of them.
jaw drops
It helps to remember that the Hindenburg was more or less parked when it exploded… I think it was like 30 feet in the air? (I’m probably wrong about the number, but I don’t think I’m very wrong.) Most of the passengers basically jumped off. And, sure, a 30 foot drop is no walk in the park, but it’s not that surprising that most people survive it.
(Well, then “out of the sky” is kind of an exaggeration, since you wouldn’t normally consider yourself to be in the sky when on a balcony on the fourth floor.)
Well, unlike the balcony of a building, a floating blimp (even close to the ground) is floating, rather than resting on the ground, so I suppose one could make the argument. But yeah, I’m inclined to agree that wherever “the sky” is understood to be, and I accept that this is a social construct rather than a physical entity, it’s at least a hundred feet or so above ground
Wait, 32% probability of dying “ain’t that dangerous”? Are you f***ing kidding me?
If I expect to be hit by a train, I certainly don’t expect a ~68% survival chance. Not intuitively, anyways.
I’m guessing that even if you survive, your quality of life is going to take a hit. Accounting for this will probably bring our intuitive expectation of harm closer to the actual harm.
Hmmm, I can’t think of any way of figuring out what probability I would have guessed if I had to guess before reading that. Damn you, hindsight bias!
(Maybe you could spell out and rot-13 the second figure in the ad...)
I would expect something like that chance. Being hit by a train will be very similar to landing on your side or back after falling 3 to 10 meters (I’m guessing most people hit by trains are at or near a train station, so the impacts will be relatively slow). So the fatality rate should be similar.
Of course, that prediction gives a fatality rate of only 5-20%, so I’m probably missing something.
There’s the whole crushing and high voltage shock thing, depending on how you land.
Well, lightning strikes kill less than half the people they hit.
Lightning strikes usually do not involve physical impacts—I think “falling from 3-10 meters and getting struck by lightning” would be worse. In addition, the length of the current flow depends on the high voltage system.
This seems overwhelmingly likely.
I can’t help but think:
Subway ad: “146 people were hit by trains in 2011. 47 were killed.”
Guy at Subway: “What does that have to do with sandwiches?”
“If your plan is for one year plant rice. If your plan is for 10 years plant trees. If your plan is for 100 years educate children”—Confucius
...If your plan is for eternity, invent FAI?
Depends how you interpret the proverb. If you told me the Earth would last a hundred years, it would increase the immediate priority of CFAR and decrease that of SIAI. It’s a moot point since the Earth won’t last a hundred years.
Sorry, Earth won’t last a hundred years?
Nanotech and/or UFAI.
The idea seems to be that even if there is a friendly singularity, Earth will be turned into computronium or otherwise transformed.
I am surprised that this claim surprises you. A big part of SI’s claimed value proposition is the idea that humanity is on the cusp of developing technologies that will kill us all if not implemented in specific ways that non-SI folk don’t take seriously enough.
Of course you’re right. I guess I haven’t noticed the topic come up here for a while, and haven’t seen the apocalypse predicted so straightforwardly (and quantitatively) before so am surprised in spite of myself.
Although, in context, it sounds like EY is saying that the apocalypse is so inevitable that there’s no need to make plans for the alternative. Is that really the consensus at EY’s institute?
I have no idea what the consensus at SI is.
I guess he means “only last a hundred years”, not “last at least a hundred years”.
Just to make sure I understand: you interpret EY to be saying that the Earth will last more than a hundred years, not saying that the Earth will fail to last more than a hundred years. Yes?
If so, can you clarify how you arrive at that interpretation?
“If you told me the Earth would only last a hundred years (i.e. won’t last longer than that) …. It’s a moot point since the Earth won’t only last a hundred year (i.e. it will last longer).” At least that’s what I got on the first reading.
I think I could kind-of make sense “it would increase the immediate priority of CFAR and decrease that of SIAI” under either hypothesis about what he means, though one interpretation would need to be more strained than the other.
The idea is that if Earth lasts at least a hundred years, (if that’s a given), then the possibility of a uFAI in that timespan severely decreases—so SIAI (which seeks to prevent a uFAI by building a FAI) is less of an immediate priority and it becomes a higher priority to develop CFAR that will increase the public’s rationality for the future generations, so that the future generations don’t launch a uFAI.
(The other interpretation would be “If the Earth is going to only last a hundred years, then there’s not much point in trying to make a FAI since in the long-term we’re screwed anyway, and raising the sanity waterline will make us enjoy more what time there is left.)
EDIT: Also, if your interpretation is correct, by saying that the Earth won’t last 100 years he’s either admitting defeat (i.e. saying that an uFAI will be built) or saying that even a FAI would destroy the Earth within 100 years (which sounds unlikely to me—even if the CEV of humanity would eventually want to do that, I guess it would take more than 100 years to terraform another place for us to live and for us all to move there).
I was just using “Earth” as a synonym for “the world as we know it”.
I think I disagree; care to make it precise enough to bet on? I’m expecting life still around, Earth the main population center, most humans not uploaded, some people dying of disease or old age or in wars, most people performing dispreferred activities in exchange for scarce resources at least a couple months in their lives, most children coming out of a biological parent and not allowed to take major decisions for themselves for at least a decade.
I’m offering $100 at even odds right now and will probably want to bet again in the next few years. I can give it to you (if you’re going to transfer it to SIAI/CFAR tell me and I’ll donate directly), and you pay me $200 if the world has not ended in 100 years as soon as we’re both available (e.g. thawed). If you die you can keep the money; if I die then win give it to some sensible charity.
How’s that sound? All of the above is up for negotiation.
As wedifrid says, this is a no-brainer “accept” (including the purchasing-power-adjusted caveat). If you are inside the US and itemize deductions, please donate to SIAI, otherwise I’ll accept via Paypal. Your implied annual interest rate assuming a 100% probability of winning is 0.7% (plus inflation adjustment). Please let me know whether you decide to go through with it; withdrawal is completely understandable—I have no particular desire for money at the cost of forcing someone else to go through with a bet they feel uncomfortable about. (Or rather, my desire for $100 is not this strong—I would probably find $100,000 much more tempting.)
PayPal-ed to sentience at pobox dot com.
Don’t worry, my only debitor who pays higher interest rates than that is my bank. As long as that’s not my main liquidity bottleneck I’m happy to follow medieval morality on lending.
If you publish transaction data to confirm the bet, please remove my legal name.
Bet received. I feel vaguely guilty and am reminding myself hard that money in my Paypal account is hopefully a good thing from a consequentialist standpoint.
Bet recorded: LW bet registry, PB.com.
(Neglecting any logistic or legal isses) this sounds like a no brainer for Eliezer (accept).
Like you would be better served by making the amounts you give and expect to receive if you win somewhat more proportionate to expected utility of the resources at the time. If Eliezer was sure he was going to lose he should still take the low interest loan.
Even once the above is accounted for Eliezer should still accept the bet (in principle).
Dollar amounts are meant as purchasing-power-adjusted. I am sticking my fingers in my ears and chanting “La la, can’t hear you” at discounting effects.
That’s a nice set of criteria by which to distinguish various futures (and futurists).
Care to explain why? You sound like you expect nanotech by then.
I definitely expect nanotech a few orders of magnitude awesomer than we have now. I expect great progress on aging and disease, and wouldn’t be floored by them being solved in theory (though it does sound hard). What I don’t expect is worldwide deployment. There are still people dying from measles, when in any halfway-developed country every baby gets an MMR shot as a matter of course. I wouldn’t be too surprised if everyone who can afford basic care in rich countries was immortal while thousands of brown kids kept drinking poo water and dying. I also expect longevity treatments to be long-term, not permanent fixes, and thus hard to access in poor or politically unstable countries.
The above requires poor countries to continue existing. I expect great progress, but not abolition of poverty. If development continues the way it has (e.g. Brazil), a century isn’t quite enough for Somalia to get its act together. If there’s a game-changing, universally available advance that bumps everyone to cutting-edge tech levels (or even 2012 tech levels), then I won’t regret that $100 much.
I have no idea what wars will look like, but I don’t expect them to be nonexistent or nonlethal. Given no game-changer, socioeconomic factors vary too slowly to remove incentive for war. Straightforward tech applications (get a superweapon, get a superdefense, give everyone a superweapon, etc.) get you very different war strategies, but not world peace. If you do something really clever like world government nobody’s unhappy with, arms-race-proof shields for everyone, or mass Gandhification, then I have happily lost.
Thanks for explaining!
Of course, nanotech could be self replicating and thus exponentially cheap, but the likelihood of that is … debatable.
I feel an REM song coming on...
(I guess I had been primed to take “Earth” to mean ‘a planet or dwarf planet (according to the current IAU definition) orbiting the Sun between Venus and Mars’ by this. EDIT: Dragon Ball too, where destroying a planet means turning it into dust, not just rendering it inhabitable.)
EY does seem in a darker mood than usual lately, so it wouldn’t surprise me to see him implying pessimism about our chances out loud, even if it doesn’t go so far as “admitting defeat”. I do hope it’s just a mood, rather than that he has rationally updated his estimation of our chances of survival to be even lower than they already were. :-)
“The world as we know it” ends if FAI is released into the wild.
When I had commented, EY hadn’t clarified yet that by Earth he meant “the world as we know it”, so I didn’t expect “Earth” to exclude ‘the planet between Venus and Mars 50 years after a FAI is started on it’.
50 years after a self-improving AI is released into the wild, I don’t expect Venus and Mars to be in their present orbits. I expect that they would be gradually moving towards being in the same orbit that the Earth is moving towards (or is already established in) 120 degrees apart, propelled by a rocket which uses large reflectors in space to heat portion of the surface of the planet, which is then forced to jet in the desired vector at escape velocity. ETA: That would mean the removal of three objects from the list of planets of Sol.
I think it will only be a few hundred years after FAI before interplanetary travel requires routine ‘take your shoes off’ type of screening.
We’ll still have shoes? And terrorists? I’m disappointed in advance.
And even the right and ability (if we currently have it) to make choices, and some privacy!
IMHO you’re being provincial. Your intuitions for interplanetary travel come directly from flying in the US; if you were used to saner policies you’d make different predictions. (If you’re not from North America, I am very confused.)
Your idea of provincialism is provincial. The idea of shipping tinned apes around the solar system is the true failure of vision here, nevermind the bag check procedures.
How quickly do you think humans will give up commuting?
Why would you put them into an inherently dynamically-unstable configuration, position-corrected by a massive kludge? I mean, what’s in it for the AI?
How about a dynamically stable one?
Oh, and roughly ten to twenty times the total available living space for humans, at an order-of-magnitude guess.
If the AI is Friendly? The enhancement of humanity’s utility/happiness/wealth—I assume terraforming is a lot easier if planets are near the middle of the water zone.
We don’t know what it takes to terraform a world—it’s easy to go “well, it needs more water and air for starters,” but that conceals an awful lot of complexity. Humans, talking populations thereof, can’t live just anywhere. We don’t even have a really good, working definition of what the “habitability” of a planet is, in a way that’s more specific than “I knows it when I sees it.” Most of the Earth requires direct cultural adaptation to be truly livable. There’s no such thing as humans who don’t use culture and technology to cope with the challenges posed by their environment.
Anyway, my point is more that your prediction suggests some cached premises: why should FAI do that particular thing? Why is that a more likely outcome than any of the myriad other possibilities?
I specifically mentioned that Earth’s orbit would also be optimized- although the solar-powered jet engine concept has bigger downsides when used on an inhabited planet.
Do distinct planets necessarily have distinct orbits?
According to the modern definition, yes.
Ah! I had read the wiki article on planets, which said “and has cleared its neighbouring region of planetesimals,” and didn’t bother to look up primary sources. I should know better. Thanks!
Not thinking very ambitious I see.
That’s on the five-millennium plan.
So, we can construct an argument that CFAR would rise in relative importance over SIAIif we see strong evidence the world as we know it will end within 100 years, and an argument with the same conclusion if we see strong evidence that the world as we know it will last for at least 100 years.
There is something wrong.
Imām al-Ḥaddād (trans. Moṣṭafā al-Badawī), “The Sublime Treasures: Answers to Sufi Questions”
Reminds me of Moore’s “here is a hand” paradox (or one man’s modus tollens is another’s modus ponens).
Richard Carrier on solipsism, but not nearly as pithy:
I think that’s actually a really terrible bit of arguing.
We can stop right there. If we’re all the way back at solipsism, we haven’t even gotten to defining concepts like ‘random chance’ or ‘design’, which presume an entire raft of external beliefs and assumptions, and we surely cannot immediately say there are only two categories unless, in response to any criticism, we’re going to include a hell of a lot under one of those two rubrics. Which probability are we going to use, anyway? There are many more formalized versions than just Kolmogorov’s axioms (which brings us to the analytic and synthetic problem).
And much of the rest goes on in a materialist vein which itself requires a lot of further justification (why can’t minds be ontologically simple elements? Oh, your experience in the real world with various regularities has persuaded you that is inconsistent with the evidence? I see...) Even if we granted his claims about complexity, why do we care about complexity? And so on.
Yes, if you’re going to buy into a (very large) number of materialist non-solipsist claims, then you’re going to have trouble making a case in such terms for solipsism. But if you’ve bought all those materialist or externalist claims, you’ve already rejected solipsism and there’s no tension in the first place. And he doesn’t do a good case of explaining that at all.
Good points, but then likewise how do you define and import the designations of ‘hand’ or ‘here’ and justify intuitions or a axiomatic system of logic (and I understood Carrier to be referring to epistemic solipsism like Moore—you seem to be going metaphysical)? (or were you not referring to Moore’s argument in the context of skepticism?)
I think Moore’s basic argument works on the level of epistemic skepticism, yes, but also metaphysics: some sort of regular metaphysics and externalism is what one believes, and what provides the grist for the philosophical mill. If you don’t credit the regular metaphysics, then why do you credit the reasoning and arguments which led you to the more exotic metaphysics?
I’m not sure what skeptical arguments it doesn’t work for. I think it may stop at the epistemic level, but that may just be because I’m having a hard time thinking of any ethics examples (which is my usual interest on the next level down of abstraction).
The way I see it, Moore’s argument gets you to where you’re uncertain of the reasoning pro or contra skepticism. But If you start from the position of epistemic solipsism (I know my own mind, but I’m uncertain of the external world), then you have reason (more or less depending how uncertain you are) to side with common sense. However, if you start at metaphysical solipsism (I’m uncertain of my own mind), then such an argument could even be reason to not side with common sense (e.g., there are little people in my mind trying to manipulate my beliefs; I must not allow them to).
A hypothesis like… I’m dreaming.
This also made me think of the aphorism “if water sticks in your throat, with what will you wash it down?”
Or “if salt loses its savor”, although I wonder if they’re really making the same philosophical point about relative weights of evidence on two sides of a contradiction/paradox.
“In a society in which the narrow pursuit of material self-interest is the norm, the shift to an ethical stance is more radical than many people realize. In comparison with the needs of people starving in Somalia, the desire to sample the wines of the leading French vineyards pales into insignificance. Judged against the suffering of immobilized rabbits having shampoos dripped into their eyes, a better shampoo becomes an unworthy goal. An ethical approach to life does not forbid having fun or enjoying food and wine, but it changes our sense of priorities. The effort and expense put into buying fashionable clothes, the endless search for more and more refined gastronomic pleasures, the astonishing additional expense that marks out the prestige car market in cars from the market in cars for people who just want a reliable means to getting from A to B, all these become disproportionate to people who can shift perspective long enough to take themselves, at least for a time, out of the spotlight. If a higher ethical consciousness spreads, it will utterly change the society in which we live.” —Peter Singer
As it is probably intended, the more reminders like this I read, the more ethical I should become. As it actually works, the more of this I read, the less I become interested in ethics. Maybe I am extraordinarily selfish and this effect doesn’t happen to most, but it should be at least considered that constant preaching of moral duties can have counterproductive results.
I suspect it’s because authors of “ethical remainders” are usually very bad at understanding human nature.
What they essentially do is associate “ethical” with “unpleasant”, because as long as you have some pleasure, you are obviously not ethical enough; you could do better by giving up some more pleasure, and it’s bad that you refuse to do so. The attention is drawn away from good things you are really doing, to the hypothetical good things you are not doing.
But humans are usually driven by small incentives, by short-term feelings. The best thing our rationality can do is better align these short-term feelings with out long-term goals, so we actually feel happy when contributing to our long-term goals. And how exactly are these “ethical remainders” contributing to the process? Mostly by undercutting your short-term ethical motivators, by always reminding you that what you did was not enough, therefore you don’t deserve the feelings of satisfaction. Gradually they turn these motivators off, and you no longer feel like doing anything ethical, because they convinced you (your “elephant”) that you can’t.
Ethics without understanding human nature is just a pile of horseshit. Of course that does not prevent other people from admiring those who speak it.
Yes. And it works this way even without insisting that more can be done; even if you live up to the demands, or even if the moral preachers recognise your right to be happy sometimes, the warm feeling from doing good is greatly diminished when you are told that philantrophy is just being expected, that helping others is not something one does naturally with joy, but that it should be a conscious effort, a hard work, to be done properly.
xkcd reference.
Not to mention the remarks of Mark Twain on a fundraiser he attended once:
It might be worth taking a look at Karen Horney’s work. She was an early psychoanalyst who wrote that if a child is abused, neglected, or has normal developmental stages overly interfered with, they are at risk of concluding that just being a human being isn’t good enough, and will invent inhuman standards for themselves.
I’m working on understanding the implications (how do you get living as a human being right? :-/ ), but I think she was on to something.
I wasn’t abused or neglected. Did she check experimentally that abuse or neglect is more prevalent among rationalists than in the general population?
Of course that’s not something a human would ordinarily do to check a plausible-sounding hypothesis, so I guess she probably didn’t, unless something went horribly wrong in her childhood.
Second thought: Maybe I should have not mentioned her theory about why people adopt inhuman standards, and just focused on the idea that inhuman standards are likely to backfire, Viliam_Bur did.
Also—if I reread I’ll check this—I think Horney focused on inhuman standards of already having a quality, which is not quite the same thing as having inhuman standards about what one ought to achieve, though I think they’re related.
I was thinking about prase in particular, who sounds as though he might have some problems with applying high standards in a way that’s bad for him.
Horney died in 1952, so she might not have had access to rationalists in your sense of the word.
When I said it might be worth taking a look at Horney’s work, I really did mean I thought it might be worth exploring, not that I’m very sure it applies. It seems to be of some use for me.
To be clear, I don’t have problems with applying high standards to myself, unless not wishing to apply such standards qualifies as a problem. However I am far more willing to consider myself an altruist (and perhaps behave accordingly) when other people don’t constantly remind me that it’s my moral obligation.
Thanks for the explanation, and my apologies for jumping to conclusions.
I’ve been wondering why cheerleading sometimes damages motivation—there’s certainly a big risk of it damaging mine. The other half would be why cheerleading sometimes works, and what the differences are between when it works and when it doesn’t.
At least for me, I tend to interpret cheerleading as “Let me take you over for my purposes. This project probably isn’t worth it for you, that’s why I’m pushing you into it instead of letting you see its value for yourself.” with a side order of “You’re too stupid to know what’s valuable, that’s why you have to be pushed.”
I’m not sure what cheerleading feels like to people who like it.
No need to apologise.
The feeling of being forced to pursue someone else’s goals is certainly part of it. But even if the goals align, being pushed usually means that one’s good deeds aren’t going to be fully appreciated by others, which too is a great demotivator.
I think the feeling that one’s good deeds will be unappreciated is especially a risk for altruism.
I’m not at all convinced that this is the case. After all, the shampoos are being designed to be less painful, and you don’t need to test on ten thousand rabbits. Considering the distribution of the shampoos, this may save suffering even if you regard human and rabbit suffering as equal in disutility.
I’m not at all convinced of this. It seems to me that a genuinely ethical life requires extraordinary, desperate asceticism. Anything less is to place your own wellbeing above those of your fellow man. Not just above, but many orders of magnitude above, for even trivial luxuries.
Julia Wise would disagree, on the grounds that this is impossible to maintain and you do more good if you stay happy.
And the great philosopher Diogenes would disagree with her.
So, how many lives did he save again?
Clever guy, but I’m not sure if you want to follow his example.
That sounds to me like exactly the sort of excuse a bad person would use to justify valuing their selfish whims over the lives of other people. If we’re holding our ideas to scrutiny, I think the idea that the ‘Sunday Catholic’ school of ethics is consistent could take a long, hard look.
We’re talking about a person who, along with her partner, gives to efficient charity twice as much money as she spends on herself. There’s no way she doesn’t actually believe what she says and still does that.
That she gives more than most others doesn’t imply that her belief that giving even more is practically impossible isn’t hypocritical. Yes, she very likely believes it, thus it is not a conscious lie, but only a small minority of falsities are conscious lies.
Yeah, but there’s also a certain plausibility to the heuristic which says that you don’t get to second-guess her knowledge of what works for charitable giving until you’re—not giving more—but at least playing in the same order of magnitude as her. Maybe her pushing a little bit harder on that “hypocrisy” would cause her mind to collapse, and do you really want to second-guess her on that if she’s already doing more than an order of magnitude better than what your own mental setup permits?
I am actually inclined to believe Wise’s hypothesis (call it H) that being overly selfless can hamper one’s ability to help others. I was only objecting to army1987′s implicit argument that because she (Wise) clearly believes H, Dolores1984′s suspicion of H being a self-serving untrue argument is unwarranted.
There’s an Italian proverb “Everybody is a faggot with other people’s asses”, meaning more-or-less ’everyone is an idealist when talking about issues that don’t directly affect them/situations they have never experienced personally”.
You’re using hypocritical in a weird way—I’d only normally use it to mean ‘lying’, not ‘mistaken’.
I use “hypocrisy” to denote all instances of people violating their own declared moral standards, especially when they insist they aren’t doing it after receiving feedback (if they can realise what they did after being told, only then I’d prefer to call it a ‘mistake’). The reason why I don’t restrict the word to deliberate lying is that I think deliberate lying of this sort is extremely rare; self-serving biases are effective in securing that.
You underestimate force of habit, prase.
Can you explain?
I don’t believe it’s practically impossible to give more than I do. I could push myself farther than I do. I don’t perfectly live up to my own ideals. Given that I’m a human, I doubt any of you find that surprising.
This is why I think it’s not too terribly useful to give labels like “good person” or “bad person,” especially if our standard for being a “bad person” is “someone with anything less than 100% adherence to all the extrapolated consequences of their verbally espoused values.” In the end, I think labeling people is just a useful approximation to labeling consequences of actions.
Julia, Jeff, and others accomplish a whole lot of good. Would they, on average, end up accomplishing more good if they spent more time feeling guilty about the fact that they could, in theory, be helping more? This is a testable hypothesis. Are people in general more likely to save more lives if they spend time thinking about being happy and avoiding burnout, or if they spend time worrying that they are bad people making excuses for allowing themselves to be happy?
The question here is not whether any individual person could be giving more; the answer is virtually always “yes.” The question is, what encourages giving? How do we ensure that lives are actually being saved, given our human limitations and selfish impulses? I think there’s great value in not generating an ugh-field around charity.
Julia Wise holds the distinction of having actually tried it though. Few people are selfless enough to even make the attempt.
I believe Peter Singer actually originally advocated the asceticism you mention, but eventually moved towards “try to give 10% of your income”, because people were actually willing to do that, and his goal was to actually help people, not uphold a particular abstract ideal.
An interesting implication, if this generalizes: “Don’t advocate the moral beliefs you think people should follow. Advocate the moral beliefs which hearing you advocate them would actually cause other people to behave better.”
Just a sidenote: If you are the kind of person who is often worried about letting people down, entertaining the suspicion that most people follow this strategy already is a fast, efficient way to drive yourself completely insane.
“You’re doing fine.”
“Oh, I know this game. I’m actually failing massively, but you thought, well, this is the best he can do, so I might as well make him think he succeeded. DON’T LIE TO ME! AAAAH...”
Sometimes I wonder how much of LW is “nerds” rediscovering on their own how neuro-typical communication works.
I don’t mean to say I am not a “nerd” in this sense :).
The result bears about as much resemblance to real people as an FRP character sheet and rulebook.
Is it justified? Pretend we care nothing for good and bad people. Do these “bad people” do more good than “good people”?
Do you live a life of extraordinary, desperate asceticism? If not, why not? If so, are you happy?
Well, Jeff and I give about a third of our income, so I’d say we’re not Sunday Catholics but Sunday-Monday-and-part-of-Tuesday Catholics.
Seriously, though, I advocate that people do what will result in the most good, which is usually not to try for perfection. Dolores1984, you’ve said before that rather than fail at a high standard of helping you’d rather not help at all. (Correct me if that summary is wrong). I’d rather see people set a standard in keeping with their level of motivation, if that’s what will mean they take a stab at helping.
That’s fair. In my case, I think I’ve decided that, so long as we’re all going to be bad people, and value some human life much more than others, I’d rather care a lot about a few people than a little about a lot of people, and calibrate my charitable giving accordingly. It does not seem, in particular, less morally defensible, and it’s certainly more along the lines of what humans were built to do. To that end, I adopted a shelter cat who was about to be put down. My views may change slightly, however, when I am less thoroughly and completely broke.
Fallacy of grey much? We’re all going to be bad people, but some of us are going to be worse people than others.
The coalition of modules in your mind that believes in ascetism being the only acceptable solution is most likely vastly outnumbered by the hedonistic modules. (Most people for which this wasn’t the case were most likely filtered out of the gene pool.) As with politics, if you refuse to make compromises and insist on pushing your agenda while outnumbered, you will lose, or at best (worst?) create a deadlock in which nobody is happy. If you’re not so absolute, you’re more likely to achieve at least some of your aims.
Or, as Carl Shulman put it:
If I may be so bold as to summarize this thread:
Whatever utility calculus you follow, it is a mathematical model.
“All models are false.”
In particular, what’s going wrong here is your model is treating you, the agent, as atomic. In reality, as Kaj Sotala described very well below, you are not an atomic agent, you have an internal architecture, and this architecture has very important ramifications for how you should think about utilities.
If I may make an analogy from the field of AI. In the old days, AI was concerned about something called “discrete search,” which is just a brute force way to look for an optimum in a state space, where each state is essentially an atomic point. Alpha-beta pruning search Deep Blue uses to play chess is an example of discrete search. At some point it was realized that for many problems atomic point-like states resulted in a combinatorial explosion, and in addition states had salient features describable by, say, logical languages. As this realization was implemented, you no longer had a state-as-a-point, but state-as-a-collection-of-logical-statements. And the field of planning was born. Planning has some similarities to discrete search, but because we “opened up” the states into a full blown logical description, the character of the problem is quite different.
I think we need to “open up the agent.”
-- Iain McKay et al., An Anarchist FAQ, section C.7.3
Matt Ridley, in The Origins of Virtue
“Better off” according to whose utility function?
yeah, it’s not obvious from this quote, but having read the book, I know what he means. The utility function of the tree is the sum, over all individuals, of the fraction of genes that each other individual has in common with it. He constantly talks as if plants, chromosomes, insects etc. desire to maximize this number.
I think it works, because when an organism is in its environment of evolutionary adaptation, finding that a behavior makes this number bigger than alternative behaviors would explains why the organism carries out that behavior. And if the organism does not carry out the behavior, then you need some explanation for why not. Right?
That’s a really important caveat. Adaptation-Executers, not Fitness-Maximizers.
They’d expend less energy per surviving descendent produced.
--Kate Evans on Twitter
Don’t we all choose for ourselves on which side to drive? There’s usually nobody else ready to grab the wheel away from you...
There are police ready to pull you over, for certain values of “ready”. (Not commenting on whether that relates to Evans’ point.)
Have successfully quoted this to counter a relativist-truth argument that was aimed towards supporting “freedom of faith” even in hypothetical scenarios where the majority of actors would end up promoting and following harmful faiths.
While counterintuitive to me, it was apparently a necessary step before the other party could even comprehend the fallacy of gray that was being committed.
You may find it felicitous to link directly to the tweet.
You responded to the wrong post or gave the wrong link. I do see your point, fixed both quotes.
-- Kaiki Deishū, Episode 7 of Nisemonogatari.
Friedman and Felleisen, The Little Schemer
Could you unpack that for me?
Sure. The book is a sort of resource for learning the programming language Scheme, where the authors will present an illustrative piece of code and discuss different aspects of its behavior in the form of a question-and-answer dialogue with the reader.
In this case, the authors are discussing how to perform numerical comparisons using only a simple set of basic procedures, and they’ve come up with a method that has a subtle error. The lines above encourage the reader to figure out if and why it’s an error.
With computers, it’s really easy to just have a half-baked idea, twiddle some bits, and watch things change, but sometimes the surface appearance of a change is not the whole story. Remembering to “think first, then try” helps me maintain the right discipline for really understanding what’s going on in complex systems. Thinking first about my mental model of a situation prompts questions like this:
Does my model explain the whole thing?
What would I expect to see if my model is accurate? Can I verify that I see those things?
Does my model make useful predictions about future behavior? Can I test that now, or make sure that when it happens, I gather the data I need to confirm it?
It’s harder psychologically (and maybe too late) to ask those questions in retrospect if you try first, and then think, and if you skip asking them, then you’ll suffer later.
You know, I’ve seen a lot on here about how programming relates to thinking relates to rationality. I wonder if it’d be worth trying and where/how I might get started.
It’s certainly at least worth trying, since among things to learn it may be both unusually instructive and unusually useful. Here’s the big list of LW recommendations.
Khan Academy has a programming course? I might try it.
Mostly, I want the easiest, most handholdy experience possible. Baby talk if necessary. Every experience informs me that programming is hard.
This is the easiest, most handholdy experience possible: http://learnpythonthehardway.org/book/
A coworker of mine who didn’t know any programming, and who probably isn’t smarter than you, enjoyed working through it and has learned a lot.
Programming is hard, but a lot of good things are hard.
The first trick is to be able to describe how to solve a problem; and then break that description down into the smallest possible units and write it out such that there’s absolutely no possibility of a misunderstanding, no matter what conditions occur.
Once you’ve got that done, it’s fairly easy to learn how to translate it into a programming language.
Which is also why it helps, conversely, for reduction and rational thinking: The same skill that applies to formulating clear programs applies to formulating clear algorithms and concepts in any format, including thought.
I would recommend trying, although I’m not really the right person to ask on starting points, if for no other reason than to test the hypothesis that learning programming aids the study of rationality.
This is a great reminder, and is not always easy advice to follow, especially if your edit-compile-run cycle tightens or collapses completely. I think there’s a tricky balance between understanding something explicitly, which you can only do by training your model by thinking carefully, and understanding something intuitively, which is made much easier with tools like the ones I linked.
Do you have a sense for which kind of understanding is more useful in practice? I suspect that when I design or debug software I am making heavier use of System 1 thinking than it seems, and I am often amazed at how detailed a model I have of the behavior of the code I am working with.
No, I don’t, which I realized after spending half an hour trying to compose a reply to this. Sorry.
Bill Clinton
-- xkcd 667
Ernest Hemingway
Excellent. A shortcut to nobility. One day of being as despicable as I can practically manage and I’m all set.
It does not state which (!) former self, so I would expect some sort of median or mean or summary of your former self and not just the last day. So I’m sorry but there is no shortcut ;-)
Indeed: if you were to be ignoble one day and normal the next, then your nobility would have gone up significantly.
“If at first you don’t succeed, switch to power tools.” -- The Red Green Show
I can confirm that this works.
Michael Welfare, quoted in The Autobiography of Benjamin Franklin
“Our planet is a lonely speck in the great enveloping cosmic dark. In our obscurity—in all this vastness—there is no hint that help will come from elsewhere to save us from ourselves. It is up to us.”—Sagan
EDIT: Quote above is from the movie.
Verbatim from the comic:
I personally think that Watchmen is a fantastic study* on all the different ways people react to that realisation.
(“Study” in the artistic sense rather than the scientific.)
Douglas Hubbard, How to Measure Anything
This is the second time I’ve come across you mentioning Hubbard. Is the book good and, if so, what audience is it goo for?
How to Measure Anything is surprisingly good, so I added it here.
--Kate Evans on Twitter
I was ready to applaud the wise contrarianism here, but I’m having trouble coming up with actual examples… marriage, maybe?
I don’t know if this is what she was thinking of but church is what I thought of when I read it.
I thought of that too but dismissed it on the grounds that church is hardly “contingently associated” with religion.
But I think you’re probably right that that’s what she meant… and that being the case it is a pretty good point. I wish I belonged to something vaguely churchlike.
I disagree, but won’t present any arguments to avoid derailing or getting involved in a debate.
G. K. Chesterton
Chesterton doesn’t understand the emotion because he doesn’t know enough about psychology, not because emotions are deep sacred mysteries we must worship.
I read “irrational” as a genuflection in the direction of the is-ought problem more than anything else.
My beef isn’t with “irrational”, he meant “arational” anyway. It’s with the idea that this property of emotions make our ignorance about them okay.
Ah—I missed that implication. Agreed.
Or better, arational.
That is an incredible term. Going to use it all the time.
Pierre Proudhon, to Karl Marx
Edward Tufte, “Beautiful Evidence”
Evolution
Relativity
Foundational assumptions of standard economics
...what else?
Bayes’ theorem
Status
Computation
Utility
Optimisation
Quantum physics
Douglas Hubbard, How to Measure Anything
-- Bryan Caplan
Hitler was at least a hypocrite—he got his Jewish friends to safety, and accepted same-sex relationships in himself and people he didn’t want to kill yet. The kind of corruption Caplan is pointing at is a willingness to compromise with anyone who makes offers, not any kind of ignoring your principles. And Nazis were definitely against that—see the Duke in Jud Süß.
?
Please provide evidence for this bizarre claim?
Spared Jews:
Ernst Hess, his unit commander in WWI, protected until 1942 then sent to a labor (not extermination) camp
Eduard Bloch, his and his mother’s doctor, allowed to emigrate out of Austria with more money than normally allowed
I’ve heard things about fellow artists (a commenter on Caplan’s post mentions an art gallery owner) but I don’t have a source.
There are claims about his cook, Marlene(?) Kunde, but he seems to have fired her when Himmler complained. Anyone has Musmanno’s book or some other non-Stormfronty source?
Whether Hitler batted for both teams is hotly debated. There are suspected relationships (August Kubizek, Emil Maurice) but any evidence could as well have been faked to smear him.
Hitler clearly knew that Ernst Röhm and Edmund Heines were gay and didn’t care until it was Long Knives time. I’m less sure he knew about Karl Ernst’s sexuality.
Wittgenstein paid a huge bribe to allow his family to leave Germany. Somewhere I read that this particular agreement was approve personally be Hitler (or someone very senior in the hierarchy).
That doesn’t contradict the general point that Nazi Germany was generally willing to kill and steal from its victims (especially during the war) rather than accept bribes for escape.
This may have happened some of the time, but everything I read suggests it was the exception and not the rule.
The reason Jews did not emigrate out of Germany during the 30s was that Germany had a big foreign balance problem, and managed tight government control over allocation of foreign currency. Jews (and Germans) could not convert their Reichsmarks to any other currency, either in Germany or out of it, and so they were less willing to leave. And no other country was willing to take them in in large numbers (since they would be poor refugees). This continued during the war in the West European countries conquered by Germany. (Ref: Wages of Destruction, Adam Tooze)
Later, all Jewish property was expropriated and the Jews sent to camps, so there was no more room for bribes—the Jews had nothing to offer since the Nazis took what they wanted by force.
The last bit is most famously true of Rohm, though of course there’s a dozen different things going on there.
That sums it up.
Bargains and bribes seem of questionable use when a power is willing and able to kill you and seize all your assets anyway. I suppose there’s the odd Swiss bank account or successful smuggling case to deal with, or people willing to destroy their possessions rather than let them fall into the hands of a murderous authority, but I’d be surprised if any of these weren’t fairly small minorities in the face of the total. We’re certainly not talking millions.
Corruption at lower levels could have reduced the death toll of many famous genocides (in fact, I’d imagine it did), but at the level of Hitler or Pol Pot I can only see it helping if the bribes or bargains being offered are quite large and originate outside of the regions where the repression’s taking place. Much like the present situation with North Korea, come to think of it.
-- Jianzhi Sengcan
Edit: Since I’m not Will Newsome (yet!) I will clarify. There are several useful points in this but I think the key one is the virtue of keeping one’s identity small. Speaking it out loud is a sort of primer, meditation or prayer before approaching difficult or emotional subjects has for me proven a useful ritual for avoiding motivated cognition.
For the curious, it’s the opening of 信心铭 (Xinxin Ming), whose authorship is disputed (probably not the zen patriarch Jiangzhi Sengcan). In Chinese, that part goes:
(The Wikipedia article lists a few alternate translations of the first verses, with different meanings)
Do I understand you to be saying that you avoid “the struggle between ‘for’ and ‘against’” to an unusual degree compared to the average person? Compared to the average LWer?
No. I’m claiming this helps me avoid it more than I otherwise could. Much for the same reason I try as hard as I can to maintain an apolitical identity. From my personal experience (mere anecdotal evidence) both improve my thinking.
Respectfully, your success at being apolitical is poor.
Further, I disagree with the quote to extent that it implies that taking strong positions is never appropriate. So I’m not sure that your goal of being “apolitical” is a good goal.
Since we’ve already had exchanges on how I use “being apolitical”, could you please clarify your feedback. Are you saying I display motivated cognition when it comes to politically charged subjects or behave tribally in discussions? Or are you just saying I adopt stances that are associated with certain political clusters on the site?
Also like I said it is something I struggle with.
My impression that you are unusually NOT-mindkilled compared to the average person with political positions/terminal values as far from the “mainstream” as your positions are.
You seem extremely sensitive to the facts and the nuances of opposing positions.
Now I feel embarrassed by such flattery. But if you think this an accurate description then perhaps me trying evicting “the struggle between ‘for’ and ‘against’” from my brain might have something to do with it?
I’m not sure I understand what you mean by this then. Let’s taboo apolitical. To rephrase my original statement: “I try as hard as I can to maintain an identity, a self-conception that doesn’t include political tribal affiliations.”
You certainly seem to have succeeded in maintaining a self-identity that does not include a partisan political affiliation. I don’t know whether you consider yourself Moldbuggian (a political identity) or simply think Moldbug’s ideas are very interesting. (someday, we should hash out better what interests you in Moldbug).
My point when I’ve challenged your self-label “apolitical” is that you’ve sometime used the label to suggest that you don’t have preferences about how society should be changed to better reflect how you think it should be organized. At the very least, there’s been some ambiguity in your usage.
There’s nothing wrong with having opinions and advocating for particular social changes. But sometimes you act like you aren’t doing that, which I think is empirically false.
I disagree with the quote too. On the other hand, the idea of keeping one’s identity small is not the same as being apolitical. It means you have opinions on political issues, but you keep them out of your self-definition so that (a) changing those opinions is relatively painless, (b) their correlations with other opinions don’t influence you as much.
(Caricatured example of the latter: “I think public health care is a good idea. That’s a liberal position, so I must be a liberal. What do I think about building more nuclear plants, you ask? It appears liberals are against nuclear power, so since I am a liberal I guess I am also against nuclear power.”)
I agree with everything you just said—keeping one’s identity small does not imply that one cannot be extremely active trying to create some kind of social/political change.
I understand how a position can be correct or incorrect. I don’t understand how a position can be strong or weak.
In a world of uncertainty, numbers between 0 and 1 find quite a bit of use.
I understand what it means to believe that an outcome will occur with probability p. I don’t know what it means to believe this very strongly.
It means that many kinds of observation that you could make will tend to cause you to update that probability less.
Concretely: Beta(1,2) and Beta(400,800) have the same mean.
I don’t understand K to be arguing in favor of high-entropy priors, or T to be arguing in favor of low-entropy priors. My guess is that TimS would call a position a “strong position” if it was accompanied by some kind of political activism.
I think of a strong position as a low-entropy posterior, but rereading I am not confident that’s what TimS meant, and I also don’t see the connection to politics.
E.T. Jaynes’ Probability Theory goes into some detail about that in the chapter about what he calls the A_p distribution.
It means roughly that you give a high probability estimate that the thought process you used to come to that conclusion was sound.
A possible interpretation is that the “strength” of a belief reflects the importance one attaches to acting upon that belief. Two people might both believe with 99% confidence that a new nuclear power plant is a bad idea, yet one of the two might go to a protest about the power plant and the other might not, and you might try to express what is going on there by saying that one holds that belief strongly and the other weakly.
You could of course also try to express it in terms of the two people’s confidence in related propositions like “protests are effective” or “I am the sort of person who goes to protests”. In that case strength would be referring to the existence or nonexistence of related beliefs which together are likely to be action-driving.
They might also differ in just how bad an idea they think it is.
As I was using the term, “strong” is a measure of how far one’s political positions/terminal values are from the “mainstream.”
I’m very aware that distance from mainstream is not particularly good evidence of the correctness of one’s political positions/terminal values.
The claim looks narrower: repeating the poem makes Konkvistador more likely to avoid the struggle.
I like his contributions, but Konkvistador is not avoiding the struggle, when compared to the average LWer.
Sick people for some reason use up more medicine and may end up talking a lot about various kind of treatments.
Case in point:
-- Ro-Man
I don’t get it. Is this saying “Don’t be prejudiced or push for any overarching principle; take each situation as new and unknown, and then you’ll find easily the appropriate response to this situation”, or is this the same old stoicist “Don’t struggle trying to find food, choose to be indifferent to starvation” platitude?
Edited in a clarification. Though it will not help you since I have shown you the path you can not find it yourself. Sorry couldn’t resist teasing or am I? :P
-- 21st c. AI Clippy
To clarify the quote contains several useful interpretations, but I think the key one is the virtue of keeping one’s identity small.
-- GoodDamon (this may skirt the edge of the rules, since it’s a person reacting to a sequence post, but a person who’s not a member of LW.)
...and, more importantly, not on LessWrong.com.
Er… actually the genie is offering at most two rounds of feedback.
Sorry about the pedantry, it’s just that as a professional specialist in genies I have a tendency to notice that sort of thing.
Rather than a technical correction you seem just to be substituting a different meaning of ‘feedback’. The author would certainly not agree that “You get 0 feedback from 1 wish”.
Mind you I am wary of the the fundamental message of the quote. Feedback? One of the most obviously important purposes of getting feedback is to avoid catastrophic failure. Yet catastrophic failures are exactly the kind of thing that will prevent you from using the next wish. So this is “Just Feedback” that can Kill You Off For Real despite the miraculous intervention you have access to.
I’d say “What the genie is really offering is a wish and two chances to change your mind—assuming you happen to be still alive and capable of constructing corrective wishes”.
One well-known folk tale is based on precisely this interpretation. Probably more than one.
0 feedback is exactly what you get from 1 wish. “Feedback” isn’t just information, it’s something that can control a system’s future behavior—so unless you expect to find another genie bottle later, “Finding out how your wish worked” isn’t the same as feedback at all.
...or unless genies granting wishes is actually part of the same system as the larger world, such that what I learn from the results of a wish can be applied (by me or some other observer) to better calibrate expectations from other actions in that system besides wishing-from-genies.
I think it was clear that I inferred this as the new definition you were trying to substitute. I was very nearly as impressed as if you ‘corrected’ him by telling him that it isn’t “feedback” if nobody is around to hear it, or perhaps told him that oxygen is a metal.
Why only 2 rounds of feedback if you have 3 wishes?
The third one’s for keeps: you can’t wish the consequences away.
Morte’s Tale to Yves (Planescape: Torment)
I should like to point out that anyone in this situation who wishes what would’ve been their first wish if they had three wishes is a bloody idiot.
So: A genie pops up and says, “You have one wish left.”
What do you wish for? Because presumably the giftwrapped FAI didn’t work so great.
“I wish to know what went wrong with my first wish.”
This way, I at least end up with improved knowledge of what to avoid in the future.
Alternatively, “I wish for a magical map, which shows me, in real time, the location of every trapped genie and other potential source of wishes in the world.” Depending on how many there are, I can potentially get a lot more feedback that way.
I bet he’d wish “to erase all uFAI from existence before they’re even born. Every uFAI in every universe, from the past and the future, with my own hands.”
Nobody believes in the future.
Nobody accepts the future.
Then -
Perhaps I’m simply being an idiot, but … huh?
It’s a reference to an anime; you’re not an idiot, just unlikely to get the reference and its appropriateness if you’ve not seen it yourself. PM me for the anime’s name, if you are one of the people who either don’t mind getting slightly spoiled, or are pretty sure that you would never get a chance to watch it on your own anyway.
Could you just rot13 it? I’m curious too, I don’t mind the spoiler, and whatever it is, I’d probably be more likely to watch it (even if only 2epsilon rather than epsilon) for knowing the relevance to LW.
I’ll just PM you the title too, and anyone else who wants me to likewise. Sorry, it just happens to be one of my favourite series, and all other things being equal I tend to prefer that people go into it as completely unspoilered as possible… Even knowing Eliezer’s quote is a reference to it counts as a mild spoiler… explanation about how it is a reference would count as a major spoiler.
I think that’s Eliezer’s prediction of the results of siodine’s wish. Because wishes are NOT SAFE.
But what is he predicting, exactly?
“I wish for this wish to have no further effect beyond this utterance.”
Overwhelmingly probable dire consequence: You and everyone you love dies (over a period of 70 years) then, eventually, your entire species goes extinct. But hey, at least it’s not “your fault”.
But, alas, it’s the wish that maximizes my expected utility—for the malicious genie, anyway.
Possibly. I don’t off hand see what a malicious genie could do about that statement. However it does at least require it to honor a certain interpretation of your words as well as your philosophy about causality—in particular accept a certain idea of what the ‘default’ is relative to which ‘no effect’ can have meaning. There is enough flexibility in how to interpret your wish that I begin to suspect that conditional on the genie being sufficiently amiable and constrained that it gives you what you want in response to this wish there is likely to be possible to construct another wish that has no side effects beyond something that you can exploit as a fungible resource.
“No effect” is a whole heap more complicated and ambiguous than it looks!
You can’t use that tool to solve that problem.
Meanwhile, you have ⇐ 70 years to solve it another way.
You can’t? So much the worse for your species. I quite possibly couldn’t either. I’d probably at least think about it for five minutes first. I may even make a phone call first. And if I and my advisers conclude that for some bizarre reason “no further effect beyond this utterance” is better than any other simple wish that is an incremental improvement then I may end up settling for it. But I’m not going to pretend that I have found some sort of way to wash my hands of responsibility.
Yes, that’s better than catastrophic instant death of my species. And if I happen to estimate that my species has 90% chance of extinction within a couple of hundred years then I would be making the choice to accept a 90% chance of that doom. I haven’t cleverly tricked my way out of a moral conundrum, I have made a gamble with the universe at stake, for better or for worse.
Relevant reading: The Parable of the Talents.
“I wish for all humans to be immortal.”
Sure, you need to start heavily promoting birth control, and there can be problems depending on how you define “immortal”, but …
It’s a wish. You can wish for anything.
Unless, I suppose, that would have been your first wish. But the OP basically says your first wish was an FAI.
Immortal humans can go horribly wrong, unless “number of dying humans” is really what you want to minimize.
“Increase my utility as much as you can”?
I said:
You replied:
I am well aware that this wish has major risks as worded. I was responding to the claim that “you can’t use that tool to solve that problem.”
Yes, obviously you wish for maximised utility. But that requires the genie to understand your utility.
That would just cause them to pump chemicals in you head, I think. But it’s definitely thinking in the right direction.
Even with pseudo immortality, accidents happen, which means that the best way to minimize the number of dying humans is either to sterilize the entire species or to kill everyone. The goal shouldn’t be to minimize death but to maximize life.
That actually seems like it’d work.
It wouldn’t do that (except in some sense in which it is able to do arbitrary things you don’t mean when given complicated or undefined requests).
As long as I am not aware of that (or do not dislike it)… well, why not. However, MugaSofer is right, the genie has to understand the (future) utility function for that. But if it can alter the future without restrictions, it can change the utility function itself (maybe even to an unbounded one… :D)
“Destroy yourself as near to immediately as possible, given that your method of self destruction causes no avoidable harm to anything larger than an ant.”
They shrink the planet down to below our Schwarzschild radius, holding spacetime in place for just long enough to explain what you just did.
Alternately, they declare your wish is logically contradictory—genies are larger than ants.
A sphere whose radius equals the Earth’s Schwarzschild radius is larger than an ant.
At the start of the scenario, you are already dead with probability approaching 1. Trying to knock the gun away can’t hurt.
I was criticizing the wording of the “ant” qualifier, not the attempt to destroy the genie.
That’s not what’s going on though. The traveller is assuming, reasonably, that his third wish is reversing the amnesiac effects of his second. He’s not just starting fr om scratch.
I don’t think this follows from the text. The hag tells him “but second wish was for me to return everything to the way it was before you had made your first wish. That’s why you remember nothing; because everything is the way it was before you made any wishes”.
So she told him that he had been an amnesiac before any wishes were granted. Therefore he should have already guessed that his first wish was to know who he was—and that this proved a bad idea, since his second wish was to reverse the first.
It should be noted that night hags are sufficiently smart, powerful, and evil that your best case scenario upon meeting one is a quick and painful death.
“anyone in this situation” who believes that an elderly woman before him can grant arbitrary wishes is a bloody idiot to begin with, so the bar is set low.
But not everything is the way it was. Before he made any wishes, he had three.
She missed the chance to trap him in an infinite loop.
But then the Hag would be trapped too.
She gets delight from tormenting mortals, but tormenting the same one, in the same way eternally, would probably be too close to wireheading for her.
Well, if she got bored, she could experiment with different ways to present his wishes to him at the “beginning” and see if she can get him to wish for something to else, or word it a bit differently. Since she seems to retain memories of the whole thing. (Which is again, things not being how they were, but.)
The psuedo-meta-textual answer is that Morte is lying to Yves while the main character overhears. Morte’s making up the story just to mess around with him.
Background information is that gur znva punenpgre znqr n qrny jvgu n Unt (rivy cneg-gvzr travr), tnvavat vzzbegnyvgl naq nzarfvn. At the start of the story, the main character somehow broke out of an infinite loop of torture; he’s stopped having Anterograde Amnesia, but still cannot remember much from before the cycle broke, and is on a quest to remember who he is. Morte is trying to dissuade the main character from finding out who he is, showing that things can be terrible even without an infinite loop.
Now that would be evil.
If his first wished disappeared him forever, how did he ever get a second wish?
Apparently I suck at reading.
The old woman is the one disappearing forever, and only because the wishes ran out.
Right, but the consequences still qualify as feedback, no?
I always imagine the genie just goes back into his lamp to sleep or whatever, so in the hypothetical as it exists in my head, no. But I guess there could be a highly ambitious Genie looking for feedback after your last wish, so maybe.
I think in this case, Eliezer in talking about a genie like in Failed Utopia 4-2 who grants his wish, and then keeps working, ignoring feedback, because he just doesn’t care, because caring isn’t part of the wish.
The genie doesn’t care about consequences, he just cares about the wishes. The second wish and third wish are the feedback.
The feedback is for you, not what you happen to say to the genie.
… which one wish, carefully phrased, could also provide.
you can’t wish for more wishes
“I wish for the result of the hypothetical nth wish I would make if I was allowed to make n wishes in the limit as n went to infinity each time believing that the next wish would be my only one and all previous wishes would be reversed, or if that limit does not exist, pick n = busy beaver function of Graham’s number.”
I can see a genie taking a shortcut here.
— Granny Weatherwax, A Hat Full of Sky.
In short, the genie may well conclude that every m’th wish, for some m (Granny Weatherwax suggests here that ‘m’ is three) your wish would be to have never met the genie in the first place. At this point, if you’re lucky, the genie will use a value of n that’s a multiple of m. If you’re unlucky, the genie will use a value of n that’s km-1 for some integer k...
Alternatively, you’ll end up with a genie who can’t handle the math and does not understand what you’re asking for.
I’m pretty sure this would result in the genie killing you.
Because I would wish to kill myself eventually? It’s hard to imagine that I would do that, faced with unlimited wishes. If I got bored, I could just wish the boredom away.
Though on reflection this wish needs a safeguard against infinite recursion, and a bliss-universe for any simulated copies of me the genie creates to determine what my wishes would be.
In a sufficiently bad situation, you may wish for the genie to kill you because you think that’s your only wish. It’s not likely for any given wish, but would happen eventually (and ends the recursion, so that’s one of the few stable wishes).
If I kill myself, there is no nth wish as n → infinity, or a busy beaver function of Graham’s numberth wish, so the first wish is wishing for something undefined.
Also, the probability of any of the individually improbable events where I kill myself happening is bounded above by the some of their probabilities, and they could be a convergent infinite series, if the probability of wanting to kill myself went down each time. Even though I stipulated that it’s if I believed each wish was the last, I might do something like “except don’t grant this wish if it would result in me wanting to kill myself or dying before I could consider the question” in each hypothetical wish. Or grant myself superintelligence as part of one of the hypothetical wishes, and come up with an even better safeguard when I found myself (to my great irrational surprise) getting another wish.
There is not even necessarily a tiny chance of wanting to kill myself. Good epistemology says to think there is, just in case, but some things are actually impossible. Using wishes to make it impossible for me to want to kill myself might come faster than killing myself.
I think you’re right, though I’m not sure that’s exactly a good thing.
I see no particular reason to expect that to be the case.
Excellent point. That might just work (though I’m sure there are still a thousand ways it could go mind-bogglingly wrong).
If you did eventually wish for death, then you would have judged that death is the best thing you can wish for, after having tried as many alternatives as possible.
Are you going to try to use your prior (I don’t want to die) to argue with your future self who has experienced the results (and determines that you would be happier dying right now than getting any wish)?
I would not want to kill myself if my distant-future self wanted to die or wanted me to die immediately. I think it is much more likely that I would accidentally self-modify in a manner I wouldn’t like if I reflected on it now and that would lead to wishing for death than that my current self with additional knowledge would chose death over omnipotence.
I don’t think the factual question “would I be happier dying right now” would necessarily be the one that decided the policy question of “will I chose do die” for both me and my future self, because we could each care about different things.
And with a warning “The way things are going, you’ll end up wanting to die.” I could change my wishes and maybe get something better.
“Does my current utility function contain a global maximum at the case where I wish for and receive death right now?” is a really scary question to ask a genie.
I would prefer “I wish for the world to changed in such a manner as my present actual utility function, (explicitly distinct from my perception of my utility function), is at the global maximum possible without altering my present actual utility function.”
Or in colloquial language “Give me what I want, not what I think I want.”
That sounds pretty scary too. I don’t think I am close enough to being an agent to have a well-defined utility function. If I do (paradoxical as it sounds), it’s probably not something I would reflectively like. For example, I think I have more empathy for things I am sexually attracted to. But the idea of a world where everyone else (excluding me and a few people I really like) is a second-class citizen to hot babes horrifies me. But with the wrong kind of extrapolation, I bet that could be said to be what I want.
I can’t easily describe any procedure I know I would like for getting a utility function out of me. If I or some simulated copy of me remained to actually be deciding things, I think I could get things I would not only like, but like and like liking. Especially if I can change myself from an insane ape who wishes it was a rationalist, to an actual rationalist through explicitly specified modifications guided by wished-for knowledge.
The best way I can think of to ensure that the extrapolated utility function is something like whatever is used in making my decisions, is to just use the brain circuits I already have that do that the way I like.
I also think a good idea might be to have a crowd of backup copies of me. One of us would try making some self-modifications in a sandboxed universe where their wishes could not get outside, and then the others would vote on whether to keep them.
Well, you don’t prefer a world “where everyone else (excluding me and a few people I really like) is a second-class citizen to hot babes horrifies me.” to the current world. If you can express such a judgement preferring one universe over another, and those judgements are transitive, you have a utility function.
And if you want to like liking them, that is also part of your utility function.
One confounding factor that you do bring up- the domain of one’s utility function really doesn’t need to include things outside the realm of possibility.
Well if deep down I wish that hot babes were given higher status in society than ugly females and males I didn’t know (I’ll roll with this example even though with omnipotence I could probably do things that would make this a nonissue) and I wish that I didn’t wish it, and the genie maximizes my inferred utility function once (not maximizing it once and then maximizing the new one that doing so replaces it with), we would end up with a world where hot babes had higher status and I wished they didn’t. The only thing I can see saving us from this would be if I also wanted my wants to be fulfilled, whatever they were. But then we would end up with me assigning Flimple utility to 2 + 2 being equal to 4.
I understand the argument “If you preferred the world ruled by hot babes it wouldn’t horrify you,” but what if that’s just because when I imagine things at a global scale they turn featureless, and appearances don’t matter to me anymore, but when I see the people in person they still do? What if, being able to see and socialize with every hot babe in the world changed my mind about whether I would want them to have higher social status, even if I was also able to see and socialize with everyone else?
What if the appearance of transitivity is only from drawing analogies from the last scenario I thought about (A) to the next one (B), but if I started a chain of analogies from somewhere else (C) my view on B would be completely different, such that you could take exchange rates of certain things I considered good in each scenario and construct a cycle of preference?
There’s a difference between finding a global maximum of a function and finding the local derivative and moving infinity towards a local maximum.
Even if your actual utility function were for one group to have higher status, that does not imply that the greatest utility comes at the highers imbalance.
And if your preferences aren’t transitive, you don’t have a valid utility function. If you find yourself in a cycle of preference, you have probably failed to accurately judge two or more things which are hard to compare.
The problem isn’t that I might see a hot babe and care about her more than other people, but that when the numbers are bigger I care about everyone the same amount. It’s that it probably dependent on things like whether i have seen them, and how much I know about them. If I was told by a perfectly reliable source that X is a hot babe by my standards, that would not make me care about her more than other people. But it would if I saw her. So what I want is not just dependent on what I believe, but on what I experience. On some figurative level, I’m a Mealy machine, not a Moore machine.
Why do you think this?
When I prefer A to B, B to C, and C to A (a cycle of preference), then I typically find that there is something moral in the general sense about A, something moral in the direct sense about B, and something which is personally advantageous about C.
For instance, I would rather my donations to charity have a larger total effect, which in the current world means donating to the best aggregator rather to individual causes. I would rather donate to individual causes than ignore them in selfish self-interest. I would rather spend my money in my own self interest than redistribute it in order to achieve maximum benefits. I believe that the reason I think my preferences lie this way is that I am unable to judge the value of diversified charity compared to selfish behavior.
In short, I am extrapolating from the single data point I have. I am assuming that there is one overwhelmingly likely reason for making that type of error, in which case it is likely that mine is for that reason and that yours is as well.
A much more real concern is that the genie is going to need to create and destroy at least BB(G) copies of you in order to produce the data you seek. Which is not good.
but does “I wish my wishes were divided up into 100 rounds of feedback each, each roughly equivalent to 1⁄100 of a wish” fall under that?
My impression would be that you could use one wish to divide a second wish up in such a manner. Using it to divide itself up would ‘fall under’. (I’m not sure how the genie would resolve the wish into a practical outcome.)
Yea, what it does is give you 200 subwishes, not 300.
Kate Evans https://twitter.com/aristosophy/status/240589485098795008
Wish 1: “I wish for a paper containing the exact wording of a wish that, when spoken to you, would meet all my expectations for a wish granting X.” For any value of X.
Wish 2: Profit.
Three wishes is overkill.
Genie provides a 3,000 foot long scroll, which if spoken perfectly will certainly do as you ask, but if spoken imperfectly in any of a million likely ways affords the genie room to screw you over.
Or the scroll is written in Martian.
I just take this as evidence that I -can’t- beat the genie, and don’t attempt any more wishes.
Whereas, if it’s something simple then I have pretty strong evidence that the genie is -trying- to meet my wishes, that it’s a benevolent genie.
Wish 2: I wish for a text-to-speech device capable of reading from this scroll with perfect accuracy.
Wish 3: delegated to the device from #2.
Are we going to keep patching up every hole she points out? Or admit that a UFAI genie can be smarter than any human (even if that human is our esteemed Alicorn, or (gasp!) Eliezer)?
Too bad Martian words sound exactly like lethal sonic weapons and your original X that your wish is about doesn’t, strictly speaking, require resurrecting you to enjoy it.
Or, the genie doesn’t have to respond to wishes that don’t come out of Master’s mouth.
Text-to-speech device provided. It reads from the scroll with perfect accuracy and low speed. It will take a few hundred years to complete this task.
You will need to change the batteries once an hour; it you forget, it starts reading from the start of the scroll again. (And where do you get a large supply of size Q batteries, in any case?)
I know some electrical engineers. It’s not all that hard to rig up an uniniterruptible power supply that runs off line voltage. The delay is inconvenient, but for the right wish it’s acceptable.
“I wish for everything written on this scroll.” Or some variation thereof that more exactly expresses that general idea.
All of the nouns named on the scroll appear. Some of them are things that the wording of the scroll expressly insists that the wish must avoid, due to their being lethal or otherwise undesirable.
“I wish for everything that would happen if I read this scroll perfectly.”
Among other things, you would suffocate due to that four-minute no-breathing-allowed Martian word in paragraph nine.
Ooooh. Well played.
4 minutes is survivable if trained.
Fine, thirty.
Not while speaking.
In the likely event that it’s impossible for you to read the scroll perfectly, it’s true for all X that “X would happen if you read this scroll perfectly”. Which means that anything the genie feels like doing satisfies that wish. Or possibly the genie has to make everything happen that could possibly happen. Neither of those seems like a good outcome.
Hm… how about “I wish to have all the skills and abilities required to formulate an unambiguous wish in standard English that would allow me to fulfill any of my non-contradictory desires that I choose, and to be able to choose which of any desires that are contradictory said wish would fulfill, and to be able to express that unambiguous wish in an unambiguous way in less than thirty seconds and with no consequences to incorrectly expressing that wish apart from the necessity of trying again to express it.”
The scroll modifies your expectations. The genie twist-interprets X, and then assesses your expectations of the result of the genie’s interpretation of X. (“Why, that’s just what you’d expect destroying the world to do! What are you complaining about?”) The complete list of expectations regarding X is at least slightly self-contradictory, so of course the genie has no option except to modify your expectations directly...
OOoh, is this now the “eliezer points out how your wish would go wrong” thread! I wanna play to! :p
“I wish for that which I’d wish for if I had an uninterrupted year of thinking about it and freely talking to a dedicated copy of Eliezer Yudovsky”
Uh oh...
Eliezer Yudkowsky:
Eliezer Yud_ov_sky:
No sleep, or anything that would interrupt thinking about it, for a year, might lead to an interesting wish.
Well, it’s obvious what happens then: the genie lets a dedicated copy of Eliezer out of a box.
No sleep, or anything else that would mean not thinking about it, for a year. That
The genie is, after all, all-powerful, so there are any number of subtle changes it could make that you didn’t specify against that would immediately make you, or someone else, wish for the world to be destroyed. If that’s the genie’s goal, you have no chance. Heck, if it can choose it’s form it could probably appear as some psycho-linguistic anomaly that hits your retina just right to make you into a person who would wish to end the world.
Really I’m just giving the genie a chance to show me that it’s a nice guy. If it’s super evil I’m doomed regardless, but this wish test (hopefully) distinguishes between a benevolent genie and one that’s going to just be a dick.
If you consider three class of genies :
(A) a genie that’s going to be “just be a dick” but is not skilled at it ;
(B) a genie that is benevolent ;
(C) a genie that’s going to be “just be a dick” but is very skilled at it.
Your test will (may at least) tell apart A from (B or C). It won’t tell apart B from C.
The “there is no safe wish” rule applies to C. Sure, if your genie is not skilled a being “evil” (having an utility function very different from yours), you can craft a wish that is beyond the genie’s ability to twist it. But if the genie is skilled, much more intelligent than you are, with like the ability to spend the equivalent of one million of years of thinking how to twist the wish in one second, he’ll find a flaw and use it.
“I wish for a paper containing the exact wording of a wish that, when spoken to you, would meet all my expectations as of September 3, 2012, for a wish granting X.”
(Then, if my expectations yesterday did contain self-contradictions, the genie will do… whatever it did if I wished that 2 + 2 = 5.)
I’m pretty sure your belief network is not coherent enough so that it is possible to “meet all your expectations”, there must be somewhere two expectations which you hold but which aren’t, in fact, compatible. So the wish will fizzle ;)
A wish is a pretty constrained thing, for some wishes.
If I wish for a pile of gold, my expectations probably constrain lots of externalities like ‘Nobody is hurt acquiring the gold, it isn’t taken from somewhere else, it is simply generated and deposited at my feet, but not, like, crushing me, or using the molecules of my body as raw material, or really anything that kills me for that matter’. Mostly my expectations are about things that won’t happen, not things that will happen that might conflict (that consists only of: the gold will appear before me and will be real and persistent).
If you try this with a wish for world peace, you’re likely to get screwed. But I think that’s a given no matter your strategy. Don’t wish for goals, wish for the tools to achieve your goals—you’ll probably be more satisfied with the result to boot.
You’re already lowering your claim, it’s not longer “for any value of X”.
But even so...
“Nobody is hurt acquiring the gold” does that include people hurt because your sudden new gold decrease the market value of gold, so people owning stocks of gold or speculating on an increase of the gold price are hurt ? Sure, you can say “it’s insignificant”, but how will a genie tell that apart ? Your expectation of what having a sudden supply of gold on the market would do and the reality of how it’ll unfold probably don’t match. So the genie will have to do corrections for that… which will themselves have side-effects...
Also, you’ll probably realize once you’ve some gold that gold doesn’t bring you as much as you thought it would bring you (at least, it happens to most lottery winner), so even if you genuinely get the gold, it’ll fail to “meet all your expectations” of having gold. Unless the genie also fixes you so you get as much utility/happiness/… from the gold as you expected to get from it. And as soon as the genie has to start fixing you… game over.
I simplify here because a lot of people think I will have contradictory expectations for a more complex event.
But I think you’re being even more picky here. Do I -expect- that increasing the amount of gold in the world will slightly affect the market value? Yes. But I haven’t wished anything related to that, my wish is -only- about some gold appearing in front of me.
Having the genie magically change how much utility I get from the gold is an even more ridiculous extension. If I wish for gold, why the heck would the genie feel it was his job to change my mental state to make me like gold more?
Possibly we just think very differently, and your ‘expectation’ of what would happen when gold appears also includes every thing you would do with that gold later, despite, among many, many things, not even knowing -when- you would speak the wish to get the gold, or what form it would appear in. And you even have in mind some specific level of happiness that you ‘expect’ to get from it. If so, you’re right, this trick will not work for you.
If you wish for gold, it’s because you have expectation on what you’ll do with that gold. Maybe fuzzy ones, but if you didn’t have any, you wouldn’t wish for gold. So you can’t dissociate the gold and the use when what you’re speaking about is “expectations”.
Or else, solutions like “the world is changed so that the precious metal is lead, and gold has low value, but all the rest is the same” would work. And that wouldn’t meet your “expectations” about the wish at all.
An altogether fairly harmless outcome.
Couldn’t you just wish that all your expectations for a wish granting X were granted, and take out the second step?
This presumes, of course, that my expectations for a wish granting X, for some value of X, is such that having a wish granted that meets them is profitable.
How about wishing for enough judgment to make wishes that you won’t regret? Additional clause: the genie isn’t allowed to deteriorate your mental capacities below their current level.
-Lewis Carroll, The Hunting of the snark
.
“Do you want 1111 1111 0000 0000 1111 1111 or 1111 1101 0000 0100 1111 1111? ”
— Michael Kirkbride / Vivec, “The Thirty Six Lessons of Vivec”, Morrowind.
Am I the only one who thinks we should stop using the word “simple” for Occam’s Razor / Solomonoff’s Whatever? In 99% of use-cases by actual humans, it doesn’t mean Solomonoff induction, so it’s confusing.
How would you characterise the in your opinion most prevalent use-cases?
“Easy to communicate to other humans”, “easy to understand”, or “having few parts”.
“Having few parts” is what Occam’s razor seems to be going for. We can speak specifically of “burdensome details,” but I can’t think of a one-word replacement for “simple” used in this sense.
It is a problem that people tend to use “simple” to mean “intuitive” or “easy to understand,” and “complicated” to mean “counterintuitive.” Based on the “official” definitions, quantum mechanics and mathematics are extremely simple while human emotions are exceedingly complex.
I think human beings have internalized a crude version of Occam’s Razor that works for most normal social situations—the absurdity heuristic. We use it to see through elaborate, highly improbable excuses, for example. It just misfires when dealing with deeper physical reality because its focus is on minds and emotions. Hence, two different, nearly opposite meanings of the word “simple.”
Yeah, various smart people have made that point repeatedly, but Eliezer and Luke aren’t listening and most people learn their words from Eliezer and Luke, so the community is still being sorta silly in that regard.
-L. A. Rollins, Lucifer’s Lexicon: An Updated Abridgment
John Adams
-- W.H. Press et al., Numerical Recipes, Sec. 15.1
I don’t think it’s fair to blame the mathematical statisticians. Any mathematical statistician worth his / her salt knows that the Central Limit Theorem applies to the sample mean of a collection of independent and identically distributed random variables, not to the random variables themselves. This, and the fact that the t-statistic converges in distribution to the normal distribution as the sample size increases, is the reason we apply any of this normal theory at all.
Press’s comment applies more to those who use the statistics blindly, without understanding the underlying theory. Which, admittedly, can be blamed on those same mathematical statisticians who are teaching this very deep theory to undergraduates in an intro statistics class with a lot of (necessary at that level) hand-waving. If the statistics user doesn’t understand that a random variable is a measurable function from its sample space to the real line, then he/she is unlikely to appreciate the finer points of the Central Limit Theorem. But that’s because mathematical statistics is hard (i.e. requires non-trivial amounts of work to really grasp), not because the mathematical statisticians have done a disservice to science.
New Peter / Orson Scott Card, Children of the Mind
That’s a modest thing to say for a vain person. It even sounds a bit like Moore’s paradox—I need advice, but I don’t believe I do.
(Not that I’m surprised. I’ve met ambivalent people like that and could probably count myself among them. Being aware that you habitually make a mistake is one thing, not making it any more is another. Or, if you have the discipline and motivation, one step and the next.)
I love New Peter. He’s so interesting and twisted and bizarre.
Baruch Spinoza Ethics
http://lesswrong.com/lw/i0/are_your_enemies_innately_evil/
That seems really odd to me, coming from Spinoza. I’ve never read him, but I thought that he was supposed to believe that God and Nature are the same thing. Does he do that, but then also investigate the nature of God through analyzing the way that Nature’s laws work? How does he reconcile those two positions, I guess, is what I’m asking.
Can someone more familiar with his work than I help me out here?
Spinoza held that God and Nature are the same thing.
His reasoning in a nutshell: an infinite being would need to have everything else as a part of it, so God has to just be the entire universe. It’s not clear whether he really thought of God as a conscious agent, although he did think that there were “ideas” in God’s mind (read: the Universe) and that these perfectly coincided with the existance of real objects in the world. As an example, he seems to reject the notion of God as picking from among possible worlds and “choosing” the best one, opting instead to say that God just is the actual world and that there is no difference between them.
So basically, studying nature for Spinoza is “knowing the mind of God.”
He may also have been reacting to his excommunication, in fact, that’s pretty likely. So the quote may have some sour grapes hidden inside of it.
That doesn’t hold in maths at least. N, Z, Q have the same size, but clearly Q isn’t part of N. And there are as many rational numbers between 0 and 1 (or between 0 and 0.0000000000000000000001) than in Q as a whole, and yet, we can have an infinity of such different subsets. And it goes even worse with bigger sets.
It saddens me how much philosopher/theologists speak about “infinity” as if we had no set theory, no Peano arithmetic, no calculus, nothing. Intuition is usually wrong on “infinity”.
Baruch Spinoza: 1632-1677 Isaac Newton: 1642-1727 Georg Cantor: 1845-1918 Richard Dedekind: 1831-1916 Guiseppe Peano: 1858-1932
Ok, I stand corrected on the dates, my mistake.
But still, didn’t we already know that if you take a line, two distinct points A and B on it, there are an infinite number of points between A and B, and yet an infinite number of points outside [AB] ? Didn’t we know that since the ancient greeks ?
First, Spinoza is not using infinite in its modern mathematical sense. For him, “infinite” means “lacking limits” (see Definition 2, Part I of Ethics). Second, Spinoza distinguished between “absolutely infinite” and “infinite in its kind” (see the Explication following Definition 6, Part I).
Something is “infinite in its kind” if it is not limited by anything “of the same nature”. For example, if we fix a Euclidean line L, then any line segment s within L is not “infinite in its kind” because there are line segments on either side that limit the extent of s. Even a ray r within L is not “infinite in its kind”, because there is another ray in L from which r is excluded. Among the subsets of L, only the entire line is “infinite in its kind”.
However, the entire line is not “absolutely infinite” because there are regions of the plane from which it is excluded (although the limits are not placed by lines).
I suspect “infinite” was supposed to mean “having infinite measure” rather than “having infinite number of points / subsets”. In the latter sense every being, not only God, would be infinite.
That’s a good point. Spinoza himself was a mathematician of no mean talent, so we should assume that he was aware of it as well. So the question is, does his argument avoid the mistake of taking ‘infinite’ to mean ‘all encompassing?’ without any argument to that effect? There are certainly questions to be raised about his argument, but I don’t think this is one of his mistakes. If you don’t want to take my word for it, here’s the opening argument of the Ethics. Good luck, it’s quite a slog.
The idea seems to be that the one substance has to be infinite and singular, because substances can’t share attributes (see his definitions), and things which have nothing in common can’t interact. Therefore substances can’t cause each other to exist, and therefore if any exists, it must exist necessarily. If that’s true, then existence is an attribute of a substance, and so no other substance could exist.
At any rate, the argument concerns an ‘infinity’ of attributes, and I think these are reasonably taken as countably infinite. Spinoza also defines infinite as ‘not being limited by anything of the same kind’, so by that definition he would say that with reference to the ‘kind’ ‘number’, the even numbers are finite, though they’re infinite with reference to the ‘kind’ ‘even number’.
Thanks.
My understanding was basically correct then. I just didn’t understand why he’d go from that overall position to talk about why we need to investigate nature, when his whole approach really seemed more like laid back speculation than any form of science, or advocacy of science. The excommunication detail clarifies a lot though, as Spinoza’s approach seems much more active and investigative when compared to the approach of the church.
Excellent, thanks again.
It’s notable that Spinoza was a part of a Jewish community, rather than “a church.” I’ve actually read the letter of his excommunication, and WOW. They really went all out. You’re considered cursed just for reading what he wrote.
Are you reacting to Spinoza’s mention of “miracles” and “gods”?
Spinoza held that there are no miracles in the sense that everything without exception proceeds according to cause and effect. So he must mean something like “alleged miracles”. As for the “gods”, they are mentioned only as part of the belief system of the “common people”.
What do you- he’s a pantheist. Contemporaries called him an atheist because his position works exactly like atheism.
Downvoted for glibness and vagueness.
― Robert M. Pirsig, Zen and the Art of Motorcycle Maintenance: An Inquiry Into Values
Well. Surely that’s only part of the real purpose of the scientific method.
G. K. Chesterton, “The Absence of Mr Glass”
Note: this was put in the mouth of the straw? atheist. It’s still correct.
Then Chesterton didn’t say it.
It is typical to quote the author of fictional works for quotes from that fictional work, though I think it’s somewhat more conventional here on LW to quote the character.
This is my home, the country where my heart is;
Here are my hopes, my dreams, my sacred shrine.
But other hearts in other lands are beating,
With hopes and dreams as true and high as mine.
My country’s skies are bluer than the ocean,
And sunlight beams on cloverleaf and pine.
But other lands have sunlight too and clover,
And skies are everywhere as blue as mine.
-Lloyd Stone
Duplicate, please delete the other.
obviously he never visited the British Isles :D
-- Tenzin Gyatso, 14th Dalai Lama
Not all that rational. Note that he requires scientific proof before he is willing to change his beliefs. The standards should be much lower than that.
Wikiquote (http://en.wikiquote.org/wiki/Tenzin_Gyatso,_14th_Dalai_Lama) quotes this as
I like that. It is a bit careful, but better than everything else I saw from other religions.
Fortunately for him, Buddhism is cleverly designed to contain no scientifically falsifiable claims.
Well, maybe some of the 14 (really only 4) unanswerable questions can be answered some day. Cosmologists might prove that the universe is finite (current odds are slim), AI researchers might prove that self is both identical with the body and different from it by doing a successful upload. Would it make Buddhists abandon their faith? Not a chance.
A Buddhist friend told me that at a class in a temple, the teacher mentioned four types of creation, of which one was spontaneous generation- like maggots spontaneously generating in meat. My friend interrupted to say that, no, that’s not actually what happens, and that people did experiments to prove that it didn’t happen. (My friend was too polite to mention that the experiments were 350 years old.) If I remember correctly, the teacher said something like “huh, okay,” and went on with the lesson, leaving out any parts relevant to spontaneous generation.
That’s an unfortunate example. The teacher should have amended his maggot example to ‘the first living cell’, then.
Traditional Spontaneous generation and modern abiogenesis are very different things, and comments that assume the first may be invalid if only the second is true.
They may have differences… but are they any that matter for that Buddhist typology of creation?
I’d say “careful” would be the other way around, not believing doctrines that make complex claims about reality until he has good evidence that those beliefs are true. Giving up hard-to-test beliefs only in the extreme case where scientific research conclusively proves them wrong is just a small concession in the direction of being epistemically responsible.
I meant “careful with respect to ‘admitting that religious claims can be wrong’ ”—in other words, the same as you.
Christians have given up virtually all of the bible on the basis of science. Whether or not they are still christians is another issue.
That’s a bold claim. The Old Testament historical narrative post-Genesis is still controversial-to-accepted; anthropology hasn’t for example turned up any evidence that I know of for Hebrew slavery under Pharaonic Egypt, but it’s still presented as fact in many Christian circles that are not Biblical literalists. Deuteronomy and Leviticus have largely been rejected, but for cultural rather than scientific reasons. Psalms still seems to be taken in the spirit it was intended.
On the New Testament side of things, the Gospels still generally seem to be taken as, well, gospel, miracles and all. Acts is mostly accepted. The epistles are very short on supernaturalist claims, concerning themselves mostly with organization and ethics. Revelation’s supernaturalist as hell but it’s in prophecy form, and interpretations of it vary widely anyway.
Really, aside from scattered references like that odd pi == 3 thing, about the only parts of the Christian Bible that mainstream churches have widely dropped on scientific grounds are in Genesis—and these days it’s got to be pretty hard for any religion to maintain a literalist interpretation of its creation myth, if it has any regard for science whatsoever.
Ok, I guess I underestimated how many people believe in miracles.
I underestimated how bad survey questions can be.
”Do you completely agree / mostly agree / mostly disagree / completely disagree with: Miracles still occur as in ancient times” (I shortened the first part a bit, without changing the context)
Seriously, wtf? The question assumes that miracles occured in ancient times. It does not define what “miracle” means at all, and it does not ask if miracles occur at all, it asks for a trend. 79% of the answers were counted as “belief in”, I think that those were the first two groups only (but I do not see that in the study).
However, the questions about heaven and hell are fine, and the large amount of “yes” answers (heaven 74%, hell 59%) makes me sad.
Funny numbers:
At least 15% believe that “good” people come to heaven, but “bad” do not come to hell. So where do “bad” people go? To heaven, too?
In the group of age 65+, 74% believe in heaven, but only 71% believe in a life after death. So at least 3% believe that “good” people will live in heaven after death, without living at all.
That’s one option. Another option would be that bad people just cease to exist. Or they get reincarnated until they become non-bad enough for heaven.
The mormons would tell you, for the most part, yes. And they generally believe in heaven and not hell.
Indeed, I might have given a “completely agree” there given that miracles occurred none of the time in ancient times, and are still going strong at that rate. But maybe other respondents have less trouble with loaded questions.
Or those 3% believe in heaven but don’t believe that dead people get to go there. It might just mean “God’s house”, or be reserved for those who are raptured.
No, the claim is that a scientific proof is sufficient from him to feel the need to change his beliefs. It isn’t that it’s necessary.
Technically true, but nice though that is, saying that scientific proof would force you to change your beliefs still isn’t a very impressive show of rationality. It would be better if he had said “Whenever science and Buddhism conflict, Buddhism should change”.
I know, it is good to hear it from a religious figure, but if it were any other subject the same claim would leave you indifferent. “If it were scientifically proven that aliens don’t exist I will have to change my belief in them.” Sound impressive? No? Then the Dalai Lama shouldn’t get any more praise just because it’s about religion.
When would you say that science and X is in conflict when there isn’t scientific proof that X is wrong?
Science is a method. In itself it’s about doing experiments. It’s not about the ideology of the scientist that might conflict with X even if there’s no proof that X is wrong.
Science and X are in conflict when on the whole there is more scientific evidence that X is wrong than there is evidence that it is right. Saying “I’ll change my belief if science proves me wrong” SOUNDS reasonable, but it is the kind of thing homeopaths say to pretend to be scientific while resting secure in the knowledge that they will never have to actually change their mind, because they can always say that it hasn’t been “proven” yet.
There is no “scientific proof” that there are no aliens. There is no “scientific proof” that the earth is 4.7 billions of years old. Not in the sense that there is a “proof” of Bayes’ theorem. And that’s where all the problem is. You can’t limit yourself with changing your believes when they are “proven” wrong. You should change your belief when they are at odd with evidence and Occam’s Razor.
The Dalaï Lama believes in reincarnation (or at least he officially says so, I don’t know what are his true believes and what is political position, but let’s assume he’s honest). There is no “scientific proof” that reincarnation is not possible, so he can’t bolster how much he is open to science. But yet, if you understand science, the evidence that there is no such thing as reincarnation is overwhelming.
People like Carl Sagan were pretty confident that there are aliens out there in the universe. He’s still has a reputation as a great rationalist.
The fact that you would personally Occam’s Razor away reincarnation given what you know, doesn’t mean that it’s also rational for other people to Occam’s Razor it away. Someone who remembers a past life and who knows other people who do it can be common sense to have a prior that reincarnation exists.
If you start with a piror that reincarnation exists I don’t see the scientific evidence that suggest that you should drop the belief. Assuming that reincarnation is true makes some things that involve working with memories of past lifes easier. Occam’s Razor is all about making things easier.
On a practical side there are scientists who believe that reincarnation into Boltzmann brains is plausible, given their current models of how the world works.
If you belief that there are random fluctuation in vaccum, time doesn’t end somewhere in the future and that the existance of humans is completely accidental it’s hard to avoid the conclusion that reincarnation will happen.
If you think you “understand” science than you aren’t rational. Any good skeptic should belief that he doesn’t understand it. Human have the habit of being much to confident in the beliefs that they hold. The Black Swan by Nassim Taleb is a great book.
Has anyone taken the time to present to the Dalai Lama a list of things about Buddhism that science proves (or can convincingly demonstrate to be) wrong?
-- Nate Silver, The Signal and the Noise
(Baruch Spinoza)
-- Dienekes Pontikos, Citizen Genetics
Michael Lewis, Moneyball, ch. 4 (“Field of Ignorance”)
CS Lewis, The Screwtape Letters
-- William Thurston
Marcus Aurelius
Meh, there are worse things to be than a mean man.
There are considerably more worse things to be than a noble one.
— Robert A. Heinlein
I think that quote is much too broad with the modifier “might.” If you should procrastinate based on a possibility of improved odds, I doubt you would ever do anything. At least a reasonable degree of probability should be required.
Not to mention that the natural inclination of most people toward procrastination means that they should be distrustful of feelings that delaying will be beneficial; it’s entirely likely that they are misjudging how likely the improvement really is.
That’s not, of course, to say that we should always do everything as soon as possible, but I think that to the extent that we read the plain meaning from this quote, it’s significantly over-broad and not particularly helpful.
There’s also natural inclinations towards haste and impatience. (They probably mostly crop up around different things / in different people than procrastinatory urges, but the quote is not specific about what it is you could put off.)
That’s certainly a fair point.
I suppose it’s primarily important to know what your own inclinations are (and how they differ in different areas) and then try to adjust accordingly.
I’m reminded of the saying, “A weed is just a plant in the wrong place.” Different people require different improvements to their strategies.
Do it today, and fix/retry tomorrow on failure?
Perhaps it’s a one-time thing.
Albert Einstein (maybe)
Cf. this and this.
--Oliver Sacks regarding patients suffering from “developmental agnosia” who first learned to use their hands as adults.
Can you provide a citation for this? I am very interested in this topic and would like to read the book or article.
It’s in the book “The Man Who Mistook His Wife For A Hat and other clinical tales” chapter 5 “Hands”.
They Might Be Giants
Jon Skeet
--Zhuangzi, being a trendy metacontrarian post-rationalist in the 4th century BC
Zhuangzi says knowledge has no limit, one could spend his entire life making a good map of a vast and diverse territory and it would not be enough to make a good map.
If one does not know this and makes maps for travel, he may be travelling to safe lands. This is weak evidence one is in danger.
If one knows this and still makes such maps, this is strong evidence one is in danger, for to travel to safe lands he would not make such foolhardy attempts.
Michael Lewis, Moneyball, Chapter Ten, “Anatomy of an Undervalued Pitcher”.
Steven Johnson, Everything Bad is Good For You
(His book argues that pop culture is increasing intelligence, not dumbing it down. He argues that plot complexity has increased and that keeping track of large storylines is now much more common place, and that these skills manifest themselves in increased social intelligence (and this in turn might manifest itself in overall intelligence, I’m not sure). Here, he’s specifically discussing video games and the internet.)
I highly recommend the book, it’s interesting in terms of cognitive science as well as cultural and social analysis. I thought it sounded only mildly interesting when I first picked it up, but now I’m thinking more along the lines that it’s extremely interesting. At least give it a try, because it’s difficult to describe what makes it so good.
Really? I thought it was very short and not in depth at all; yeah, his handful of graphs of episodes was interesting from the data visualization viewpoint, but most of his arguments, such as they were, were qualititative and hand-wavey. (What, there are no simplistic shows these days?)
It was rather broad and not very in depth, but it was largely conceptually oriented. He conceded that there were simplistic shows, but argued that the simplistic shows of today tend to be more complicated than the simplistic shows of yesterday. If you disagree...
I don’t know how I’d refute him—there are so many TV shows, both now and then! One can cherrypick pretty much anything one likes, although I don’t personally watch TV anymore and couldn’t do it.
(I’m reminded how people online sometimes say ‘anime really sucked in time period X’, because they’re only familiar with anime released in the ’00s and ‘10s, while if you look at an actual full 30+ strong roster of one of their example ‘sucking’ years like eg. 1991, you’ll often see a whole litany of great or influential series like Nadia, City Hunter, Ranma 1⁄2, Dragon Ball Z, and Gundam 0083: Stardust Memory. Well, yeah, if you forget entirely about them, I suppose 1991 seems like a really sucky year compared to 2010 or whatever.)
Those anime you cite all sucked though, they were considered “great” or “influential” at the time because people didn’t know any better. Anime technology has advanced vastly in the past twenty years.
Anime technology has advanced, yes, but I don’t know how you go from that to ‘all my examples sucked’.
Out of curiosity, what in the 90s compares to Hikaru no Go or Madoka Magica?
Serial Experiments Lain.
Serial Experiments Lain severely disappointed me. It’s nicely creepy and atmospheric but....
(rot13) vg’f ernyyl n fgnaqneq “punatryvat” fgbel—ohg vafgrnq bs snvevrf naq zntvpny jbeyqf naq punatryvat puvyqera, jr unir cebtenzf naq gur Vagrearg naq cebtenzf orpbzvat syrfu.
Gur fpvrapr-svpgvba ryrzragf srry whfg pbfzrgvp punatrf jura gur pber bs gur fgbel vf cher snvel-gnyr… N tevz snvel-gnyr gb or fher, ohg n snvel-gnyr abarguryrff.
I don’t consider Hikaru no Go to be anything more than a gimmick anime like Moyashimon, so I have no idea for it.
The most obvious counterpart to Madoka would be Evangelion (yeah I know Sailor Moon was airing in the ’90s and was more popular and influential than Madoka will ever be, but I think Eva is a better comparison).
Exactly. It doesn’t look like I’m going to finish Hikaru no Go by the end of the year, but I finished Serial Experiments Lain (a 90s anime) in less than 3 days.
Seconding Serial Experiments Lain and Evangelion. Also Cowboy Bebop was in the 90s.
Irresponsible Captain Tylor, Berserk, Excel Saga and Trigun are uneven, but have their moments.
I also have a soft spot for the trashy ultraviolent OVA stuff from the early 90s, like Doomed Megalopolis and AD Police Files, but I’m not sure if it’s good in any objective sense.
Heh… In my myanimelist profile I’ve only listed three anime series as favourites, and Hikaru no Go and Madoka Magica are two of them.
The third one is “Revolutionary Girl Utena”, from the 1990s. I think it’s the sort of series that one either loves or hates—but I loved it.
That was explanation or elaboration, not evidence. I was going to just leave “they sucked” as a bare assertion rather than get into an anime slapfight on LessWrong. If you link me to your anime blog I will be happy to take it up in the comments section there, though.
Alas, I have no anime blog!
You could analyze the way that people in the TV business think and talk about complexity, while assuming that they know what they’re doing. He seemed to do a bit of this.
I’d start by looking at the shows with the highest ratings.
Does he look at the possibility that people are getting more intelligent for some other reason, and popular art is the result of creators serving a more intelligent audience rather than more complex art making people smarter?
No. But your question seems odd. I didn’t interpret the book as an attempt to start with the increase in intelligence and then to assume/explain why pop culture was the cause. Rather, I interpreted the book as an attempt to analyze pop culture, which then found that pop culture did things that seemed like they would have beneficial effects. His analysis of the things that pop culture does in our minds is what I found interesting, not so much the parts which talked about intelligence more generally.
Additionally, I’m not really sure what someone would do to identify pop culture as the cause of this increase as opposed to something else. I’m not sure what other factors could be responsible.
I was reacting to the title of the book.
I don’t believe the title implies that his primary concern is explaining an intelligence increase.
There are two ways of looking at the interaction between pop culture and intelligence. You can start by analyzing intelligence and noticing that it seems to increase, and then trying to figure out why, and then figuring out that pop culture caused it. Or, you can start by analyzing pop culture, and then noticing that it seems to do things that would have cognitive benefits, and then attaching this to the increase in intelligence as a factor that helps explain it. The book does the latter, not the former.
I think any link between tv and intelligence is unproven, but at least the book does something to debunk the common idea that television is making people stupider.
-William of Ockham
This is an interesting quote for historical reasons but it is not a rationality quote.
It makes a very important reply to anyone who claims that e.g. you should stick with Occam’s original Razor and not try to rephrase it in terms of Solomonoff Induction because SI is more complicated.
Humans and their silly ideas of what’s complicated or not.
What I find ironic is that SI can be converted into a similarly terse commandment. “Shorter computable theories have more weight when calculating the probability of the next observation, using all computable theories which perfectly describe previous observations”—Wikipedia.
Hrm.… I’m not sure if I can sufficiently state the speed prior in natural language faster than that, without some really good auctioneer training.
I take it you mean in the sense “Really? Look how terrible the original is! You’ve got to be kidding.”
I read this as a reminder not to add anything to that map that won’t help you navigate the territory. How is this not a rationality quote? Are you rejecting it merely because of the third disjunct?
The quote doesn’t say that, this is (only) a fact about your reading.
I’m not especially impressed with the first two either, nor the claim to be exhaustive (thus excluding other valid evidence). It basically has very little going for it. It is bad epistemic advice. It is one of many quotes which require abandoning most of the content and imagining other content that would actually be valid. I reject it as I reject all such examples.
Luther Sloan to Juilian Bashir in Star Trek: Deep Space Nine, “Inquisition”, written by Bradley Thompson and David Weddle, created by Rick Berman and Michael Piller
Presuming that Starfleet Medical has limited enrollment, and that if he hadn’t lied, a superior candidate would have enrolled, then that superior candidate would have saved those hundreds or thousands, and then a few more.
He was lying about having had gene therapy. He was a superior candidate by virtue of same but it would have kept him out because Starfleet is anti-gene-therapy-ist. (At least I assume so—I remember the character had the therapy and had to hide it, but not whether it came out in that episode or something else did.)
That is much more justifiable than the standard case of lying on applications.
I can imagine Star Robin Hanson writing an angry blog post about what this implies about Starfleet’s priorities.
Have you seen any Star Trek? Star Robin Hanson would have a lot of angry posts to write.
Some, as a child.
There was a (flimsy) historical reason—there had been wars about “augments” in the past; the anti-augments won (somehow), determined the war was about “people setting themselves above their fellow humans”, and discouraged more people augmenting themselves/their children in this way by (ineffectively) making it a net negative.
Heck, anti-fascism beat fascism, it’s not always the stronger-seeming ideology that comes out on top.
Anti-fascism is perhaps more usefully described as pro-something else. In the event, communism.
I read somewhere that, in Star Trek land, genetic engineering of intelligent beings is highly correlated with evil, either because it’s being done for an evil purpose to begin with or because the engineered beings themselves end up as arrogant, narcissistic jerks with a strong tendency toward becoming evil. The latter implies that there’s a technical problem with the genetic engineering of humans that hasn’t been solved yet, which Bashir was lucky to have avoided.
It might not be a technical problem. It might merely be that most augments are raised by people who keep telling them that they’re genetically superior to everyone else and therefore create in them a sense of arrogance and entitlement. Which is only made worse by the fact that they actually are stronger, healthier and smarter than everyone else (but not by as big a margin as they tend to imagine).
Robin Hanson has some interesting things to say about Battle Star Galactica.
This is correct. He lied about not being genetically engineered.
I see no good reason to presume a correlation between a med school’s admissions criteria and total lives saved over a doctor’s career as tight as this reasoning requires. Or to presume that it is near certain that if he hadn’t lied, another liar wouldn’t have been accepted in his place.
This reasoning merely requires that the correlation exist and be positive.
Right but to nitpick just to show off my nerdiness if I recall right Juilian Bashir wouldn’t have been admitted to Starfleet medical because he was a genetically engineered human and those are barred from Starfleet and some other professions because of cultural baggage from the Eugenics Wars.
That was the thing he lied about, so it doesn’t seem likely someone taking his place would have saved more lives in fact he may have saved fewer lives.
Buckaroo Banzai
...
...
dur....
....
What?
I’ll take the new −5 karma hit to point out that this comment shouldn’t be downvoted. It is an interesting critique of the post it replies to.
How is it a critique? The quote is an adequate expression of Eliezer’s own third virtue of rationality, and I daresay if anyone had responded as uncharitably as that to his “Twelve Virtues”, he would have considered ‘dur’ to be an adequate summary of that person’s intellect.
The critique is of the phrase “but to be of no mind whatsoever.”
The uncharitable interpretation is that something without a mind is a rock; the charitable interpretation is to take “mind” as “opinion.”
I ended up downvoting the criticism because it doesn’t apply to the substance of the quote, but to its word choice, and is itself not as clear as it could be.
The criticism is that a martial artist or scientist is actually trying to attain a highly specific brain-state in which neurons have particular patterns in them; a feeling of emptiness, even if part of this brain state, is itself a neural pattern and certainly does not correspond to the absence of a mind.
The zeroth virtue or void—insofar as we believe in it—corresponds to particular mode of thinking; it’s certainly not an absence of mind. Emptiness, no-mind, the Void of Musashi, all these things are modes of thinking, not the absence of any sort of reified spiritual substance. See also the fallacy of the ideal ghost of perfect emptiness in philosophy.
And this critique I upvoted, because it is both clear and a valuable point. I still think you’re using an uncharitable definition of the word “mind,” but as assuming charity could lead to illusions of transparency it’s valuable to have high standards for quotes.
You’ve mentioned this before, and I don’t really know where it comes from. Do you have any specific philosopher or text in mind, or is this just a habit your perceive in philosophical argument? If so, in whose argument? Professional or historical or amateur philosophers?
Aside from some early-modern empiricists, and maybe Stoicism, I can’t think of anything.
Cf. Mushin
I’m amazed how you guys manage to get all that from “dur”. My communication skills must be worse than I thought.
Context helps.
I agree that the response was not particularly charitable, but it’s nevertheless generally a type of post that I would like to see more of on LessWrong—I think that style of reply can be desirable and funny. See also this comment.
My interpretation was that it was advising system 1 rather than system 2 reasoning, thus no mind being no explicit thoughts.
How is it uncharitable? Eliezer is emptying his mind as recommended by Doctor Banzai. Not sure how it’s a “critique” though.
See a priori, No Universally Compelling Arguments.
interesting?
Probably it would be even more interesting if I could understand it.
Eliezer posted a comment that’s essentially devoid of content. This satirizes the original quote’s claim that one should be of “no mind whatsoever” by illustrating that mindlessness isn’t particularly useful—a truly mindless individual (like that portrayed in the comment) would have no useful contributions to make.
That went completely over my head. (I guessed he was alluding to some concept whose name began with “dur”, but I couldn’t think of any relevant one.)
I interpreted ‘mind’ as ‘opinion’, so I didn’t get it either.
“No mind” is ordinary mind.
“Is this a victory or a defeat? Is this justice or injustice? Is it gallantry or a rout? Is it valor to kill innocent children and women? Do I do it to widen the empire and for prosperity or to destroy the other’s kingdom and splendor? One has lost her husband, someone else a father, someone a child, someone an unborn infant… What’s this debris of the corpses?”—Ashoka
-- Ludwig Wittgenstein, Philosophical Investigations
I’ll risk a bit of US politics, just because I like the quote:
Scott Adams on one of the two presidential candidates being skilled at the art of winning (with some liberal use of dark arts).
From “Surely You’re Joking, Mr. Feynman!”: Adventures of a Curious Character
I’ve heard it said that “Trivial” is a mathematics professor’s proof by intimidation.
The view, I think, is that anything you can prove immediately off the top of your head is trivial. No matter how much you have to know. So, sometimes you get conditional trivialities, like “this is trivial if you know this and that, but I don’t know how to get this and that from somesuch...”.
Relatedly, a mathematician friend said that he uses “obvious” to mean “there exists a very short proof of it.” He has been sometimes known to say things like “I think this is obvious but I’m not sure why yet.”
-- G.K. Chesterton
But so does lukewarm water (which is also cheaper, and doesn’t steam up the mirror in the bathroom).
I cannot tell if this is rationality or anti-rationality:
Steve Ballmer
I’d saying telling an interviewer you have sufficient confidence in your product to not need a backup plan is rational, actually not having one isn’t.
I’m reminded of a quote in Lords of Finance (which I finished yesterday) which went something like ‘Only a fool asks a central banker about the currency and expects an honest answer’. Since confidence is what keeps banks and currencies going...
See, if instead of “I’m not paid to have doubts.” he said “I am paid to address all doubts before a product is released”, that would have made more sense.
This comes across as inauthentic and slightly scared to me. At best, he’s not great at PR. At worst, he doesn’t have any back up plan. So that would support calling it irrationality.
Well. I was thinking about it, and it seems like not having a backup plan is the kind of thing that would send bad signals to investors and whatnot. It’s not clear to me that he’s better off doing this than explaining how Microsoft is a fantastically professional company that’s innovating and reaching into new frontiers, etc.
I don’t know specifically what alternate products would potentially be good ideas for them though. I agree that backup plans are good in general but I don’t know if they’re good for Microsoft specifically, based on the resources they have. Windows is kind of their thing, I don’t know if they could execute on anything else.
Ambrose Bierce
-Bryan Caplan, Selfish Reasons to Have More Kids
I’d pick gripping lies over most nonfictional shows, which are mainly irrelevant or misleading truths.
-- Iain McKay et al., An Anarchist FAQ, Sec. F.2.1
Salman Rushdie, explaining identity politics
I think “identity politics” is a term of art which covers things other than that which aren’t bad, like minority struggles.
You’ve got a point, and it’s one that gets into hard issues. It can be quite hard for some people to decide whether they’re being unfairly mistreated and to act on it, and the people for whom the decision is easy aren’t necessarily sensible. Emotions are not a reliable tool for telling whether acting on a feeling of being unfairly mistreated makes sense.
How do you tell to what extent is a particular instance of people feeling outraged them just getting worked up for the fun of it over something they should endure, and to what extent are they building up enough allies and emotional energy to deal with a problem which (by utilitarian standards?) needs to be dealt with?
I don’t know how to tell legitimate movements from illegitimate ones, but the term of art “identity politics” refers to both. ID politics is a specific kind of political advocacy, and there are both good ID politics arguments and bad ones. You’d probably just have to investigate the claims they’re making on a case by case basis.
But, I wasn’t trying to interrogate whether defining yourself by outrage can be good in some instances, I was trying to point out that the term “ID politics” refers to things outside of defining yourself in relation to outrage. Maybe I just misinterpreted what you were saying, but I thought your comment unintentionally hinted that you were unaware the phrase is a specific term of art. There are many types of identity politics that aren’t about outrage or opposition.
You’re quite right, I didn’t know about it as a term of art.
I suppose I’ve mostly heard about the outrage variety of identity politics—it tends to be more conspicuous.
More from Scott Adams:
Warning: Your milage may vary.
“The reasonable man adapts himself to the world; the unreasonable one persists in trying to adapt the world to himself. Therefore all progress depends on the unreasonable man.” -George Bernard Shaw
Sadly, duplicate.
Tenzin Gyatso, 14. Dalai Lama
That’s intriguing, but it also sounds like a case of non-apples.
Well, it is a necessary step to find other fruits.
Descartes, in Rules for the Direction of the Mind
A related Sherlock Holmes quote:
That’s from The Red-Headed League.
I’ve been collecting examples of this (or something more general but similar) under the name “brain-as-rock fallacy”, though there’s probably a better and less ambiguous name.
I don’t get what you are hinting at with “brain-as-rock”, could you please explain?
“Brain as ideal decision-making engine unaffected by mere external or physical facts.”
Open question: Do you care about what (your current brain predicts) your transhuman self would want?
If you don’t, you’re really going to regret it in a million years.
I’m rather skeptical about that, even conditioning on Ezekiel being around to care. I expect that the difference between him having his current preferences and his current preferences+more caring about future preferences will not result in a significant difference in the outcome the future Ezekiel will experience.
The chance of human augmentation reaching that level within my lifespan (or even within my someone’s-looking-after-my-frozen-brain-span) is, by my estimate, vanishingly low. But if you’re so sure, could I borrow money from you and pay you back some ludicrously high amount in a million years’ time?
More seriously: Seeing as my current brain finds regret unpleasant, that’s something that reduces to my current terminal values anyway. I do consider transhuman-me close enough to current-me that I want it to be happy. But where their terminal values actually differ, I’m not so sure—even if I knew I were going to undergo augmentation.
Yes, I think so. It surely depends on exactly how I extrapolate to my “transhuman self,” but I suspect that its goals will be like my own goals, writ larger
-- Motaigne
From “An Elementary Approach to Thinking Under Uncertainty,” by Ruth Beyth-Marom, Shlomith Dekel, Ruth Gombo, & Moshe Shaked.
Or not.
Arguably, assigning a particular floating point number between 0.0 and 1.0 to represent subjective degrees of belief is a specialized skill and it could take years of practice in order to become fluent in numerical-probability-speak.* Another possibility is that it merely adds a kind of pseudo-precision without any benefit over natural language.
In any case, it seems to be an empirical question and so should be answered with empirical data. I guess we won’t really know until we have a good-sized number of people using things such as PredictionBook for extended periods of time. I’ll keep you posted.
*There does exist rigorously defined verbal probabilities, but as far as I know they haven’t been used much since the Late Middle Ages/Early Modern Period.
I’d like to see more on those verbal probabilities, having stated to use my own since few satisfactory existing versions.
Potest legistis linguam Latinam? If not, then you might want to read The Science of Conjecture: Evidence and Probability before Pascal by James Franklin for an overview of the tradition I was referring to.
I’ll give it a look.
Nietzsche, The Gay Science
[oops; this was a repeat]
Repeat.
.
Can you elaborate on what this is getting at?
.
You shouldn’t be decieved by the use of the word formal as an applause light.
Linus Pauling
Edit: another one captured by an old thread!
From the alt-text in the above-linked comic:
Nietzsche, The Gay Science
--Satoshi Kanazawa
That would be much more convincing coming from literally anyone other than Kanazawa. It takes very little charity to interpret his critics as saying, not “Your theories are inherently racist” but “Your theories are only some of many compatible with your findings; you are privileging them because you are biased in favor of hypotheses that postulate certain races naturally do worse than others”.
I don’t know what to learn from the quote. It’s literally true, but it’s also clearly unhelpful, since Kanazawa writes this while following non-truth-seeking algorithms. Maybe the moral is “If someone calls you a mean name, address the content of the criticism and not whether the mean name applies”, or maybe “Don’t be a giant flaming hypocrite”.
He isn’t a great scientist in my mind since he seems to often just lazily reverse stupidity, but it was a good quote.
This is my home, the country where my heart is;
Here are my hopes, my dreams, my sacred shrine.
But other hearts in other lands are beating,
With hopes and dreams as true and high as mine.
My country’s skies are bluer than the ocean,
And sunlight beams on cloverleaf and pine.
But other lands have sunlight too and clover,
And skies are everywhere as blue as mine.
-Lloyd Stone
Attributed to Voltaire
St. Thomas Aquinas
Edit: Oops, accidentally created a repost.
Eminem, “Love The Way You Lie”
Baudrillard, In the Shadow of Silent Majorities
I was curious why the Baudrillard comment was downvoted when it expresses the same idea as the Nietzsche comment, it just uses a different style and approaches the problem from a different direction. Ideas, anyone?
Priming? Beaudrillard is associated with humanities, pomo and academic philosophy; Nietzche is associated with atheism, contrarianism and the idea of the ubermensch. The comment doesn’t seem to be very strongly downvoted; possibly you’re just dealing with detractors here (I daresay LW has more fans of the latter than of the former).
This was roughly my thought as well. I thought there might also have been more substantive differences though and I was curious what those might be. The only thing I could see is that Baudrillard’s quote had a tone that’s more critical of the masses and the way they do politics, and that Baudrillard’s quote could be misread as an injunction to stop trying to make people rational (which it’s not).
Well, I’m not even sure whether Boudrillard’s quote is grammatically well-formed, so there’s that. Then again, postmodernist texts tend to be imbued with near-poetical and mystical qualities. Much like Zen koans, they’re more about exemplifying a particular mind-posture and way of thinking than they are about straightforward argumentation. I think it’s unfair to expect LessWrongers to be familiar with such texts.
Oh. So this quote is difficult to read, then? More difficult than the Nietzsche one? I guess inferential gaps must be coming into play here. I’m having a difficult time trying to not-understand it, trying to emphasize with your viewpoint. I’m having a difficult time believing that you couldn’t understand the quote, honestly.
I feel like you’re generalizing too much about post modernism. I like lots of it, and don’t think that it’s mystical oriented. I would say rather that it packs a lot of information into a small amount of words through the clever use of words and through recurring concepts and subtle variations on those concepts.
Post modernism can be difficult to understand, but I don’t think it is in this case, and I think that it’s complexity is justified. Scientists use obscure terminology, but for a good purpose, generally. Some scientists use obscure terminology to hide the flaws in their ideas. I view post modern criticisms in almost exactly the same way—their complexity can be for both good and bad.
Also, Baudrillard is French. It might not be his fault if there’s problems with the translated text.
I’m using “mystical” in a rather specialized sense, actually. What I mean is that postmodernist texts seem to eschew straightforward arguments—instead they use rhetorical and poetical patterns in a functional way, to inspire a specific mental stance in the reader. This mental stance might be quite simply described as “emptying the teacup”, i.e. questioning and letting go of the “cached thoughts” which comprise one’s current understanding of reality and culture. This mental stance happens to be remarkably useful in textual criticism and social science, where one often has to come to terms with (and perhaps reconstruct, at least partially) cultures which are far apart from one’s own, so that a “filled cup” would be a significant hindrance.
Oh, and yes, I had quite a bit of trouble with trying to understand the Beaudrillard quote, although I did grok the gist of it, and I also got the similarity wrt. the Nietzsche one. But I’d say that grammar is clearer in Nietzsche’s quote, and even his rhetoric seems more direct and to the point here.
Okay, gotcha. Thanks.
Nope, I’m a native French speaker and my reaction to Baudrillard is “WTF?” and building a Markov Baudrillard quote generator to see if I can tell the difference.
Jargon is good. Vaguely defined jargon isn’t bad—sometimes all you can do is say “sweet refers to the taste of sugar, if you don’t know what that is I can’t help you”.
But structure shouldn’t be completely unclear. Baudrillard has a lot of “X is Y” statements and very few “therefore”s. I can’t tell what is a conclusion, what is an argument, what is a definition, or even whether there are anything but conclusions.
I’ve found some Baudrillard texts that clearly mean things, but they’re not very good.
Can you specify more about what parts of the quote are confusing?
This one isn’t that bad. (For utter, words-don’t-work-that-way confusion, see Debord. Or good ol’ Hegel.)
That bit is straightforward.
“The masses” has a standard denotation but various connotations. Freddy Nietzsche talks about enthusiastic young people, which is more specific.
What’s “to keep within reason”? What this evokes is talking someone down, preventing outbursts. Applied to the masses, does he mean control—propaganda, opiate of the masses? The context suggests the opposite: to present a logical argument and try to convince audiences with it as the core of communication, more important than ethos and pathos and Cheetos.
What?
Okay, “imperative” seems to mean what social justice types can “enforcement by shaming”. If you don’t talk like a Vulcan, whoever is producing those great media reform plans (pretentious elites?) will shame you.
Okay, so media becomes morally loaded: information good, fluff bad. Much like food is morally loaded: vegetables good, fat bad.
Examples! Hallelujah, hosanna in excelsis! So the media reformers want to make people better. If you say a thing and hearing it doesn’t make listeners better, you’re selling junk food.
That seems pretty clear too: logical arguments aren’t what convinces people. Nietzsche says that too, but in a more specific context: recruiting for a cause.
I assume this means: “the masses decide what they want to take from what they hear, and it’s not logical argument, it’s”
I’ll grant that “spectacle” is a totally precise and useful term of art that people clearly define whenever I’m out of earshot. But if he’s saying what Fred says, he doesn’t need the jargon; it’s not a rare concept.
Freddypants is saying “If you want a young, energetic, status-seeking enthusiast to be enthusiastic about your cause, don’t bother calmly explaining why your cause is good. Instead, make it look awesome and promise exciting heroics.”. (Which he what he does in Zarathustra, and it worked on me but I already agreed.) Baudrillard appears to be saying “If you want to convince people, calm explanations won’t work.”.
Okay, thank you.
I agree that Hegel is ridiculously opaque, too.
“The world is just a word for the things you value around you, right? That’s something I’ve had since I was born. If you tell me to rule such a world, I already rule it.” – Tohsaka Rin (Fate: stay night) on not taking over the world.
I think it is having a small core of things and people you value that keeps you grounded and healthy. Our “Something to Protect” if you like. Without that investment and connection to things that matter it’s easy to lose your way.
No, that’s never how I’ve seen anyone define ‘world’. Maybe that quote makes more sense in context.
The character was just asked whether they would wish to conquer the world if given a wish-granting machine (and are saying no, they already have what they want and value). The way I understood the quote was that when people talk about ruling the world they really just want to control and protect the things they value around them. It made me think that “the world” isn’t really a concept that people can easily grasp in the abstract, they need to look at the smaller scale to give them context.
I think “I want to protect humanity” or “I want to save the world” carry more weight and are easier to follow through on if you come at them from the angle of “I want to protect people like the people around me I love” or “I want to save the place where people like my friends and family live”.
I’m not sure I follow even with that explanation, but I’ve never really known what to make of the Nasuverse in the first place. (“This is so awesome!” “But also incredibly stupid.” “But awesome!” “But stupid. And ad hoc. And ill-thought-out.” “Aw, don’t be like that, just enjoy the Rule of Cool.”)
I imagine Eliezer would answer something like “No, that would be redundant.” (Edit: not to credit Eliezer with inventing the concept.)
--Mencius Moldbug on an experiment that has interesting results
--Some dude in the comment section of West Hunters
I need to hunt down the source for his claim so take it with a grain of salt.
Memories can be vile, repulsive little brutes. Like children, I suppose. haha.
But can we live without them? Memories are what our reason is based upon, if we can’t face them, we deny reason itself! Although, why not? We aren’t contractually tied down to rationality!
There is no sanity clause!
-The Joker, A Killing Joke
Obsoleted by sticky notes.
“A mind to which the stern character of an armchair is more immediately apparent than its use or its position in the room, is over-sensitive to expressive forms. It grasps analogies that a riper experience would reject as absurd. It fuses sensa that practical thinking must keep apart. Yet it is just this crazy play of associations, this uncritical fusion of impressions, that exercises the powers of symbolic transformation.”
Susanne Langer, Philosophy in a New Key
I don’t understand this. What are “the powers of symbolic transformation”?
To finish it: “To project feelings into outer objects is the first way of symbolizing, and thus of conceiving those feelings. This activity belongs to about the earliest period of childhood that memory can recover. The conception of ‘self,’ which is usually thought to mark the beginning of actual memory, may possibly depend on this process of symbolically epitomizing our feelings.”
I’m afraid this simply leaves me more confused—could you explain what Susanne Langer is getting at in your own words?
The book is from 1942, so is dated in thought and is linguistically frilly.
What she was getting at here, I think, is that when we think of a baby learning to speak, and it is learning the word “mama.” Probably most of the time it will come to speak the word “mama” were times that before were purely emotional. That is mother has a certain emotional valence as an object and at specific times will have even more emotional significance to the baby. Langer is exploring how language takes over for those feelings and will sit on top of those feelings (and goes on to shape and reshape and limit those feelings). “Mama” comes to represent mother, which before the word becomes accessible was some kind of blind object-feeling-sense data bundle. She claims earlier that the baby’s ~”thoughts” are very synaesthetic, sense-data is muddled, something she takes from the James view.
“Powers of symbolic transformation” just means the ability to use language, but this she claims is something that takes time. When we come to be skillful masters of language we have long separated the emotional raw feelings from the word-object relationship in itself, though our words are always very much tinged by those earlier feelings (she presents a high brow picture where only the thoroughly educated reach the best forms of these states, i.e. philosophers).
Her point: “Speech is through and through symbolic; and only sometimes signific.”
Speech (and a good “mind”) is for abstract manipulation of concepts that help us understand the world, and to accomplish this we must painstaking remove emotions and feelings that attach to our words, e.g. the sternness of chairs.
It sounds somewhat related to the idea that words are based less on definitions than prototypes—yes, penguins, ostriches, and hummingbirds are all birds taxonomically, but the word “bird” evokes ducks, sparrows, and eagles much more easily than those.
.
He is probably referring to nuclear fission reactors and/or genetically modified crops.
— Michael Kirkbride / Vivec, “The Thirty Six Lessons of Vivec”, Morrowind.
Andrew Tobias (Warning: he is a DNC treasurer, so by no means rational in the matters of politics, only finance.)
Well, probably not even rational there.
--Aleksandr Solzhenitsyn
Edit: Ooops old thread.
The potential for abuse of this quote is too high. While it’s an example of how even absurd amounts of research can fail to move a religious thought, too many people will fail to get the joke.
--”um” on me posting that quote
Harry Potter, in Harry Potter and the Methods of Rationality by Eliezer Yudkowsky
That arguably counts as LW/OB.
If HPMoR isn’t allowed, that should be specified in the rules.
I mean that it’s a nice quote, but I suspect that’s the reason for the downvotes.
I’m not overly impressed with the quote either. Sometimes you can understand things and still not command them. Sometimes you just lose and all understanding you can get will just tell you to go do something else that you can control.
It is arguably a lot more affiliated to LW than OB is. (We successfully got OB removed from the no-quote list at some stage. Unfortunately someone reverted it.)
I understand people. Imperius! People, I command you to build me a moon rocket!
-- Draco
...That might actually work, as long as he understands which people to Imperius.
Junot Diaz
A different perspective on a phrase our community holds near and dear
If that’s true, then there could be no use in finding a place because you would then follow the quote’s advice and never return again!
Per Bohr’s advice we can identify this as a meaningless ‘profound truth’ by reversing it:
I took the quote as a call to explore. Don’t just be satisfied with learning things other people have figured out, try to creatively venture into the unknown yourself.
http://lesswrong.com/lw/8n9/rationality_quotes_december_2011/5czw
If you are on Earth, then you’re at a place someone has already found.
I take Junot Diaz to mean “place” metaphorically, not geographically.
Love is experienceable, not intelligible.
Anonymous.
-Harry James Potter-Evans-Verres, Harry Potter and the Methods of Rationality
Downvoted since HPMoR is or should be included in the “Do not quote comments/posts on LW/OB” rule.
But it’s not.
Kinda mean to downvote somebody for breaking a rule that doesn’t exist (yet), don’t you think?
No mean-ness intended, and it’s only one downvoted. If we let HPMoR quotes in here we’d be spending this whole thread reading HPMoR which isn’t particularly helpful.
I was also under the impression that it had been at some previous point specifically mentioned in the rules. And considering HPMoR as a part of LW is not unreasonable: you yourself did that here.
Agreed. Retracted.
Not especially. It’s using downvotes exactly as intended. It would be fine even if there was an explicit endorsement of HPMoR quotes as being sufficiently external to qualify for the thread, if tgb just happened personally not to like them.
Incidentally, I down-voted a quote of Eliezer from TDT a week ago and just didn’t think it was worth making a comment about it. Really, Eliezer making one of his essays slightly longer than the others and publishing it in a very slightly more formal form doesn’t change the nature or role significantly. (Mind you, I would likely have ignored it or upvoted if the specific quote was sufficiently impressive to make up for it.)
EDIT: It could actually be mean, denotatively but if so it does not have the features that some other mean things do that make them not an acceptable and sometimes virtuous thing to do.
False. Criticism and optimization overlap highly in the nature of the reasoning but it is possible to construct an optimizer that is not capable of criticism. If you really want one. I’m not sure what exactly Harry’s intended human-level practical point is—some applications apply even though the claim is technically false but some may not.
I believe his human-level point is that if you are unable to find a problem with a current system, you will believe optimising it is impossible.
— Michael Kirkbride / Vivec, “The Thirty Six Lessons of Vivec”, Morrowind.
Note for the unfamiliar: this exchange occurs in the Temple of False Thinking.
You know too much.
Now I am the mightiest of your children?
— Michael Kirkbride / Vivec, “The Thirty Six Lessons of Vivec”, Morrowind.
I don’t get it.
-- Clarice Lispector.
Original in Brazilian portuguese:
Before you downvote: I think this has a lot to do with rationality because we tend to be caught up in thinking and the models about the world we create in our minds, actually science is about this. But those models have limitations and are often wrong as the history of science shows time and again.
EDIT: added the originals. Fixed typo.
Now that you have noticed this, what are you going to do with it?
Realize that your mental models might be wrong and don’t put too much weight on them, instead put more weight on your feelings.
Trying to make better models does not appeal to you?
Do you have good evidence that your feelings are more often correct than your models?
Feelings honed by millions of years of evolution.
To what extent can you expect evolution to have prepared you for your day-to-day experience?
Is this a serious question? While the modern world might have changed in a lot of aspects a big factor remains constant: people, social interactions. What use is it to choose the logically correct decision if it still makes us feel miserable?
What use is it to feel miserable despite having made a correct decision?
It might be hard to change our feelings. Should a correct decision make us feel miserable? Maybe there is a better decision that also makes us feel good? Also relevant see my answer here: http://lesswrong.com/lw/ece/rationality_quotes_september_2012/7qmo
There are situations where your feelings are more reliable than your models. Are there situations where it is the other way around? How do you decide which to use?
I don’t intended the original quote to be an admonition against all use of models/reasoning. My point was more or less along the lines of “listen to your feelings, they might be telling you something important. Don’t disregard them just because you have some neat model, your model could be wrong.”
I agree, but that does not answer the question. How do you decide which to use? What do you need in order to decide?
This boils down to: when do you know that your models are correct? And the answer is, you almost never know, unless it is already settled by science and even then there is room for error and further correction down the road(years away). But you need to make decisions now, every day.
Almost. It boils down to: when do you know that your models are correct and when do you know your feelings are correct. Well, how do you settle that question?
I don’t know, but I have the impression that you have an answer in mind, care to share?
chaosmosis said it already :)
You don’t have to treat your feelings and your models differently. Just use whichever one the evidence suggests is more likely to be correct in whichever situation you find you find yourself in. See?
Sounds good, but you still have to decide which one is more likely to be correct, so it doesn’t seem to solve the fundamental question at hand.
Unfortunately, I can’t help you with that, as you have your own models and feelings. You’ll have to collect data on your own about which works better in what situation. You can probably start by going over past experiences to see if there are any apparent trends, and then just be mindful of any opportunity you might have to confirm or disconfirm any hypothesis you might generate. Watch out for unfalsifiables!
Empiricism and logic? Just treat your emotions like a model, and judge them like you would any other. Even though you can’t see the inside of your emotions, neither can you see the inside of the thought processes that produce the model. I don’t see why there would be any difference between the two.
yes
You might want to remove the space at the beginning of the line. It’s distracting to have to use the scrollbar to read the full quote.
also, how is math the only system able to prove that it has unproveable truths? that was missing from my copy of Godel’s theorems.
Did you mean to reply to kateblue rather than me? (Or did you want to evade the karma fee?)
I don’t understand—what fee? Would Jonathan_Graehl get more downvotes if he replied directly to kateblu? Why?
There’s a new “feature” that replies to sufficiently negative karma posts instantly lose 5 karma.
i was hoping you’d let my cowardice pass unnoted.
Wait actual humans are afraid of losing karma?
...i dont even
Actual humans are afraid of being considered obnoxious, stupid or antisocial. Karma loss is just an indication that perception may be heading in that direction.
Attempts to avoid karma loss by procedural hacks are a stronger indication...
This is how lost purposes form. Once you’ve figured out that karma loss is a sign of something bad, you start avoiding it even when it’s not a sign of that bad thing.
Maybe wedrifid is taking that into account and renormalizing. It’s hard to tell.
Of something different.
Of people assigning excessive weight to very small changes in signaling?
#noshitsherlock
Yes.
I don’t think it’s any weirder than viewing losing karma as an entertaining game.
Only on LessWrong would a statement with that much insight be downvoted because it could be taken to signal something vaguely positive about religion.
The quote, phrased in a less tortuous way, says that mathematics contains true statements that cannot be proven, and is unique in being able to demonstrate that it does. So far, so good, although the uniqueness part can be debated.
But the quote also states that mathematics therefore contains an element of faith, that is, that there exist statements that have to be assumed to be true. This is not the case.
Mathematics only compels you to believe that certain things follow from certain axioms. That is all. While these axioms sometimes imply that there exist statements whose truth will never be determined, they do not imply that we should then assume that such-and-such a statement is true or false.
That is why it should be downvoted. Because not knowing something doesn’t mean having to pretend that you do.
I was tempted to downvote it because it could be taken to be negative about math.
It sounds to me like a goofy language game, akin to “How many legs does a dog have if we call a tail a leg?”
That conundrum, to which the correct answer is “four”, is not a goofy language game. It is making the point that you cannot change the truth of a proposition by changing the meanings of the words in it. When you change the meanings of the words, you are creating a different proposition. It looks like the original one, because it consists of the same string of words, but it is not. Its truth need have nothing to do with the truth of the original one.
Would you still be able to see these words if we called black white?
I always hated that question due to its ambiguity. Those who state the answer is four legs seem to interpret the question as asking: “Labeling our current language as Language-A, and mentioning a different language Language-B in which ‘leg’ also refers to tails, and keeping in mind that we do not speak Language-B, how many legs does a dog have?”
However, for some reason I first interpreted the question as asking: “Labeling our current language as Language-A, and mentioning a different language Language-B in which ‘leg’ also refers to tails, what is the answer to ‘how many legs does a dog have?’ in Language-B?”
I apologies for both the brevity and ambiguity of these interpretations. However, I doubt that I am the only person who interprets the question in something along the lines of my fashion.
Definition is the basis of language. Without a common understanding of terms, there can be no discussion. Anything that has not been falsified is theory unless it is proven to be true. Without a common understanding of terms, how can we know that a statement has been proven false? Mathematics is the most rigorous language in the sense that there is nearly universal understanding of terms among professional mathematicians, but it is still a language. The answer to your question is unambiguous; if a dog has a set of appendages that we will call “Legs” that consists of four of what we commonly call legs plus one tail, then the number of elements in the set of “Legs” is equal to 5. We could say that the set L = {a,a,a,a,b). Either way, it is simply a matter of definition—not really a ‘goofy game’.
Be wary when issuing grand proclamations about language, lest you wind up looking silly to the linguistically-knowledgeable.
I think you have it the other way around. Definitions are based on language. Language is based on meaning. I knew the meaning of the word “red” before I had any definition for it, and I’d guess that so did you.
(with a smile) Perhaps we need to define definition. True that definitions are based on language. Also true, I believe, that if language is to communicate effectively, it will need commonly understood meanings for specific sounds/symbols. I may “see as red” what you “see as orange”. My guess is that we both saw and could differentiate between colors before we knew the commonly accepted terms for them.
I had assumed the audience had heard the joke before. The punch line: “Four. Calling a tail a leg doesn’t make it one.”
Which is the sort of thing that could be called “problematic on so many levels” — or just “goofy”.
Modus pwnens, modus trollens.
That sounds positive about religion to you?
It vaguely associates math with religion in the sort of mind that is in the human that reads LessWrong. In such minds that means it is both saying something positive about religion and negative about math. That means it deserves twice the downvotes obviously. Let’s downvote it, Nominull! I’ll downvote it if you do. Let’s start a circle-jerking session, Nominull. I will if you do. Agreed?
Circle-jerkIng is like more like a stag hunt; only implicit cooperation and an explicit lack of shame are required.
-L. A. Rollins, Lucifer’s Lexicon: An Updated Abridgment
You should be less transparent about your social psychology experiments if you don’t want people like me to make them transparent to everyone else.
You should be less transparent about your social psychology experiments if you don’t want people like me to make them transparent to everyone else. I like to disrupt things, you see. So does reality.