Open thread, Nov. 17 - Nov. 23, 2014
If it’s worth saying, but not worth its own post (even in Discussion), then it goes here.
Notes for future OT posters:
1. Please add the ‘open_thread’ tag.
2. Check if there is an active Open Thread before posting a new one. (Immediately before; refresh the list-of-threads page before posting.)
3. Open Threads should be posted in Discussion, and not Main.
4. Open Threads should start on Monday, and end on Sunday.
Recently I have been thinking about imaginary expertise. It seems remarkably easy for human brains to conflate “I know more about this subject than most people” with “I know a lot about this subject”. LessWrongers read widely over many areas, and as a result I think we are more vulnerable to doing this.
It’s easy for a legitimate expert to spot imaginary expertise in action, but do principles exist to identify it, both in ourselves and others, if we ourselves aren’t experts? Here are a few candidates for spotting imaginary expertise. I invite you to suggest your own.
Rules and Tips vs Principles
At some point, a complex idea from [topic] was distilled down into a simple piece of advice for neonates. One of those neonates took it as gospel, and told all their friends how this advice formed the fundamental basis of [topic]. Examples include “if someone touches their nose, they’re lying” and “never end a sentence with a preposition”.
If someone offers a rule like this, but can’t articulate a principled basis for why it exists, I tend to assume they’re an imaginary expert on the subject. If I can’t offer a principled basis for any such rule I provide myself, I should probably go away and do some research.
Grandstanding over esoteric terminology
I’ve noticed that, when addressing a lay audience, experts in fields I’m familiar with rarely invoke esoteric terminology unless they have to. Imaginary experts, on the other hand, seem to throw around the most obscure terminology they know, often outside of a context where it makes sense.
I suspect being on the receiving end of this feels like Getting Eulered, and dishing it out feels like “I’m going to say something that makes you feel stupid”.
Heterodoxy
I have observed that imaginary experts often buy into the crackpot narrative to some extent, whereby established experts in the field are all wrong, or misguided, or slaves to an intellectually-bankrupt paradigm. This conveniently insulates the imaginary expert from criticism over not having read important orthodox material on the subject: why should they waste their time reading such worthless material?
In others, this probably rings crackpot-bells. In oneself, this is presumably much more difficult to notice, and falls into the wider problem of figuring out which fields of inquiry have value. If we have strong views on an established field of study we’ve never directly engaged in, we should probably subject those views to scrutiny.
I agree with what you wrote. Having said this, let’s go meta and see what happens when people will use the “rules and tips” you have provided here.
A crackpot may explain their theory without using any scientific terminology, even where a scientist would be forced to use some. I have seem many people “disprove” the theory of relativity without using a single equation.
If there is a frequent myth in your field that most of the half-educated people believe, trying to disprove this myth will sound very similar to a crackpot narrative. Or if there was an important change in your field 20 years ago, and most people haven’t heard about it yet, but many of them have read the older books written by experts, explaining the change will also sound like contradicting all experts.
In response to your second point, I’ve found “field myths” to be quite processable by everyday folk when put in the right context. The term “medical myth” seems to be in common parlance, and I’ve occasionally likened such facts to people believing women have more ribs than men, (i.e. something that lots of people have been told, and believe, but which is demonstrably false).
It does seem a bit hazardous to have “myths” as a readily-available category to throw ideas in, though. Such upstanding journalistic tropes as Ten Myths About [Controversial Political Subject] seem to teach people that any position for which they hold a remotely plausible counterargument is a “myth”.
Status of your Domain
There was a post some time ago which I can’t find right now which talked about the danger of one field looking down upon other fields. The prototype example is hard sciences looking down on soft sciences. The soft sciences are seen as less rigorous or less mathematically well founded and thus if you belong to the higher domain you think yourself better qualified to judge the presumably less rigorous fields.
It was this post by Robin Hanson. However, I am not sure that “status” (which suggests having something like a one-dimensional order) is the right explanation here. For example, philosophers tend to say things about a lot of other fields and people from other fields tend to say a lot of things about philosophy. Therefore, unless one considers all those fields to be of equal status, one-dimensional order doesn’t seem to be applicable as an explanation.
It seems to me that:
if field A has a lot of widely applicable methods that are abstract enough and/or have enough modelling capabilities so as to be able to model a wide variety of objects, including most/all objects that are the object of research of field B, then the experts of field A will often express their opinions about the object of research of field B, theories of field B, and the state of field B itself. In other words, if field A has a widely applicable powerful method, then experts of field A will try to apply this method to other fields as well. For example, physics has a lot of powerful methods that are applicable well beyond physics. Some ideas developed in philosophy are often abstract enough so as to be (or seem to be) applicable to a wide variety of things. Note, however, that in some cases these methods may fail to capture important features of the object of field B. On the other hand, experts of field B have incentives to claim that methods of field A are not applicable to their field, because otherwise field B is redundant. In response, they may start to focus on the aspects of their field that are harder to model using field A’s tools, whether those aspects are important or not. Yet another interesting case is when there are two fields A and A_1 whose methods are capable of modelling certain aspects of the object of research of field B. Then experts of A may denounce the application of A_1′s methods to the field B (and vice versa) as failing to capture important things and contradicting findings of field A.
if the object of research of field A is closely related to human experiences that are (very) common and/or concepts that are related to common sense concepts (e.g.cover the same subject), then people of other fields will often express their opinions about the object of field A, theories of field A and field A itself. For example, many concepts of philosophy, psychology (and maybe most/all social sciences) are directly related to human experiences and common sense concepts, therefore a layperson is more likely to speculate about the topics of those fields.
Well, of course there is the idea that the findings of two research fields A and B should not contradict each other. And there is this informal “purity” rank, i.e. if A is to the right of B, then experts of A will regard findings of B as wrong, whereas experts of B will regard theories of A as failing to capture important features of the object of research of their field. However, it doesn’t seem to me that all fields can be neatly ordered this way. One reason is there is at least two possible partial orders. One is the “reductionist” way, i.e. if the object of field B is made out of the stuff that is the object of field A, then field A is to the right of object B. Another way is to arrange fields according to their methods. If field B borrows a lot of methods from the field A, but not vice versa (in other words, methods of A can be used to model the object of field B), then field B can be said to be to the left of A. In some cases, these two orders do not coincide. For example, the object of economics is made out of the object of psychology, however, it is my impression that most subfields of economics borrow their methods directly from mathematics (or physics), i.e. in the second partial ordering psychology is not between economics and mathematics. By the way, conflating these two orders might give rise to intuitions that are not necessarily grounded in reality. For example, mathematics (despite the fact that not all physicists’ results are formalized yet), is probably to the right of physics when we order fields according to their methods. That gives us intuition that mathematics is to the right of physics in the “reductionist” order as well, i.e. the object of physics is made out of mathematical structures. Well, this might turn out to be true but you should not proclaim this as casually as some people do.
Some indications can be taken from Some Heuristics for Evaluating the Soundness of the Academic Mainstream in Unfamiliar Fields:
Low Hanging Fruit Heuristic
Thus: You are likely having imaginary expertise if all low hanging fruits in your field have been taken.
Ideological/venal Interest Heuristic
You should be doubtful about your expertise when
Note that these concern your expertise despite you actually being an expert.
You should note that these cut both ways—given a field of research, they reduce (ceteris paribus) the trustworthiness of both layperson’s independent attempts to form an opinion about this field and the trustwortiness of the established experts in that field.
Making predictions about the subject matter and seeing whether those come true, is one of the great ways to measure expertise.
I blame World of Darkness circa 1998.
So kind of like an intellectual chuunibyou.
Sounds like a pretty good idea to me. You’ve listed some symptoms, I guess if you catch yourself engaging in such behaviour, there are potential solutions?
Avoid heterodoxy by reading opposing viewpoints while steelmanning as much as possible.
To avoid grandstanding via vocab, keep things simple. Make sure you can play taboo with words, or up goer five.
For rules and tips, make sure you’re clear with yourself on when your rules don’t apply as well as when they’re useful.
But really, it is a hard one to notice in yourself, and a hard one to fix without actually becoming an expert.
I have an ongoing project to read an introductory textbook for every subject I claim to be interested in. This won’t make me an expert in those subjects, but it should hopefully stop me saying things that cause actual experts to facepalm.
I present exhibit A.
(Meant lightly, but this really is a good example of phrasing that could be unnecessarily confusing to an average reader.)
I can see what you mean, but rephrasing what someone else has said is the opposite of using unnecessarily confusing phrasing, no? It’s just good practice, and a big part of what the guide to words stuff was all about.
But you’re right, I should probably put in a link for steelmanning, just in case.
Great insight. I added it to my Anki deck. I think this could have been a Discussion post of its own (even at this length).
If I collect any more indicators of imaginary expertise, I may assemble it into a discussion post.
You’ve got a worthwhile project there, but I gather that in sufficiently complex fields (like engineering) there’are a lot of rules and tips because there isn’t a complete set of principles.
Can you provide any examples? I have introductory exposure to engineering, but can’t readily pinpoint the sort of rules or tips you’re referring to.
Page down to Assumptions, Preditions, and Simplifications.
I didn’t actually have any examples—the idea that engineering has methods which aren’t firmly based in math or physics was just something I’d heard enough times to have some belief that it was true. The link has some examples.
I’m not sure these are quite the same “rules and tips”. There’s still a rationale behind model assumptions that an expert could explain, even if that rationale is “this is essentially arbitrary, but in practice we find it works quite well”.
The reasoning for “if someone touches their nose, they’re lying” is: If someone lies they feel anxiety. It’s well known that anxiety can produce effects like someone getting red because of changes in blood circulation due to tension changes. Some tension changes can make a person want to touch the corresponding area.
Something along those lines is the motivation for that tip. The problem is that there are many reason why someone touches their nose and it’s no good tell unless you calibrated it for a person.
Of course you could say that the fact that I know the explanation suggest that I know something about the topic. I’m probably better than average at lie detection. On the other hand I have not practiced that skill well enough to trust my skills to draw strong conclusions.
On the other hand in biology there are many cases where I have true beliefs where the justification isn’t deeper than “the textbook says so”. There no underlying logic that can tell you what a neurotransmitter does in every case. Evolution is quite random. In many cases I could tell you which textbook or university lecture is responsible for me believing a certain fact about biology.
Facts backed by empirical research are better than facts backed by theoretical reasoning like the one in the nose touching example.
I’ve heard a few justifications for the nose-touching example, and the one you provided is new to me.
The nose-touching example could alternatively have an empirical basis rather than a theoretical one. It came to mind because areas like body language, “soft skills” and the like are, in my experience, rife with people who’ve learned a list of such “facts”.
That’s a good start, now take those ideas about imaginary expertise and use them to do some introspection… :)
Anyway, it’s hard to clearly define what an ‘imagined expert’ is. All it means is that the person overestimates their knowledge of a subject, which affects ALL human beings, even bona-fide experts. Expertise in anything other than a very small slice of human knowledge is obviously impossible.
It’s easy to spot lack of knowledge in another person but not so easy to spot it in yourself.
About ‘crackpottery’, I don’t like labelling people as crackpots. Instead it’s much better to talk about crackpot ideas. There is no distinction between ‘crackpot’ and ‘wrong’. All ‘crackpot’ means is that the idea has very low likelihood of being true given our knowledge about the world; this is just the definition of ‘wrong’. It is in human nature to be immediately suspicious of consensus; that’s not the problem (in fact it’s part of the reason why science exists in the first place). The problem is when people try to push alternative agendas by pushing wrong information as correct information. That’s not a problem of imagined expertise; it’s a problem of willful misguiding of others.
I don’t see sixes_and_sevens labeling people as crackpots:
(Emphasis added.)
In other words, it’s not that someone has or lacks the crackpot label. Rather, there is a “crackpot narrative”, a sort of failure mode of reasoning, which people can subscribe to (or repeat to themselves or others) to a greater or lesser extent.
The difference is significant. It’s like the difference between saying “Joe is a biased person” and saying “Joe sure does seem to exhibit fundamental attribution error an awful lot of the time, doesn’t he?”
On the “all arguments are soldiers” metaphorical battlefield, I often find myself in a repetition of a particular fight. One person whom I like, generally trust, and so have mentally marked as an Ally, directs me to arguments advanced by one of their Allies. Before reading the arguments or even fully recognizing the topic, I find myself seeking any reason, any charitable interpretation of the text, to accept the arguments. And in the contrary case, in a discussion with a person whose judgment I generally do not trust, and whom I have therefore marked as an (ideological) Enemy, it often happens that they direct me to arguments advanced by their own Allies. Again before reading the arguments or even fully recognizing the topic, I find myself seeking any reason, any flaw in the presentation of the argument or its application to my discussion, to reject the arguments. In both cases the behavior stems from matters of trust and an unconscious assignment of people to MySide or the OtherSide.
And weirdly enough, I find that that unconscious assignment can be hacked very easily. Consciously deciding that the author is really an Ally (or an Enemy) seems to override the unconscious assignment. So the moment I notice being stuck in Ally-mode or Enemy-mode, it’s possible to switch to the other. I don’t seem to have a neutral mode. YMMV! I’d be interested in hearing whether it works the same way for other people or not.
For best understanding of a topic, I suspect it might help to read an argument twice, once in Ally-mode to find its strengths and once in Enemy-mode to find its weaknesses.
Just wondering if it would make sense to consider everyone a Stupid Ally. That is, a good person who is just really really bad at understanding arguments. So the arguments they forward to you are worth examining, but must be examined carefully.
This is generally how I read self-help books, especially the ones I have to hold my nose for while reading. (Because their logic stinks.) ;-)
Basically, I try to imagine Ignaz Semmelweiss telling me his bullshit theory about handwashing, when I know there is no way that there could be such a powerful poison in corpses, that a tiny invisible amount would be deadly. So I know that he is totally full of bullshit, at the same time that I must take at face value the observation that something different is happening in his clinic. So I try to screen out the theory, and instead look for:
What are they observing (as opposed to opining about the observations)
What concrete actions are they recommending, and
What effect of these actions are they predicting
As this information can be quite useful, once the insane theories are scrubbed off the top.
While it’s easy to laugh at doctors for ignoring Semmelweiss now, there was no scientific theory that could account for his observations until well after his death. A similar phenomenon exists in self-help, where scientists are only now producing research that validates self-help ideas introduced in the 70′s and 80′s… usually under different names, and with different theories. Practice has a way of preceding theory, because useful practices can be found by accident, but good theories take hard work.
So, “Stupid Ally” makes sense in the same way: even an idiot can be right by accident; they’re just unlikely to be right about why they’re right!
I figured this would be broadly of interest to this site:
http://www.pnas.org/content/111/45/16106.abstract
“Chlorovirus ATCV-1 is part of the human oropharyngeal virome and is associated with changes in cognitive functions in humans and mice ”
A metagenomics study of the throat virome of a bunch of people in Baltimore revealed that a full 40% of them were persistently infected with a clade of chloroviruses, very large ~300 kilobase DNA viruses (possibly very distantly related to poxviruses but it’s difficult to tell) which have previously only been known to infect freshwater algae. Upon looking at correlations, they found an extremely significant correlation between infection and a mild degradation in memory and attention tasks. Infecting mice with the virus both caused a decline in memory function as measured by maze-running tasks and, since unlike a human you can crack open a mouse and extract RNA from pieces of its brain, very clear changes in the gene expression of the hippocampus. Not a clue about the mechanism.
This virus had already been noted to be odd a few years ago—a paper from 2011 (http://www.sciencedirect.com/science/article/pii/S1360138511002275) noted that they contained carbon catabolism genes and carbohydrate-processing enzymes that seem to come from animals despite the fact that they were only known to infect algae.
Interestingly, human-and-algae-infecting viruses appear to have been empirically identified in the vaginal secretions of people whose secretions killed algae by scientists in Ukraine several years ago, and this has largely gone unnoticed (possibly justifiably) since it was published (www.bioorganica.org.ua/UBAdenovo/pubs_9_2_11/Stepanova.pdf) in a small journal with poor English translation.
Fulltext: https://pdf.yt/d/dr3uP9XOtT1BPimU / https://dl.dropboxusercontent.com/u/5317066/2014-yolken.pdf / http://libgen.org/scimag7/10.1073/pnas.1418895111.pdf
I didn’t much like it. This thing reeks of data dredging in every step; I don’t see why they controlled for birth place where you’d think that current residence would be much more relevant (Baltimore has rich and poor areas like most big cities; and if nothing else, it’d give you an idea of infection vectors if carriers cluster around the harbor or specific places); I find it odd that their WAIS subtest shows zero (0.0) decrease in the infected group while their weirdo IQ test I’ve never heard of shows a fall; and I’m not sure how convincing I find their mice models* - to what extent does it really mimick human infections with no apparent symptoms? It wouldn’t surprise me if, every time you gave mice a big injection of infectious organisms, their scores fell simply because you made them sick with something, so I’m not sure whether the mice experiment part is testing the right causal hypothesis (it might be testing ‘raging infections decrease cognitive performance’, not ‘this algae, and not other infectious agents, decreases cognitive performance’).
I would not be surprised if this never replicates.
* kudos to them for trying to experimentally test it, though
Good points all—I was hoping you’d show up. It’s odd enough though that I would be quite interested in any attempts at replication. Course, that might be coming from my interest in the evolutionary history and ecological role of a virus that can apparently infect organisms as different as blue-green algae and mammals.
I keep on wondering to what degree our discovery and exploration of the all the microflora and microfauna that lives insides us will overturn “standard” medical theories of how humans work. It’s already generally accepted that the diet’s effect on nutrition is noticeably mediated by the gut biota, though the details are still very very fuzzy, and that’s only a crude start...
This stupid bot has almost 20 000 comment karma on Reddit.
I have seen it in action, and sometimes it may take some time for humans to recognize it is a bot, not a passively aggressive human. Because, well, there are many kinds of humans on internet.
But this made me think—maybe we could use “average reddit karma per comment” or something like this as a measure of Turing test. And perhaps we could make a bot-writing competition, where the participant bots would be released on Reddit, and the winner is the one which collects most karma in 3 months.
Of course the rules would have to be a bit more complex. Some bots are useful by being obvious bots, e.g. the wikipediabot who replies with summaries of wikipedia articles to comments containing links to wikipedia. Competition in making useful bots would also be nice, but I would like to focus on bots that seem like (stupid) humans. Not sure how to evaluate this.
Maybe the competition could have an additional rule, that the authors of the bots are trying to find other bots on Reddit, and if they find them, they can destroy them by writing a phrase that each bot must obey and self-destruct, such as “BOT, DESTROY YOURSELF!”. (That would later become a beautiful meme, I hope.) The total score of the bot is the karma it has accumulated until that moment. Authors would be allowed to launch several different instances of their bot code, e.g. in different subreddits, or initialized using different data, or just with different random values.
Has anyone tried something like this before? What is the reddit policy towards bots?
Related: Stealth Mountain, a twitter bot (now defunct) which would correct tweets containing the expression “sneak peak”.
Both this and the bot you link to rely less on getting machines to cleverly reproduce human behaviour, and more on identifying robotic human behaviour that can be carried out by stupid machines. Since this is probably a winning strategy, I’d recommend making that the focus of such a competition.
Core War in social media? This could get a wee bit out of hand… X-)
The only site-wide rules I’m aware of are ones against abusing the API (ie, your bot shouldn’t be functionally equivalent to a DDoS attack). Other than that, most large subreddits seem to allow bots, but it’s up to the individual moderators.
Are you familiar with the various only automatic rant generators?
I have seen various random text generators on their own web pages, but never actively participating in a forum.
Would supporting open immigration count as part of effective altruism?
Not unless you have an amazing insight / highly effective novel argument to inject into the political debate on the issue. It is an altruistic goal, but pushing for political changes that have near zero chance of actually coming to pass or have overton window shifts in the desired direction fails hard at the “Effective” part. Dont get me wrong, political action can certainly be a very effective way to do good in the world, but this particular issue is so far out of reach of the overton window it kind of isn’t part of the building anymore.
The best argument I’ve ever been able to come up with is that we should underbid the people smuggling industry—Grant anyone who can post a 3000 dollar bond and a return ticket a temporary right of residence and employment as long as their money lasts, and I’ve never been able to sell anyone on it that weren’t already opposed to controls on immigration.
I have also seen people much, much better at politics than me try to persuade people of the immorality of immigration restrictions. I have never seen it work. Anecdote, data, I know, but if this was a good attack angle, I should be able to think of examples of it not just abysmally failing.
If I could figure out a way to make seasteading actually profitable (iron fertilization fishing industry? ) doing that under an appropriate flag of convenience and recruiting labor out of the various hellholes people need to leave seems workable, because it doesn’t require persuasion. However, it also requires what would effectively be company towns to be run to first world standards of living…
It’s not an either-or thing. There are a lot of intermediate states and pushing in the direction of open immigration can be effective even if you don’t expect to get to the abolisment of national borders in the near future. EU exists.
And if we could get our politicians to stop fucking up for a couple of years, expanding the union further would be the most viable angle of attack I can think of. But that requires the european economy to be doing much better than it currently is to be viable, so step one would be to deprogram our political class of their love affair with austerity…
EU expansion doesn’t rely on a successful European economy. Croatia’s accession to the EU took place entirely post-GFC.
The real reason the EU won’t expand significantly is because there’s nowhere for it to expand to. Norway, Switzerland and Iceland don’t want to join. Turkey and the Ukraine aren’t wanted. Some of the remaining Balkan countries will join eventually, but they total only 18 million people, and none are great accession candidates; it’s now recognised that Bulgaria was probably a mistake, and there’s no wish to repeat that. I’d say there’s more chance that the UK leaves the EU than it undertakes a significant expansion in the next, say, 20 years.
Ah, well, I wouldn’t advise you to hold your breath… X-)
The most pressing problem at the moment for that expansion east is that Mr.Putin doesn’t seem to like it. And if you are thinking about expanding south, the voters don’t seem to like that.
The voters are reasoning beings—if we could get the economy created by the last expansion round to work well, they would have more of an appetite for further rounds. Nothing much is viable as long as we are collectively shooting ourselves in one foot, reloading and switching to the other, however. It hurts my head that our political class doesnt seem to be updating on evidence at all. Austerity makes things worse. Try something else. ANYTHING ELSE, Wage hikes in Germany. A monster project to build a low-pressure intercity subway between every city in europe with more than 100000 inhabitants. I’m not picky, as long as it’s less idiotic than “If we remove money from the economy it’s sure to grow!”
Are they now? They persistently elect those politicians who make you despair. And I’m not sure at all that the anti-immigrant moods can be assuaged by throwing money at the voters.
There is no austerity in Western Europe (with the possible exception of Greece). Take a look at the government budgets—did they significantly contract? Take a look at the ECB balance sheet and the current interest rates as well.
Sure. How about you deregulate the economy and let capitalism do what it does best?
From a Keynesian point of view the impression is more important than the reality. There may not be austerity, but people think there is, and they’re saving rather than consuming as a result, and that slows the economy.
Why? If savings come out the other end as investment.
They don’t always. Even if they do, investment doesn’t help if it doesn’t have anywhere to go. Right now the economy is not investment-limited, it’s demand-limited (as you can tell by e.g. low returns on investment).
Why doesn’t the market for capital equalize the discrepancy between supply and demand by adjusting the price (the interest rate)?
a) there’s a persistent market distortion because investment profits are undertaxed
b) savers are often not terribly price-sensitive; a lot of people will “save as much as they can” at any interest rate. Also interest can’t go below nominal zero (many bank accounts are very close to this).
What do you mean? Corporate tax rates, at least in the US, are higher than personal tax rates.
Sure it can.
It does for long-term interest rates. Short-term interest rates are effectively determined by the government.
“And in the long run we’re all dead”? : )
That too, but conventionally long-term interest rates are 10-years and up. In the US the term structure usually goes out to 30 years, in other countries sometimes more.
Heh. That’s an interesting question. Some people certainly believe so.
It’s something that GiveWell’s looked into over the last year or two, and so has Carl Shulman on Reflective Disequilibrium as part of his effective altruism research notes.
“Effective” assumes that you believe you make an impact. Depending on how you plan to support it, effectiveness might vary greatly.
I went to a talk which presented the horrifying fact that immigrants are several times more likely to develop psychosis (varying from 2x to 10x dependant upon group, IIRC), and they theorised that some people have trouble empathising with people of other races, and this literally drives them insane. (Their other theory was that the cause is stimulant psychosis, but there wasn’t evidence to support this as the primary cause)
I’d rather this wasn’t the case. I quite like some level of controlled immigration, especially if they bring curry. But I worry that driving people insane is a fairly severe negative externality.
Here’s a study.
I think the amount of additional schizophrenia is low enough that this isn’t a major issue for immigration.
Guessing that link’s meant to go somewhere else...
Thanks. I’ve found a better link than the one I intended, so I’ve updated the comment text.
Did this study consider the difference between white and non-white immigrants to mostly white Western countries?
Did this study consider the difference between white and non-white immigrants to non-white countries?
Did this study consider the difference between immigrants who (try to) assimilate to local communities, and those who prefer to stay within national communities?
Yes, the effect was very pronounced in Africans, possibly due to the use of the traditional African stimulant khat. But white people do drugs too, so this seems very dubious.
&
No, although maybe other studies did.
For context on the size of the potential benefit, an additional 1% migration rate would increase world GDP by about 1% (i.e. about one trillion dollars). The main question is the rate of migration if barriers are partially lowered, with estimates varying between 1% and 30%. Completely open migration could double world output. Based on Table 2 of Clemens (2011)
I am having strong doubts about this number. The paper cited is long on handwaving and seems to be entirely too fond of expressions like “should make economists’ jaws hit their desks” and “there appear to be trillion-dollar bills on the sidewalk”. In particular, there is the pervasive assumption that people are fungible so transferring a person from a $5,000 GDP/capita economy to a $50,000 GDP/capita economy immediately nets you $45,000 in additional GDP. I don’t think this is true.
This is ridiculous. If this were possible, wouldn’t the industries just relocate to the third world? (obviously, yes, some industries are inherently local, but not all industries that can move have moved)
Um....
:-D
Yes, I know about China, but (a) not all industry that could move has (b) China isn’t really a third world country.
Not any more it isn’t, now that all the industry moved there.
All the industry moved there? And so why hasn’t it now moved to the remaining third-world countries?
Not all of it, but a lot of it. My point is that its economy grew as a result.
Industries are moving out of China to cheaper places like Africa and the Philippines. But the real-world economy is not completely frictionless; it takes time for these effects to occur.
Indeed, it takes time, and for a third world person to reach the productivity of a first world person could take maybe 15 years of first-world education at a bare minimum? So people are not perfectly fungible.
I very much doubt your number. Third world people who go to first world universities do fine, so that’s 3 years, not 15. And for many jobs—even a very intellectual one like mine—I’m skeptical how much difference a degree really makes.
If immigrants didn’t have the skills that were needed to be productive, companies wouldn’t want to hire them. Instead we see companies chafing at the H-1B limits and crying out for more immigration.
But these people are not the average third world people, they are presumably the ones fortunate enough to get a decent education in a country where many people don’t.
I’m certainly not arguing some crazy position like ‘all immigrants are stupid.’ I’m also guessing that these immigrants that companies desperately want to hire are mostly not from third world countries.
Any thoughts about what will happen when there are no longer really cheap places to move to?
Productivity will equalize and then growth will slow. When I’m pessimistic I fear every national economy will look like Japan’s as soon as we catch up.
The robots will take over.
That’s not a joke—industrial automation is proceeding at a rapid clip and the range of jobs for which you want dumb but cheap human labour continues to contract.
Also, the industries that can move and would benefit already did, so the rest are in particular need of people from developed nations.
The paper cited is handwavy and conversational because it isn’t making original claims. It’s providing a survey for non-specialists. The table I mentioned is a summary of six other papers.
Some of the studies assume workers in poorer countries are permanently 1/3rd or 1/5th as productive as native workers, so the estimate is based on something more like a person transferred from a $5,000 GDP/capita economy to a $50,000 GDP/capita economy is able to produce $10-15K in value.
It looks to me as providing evidence for a particular point of view it wishes to promote. I am not sure of its… evenhandedness.
I think that social and economic effects of immigration are a complex subject and going about trillions lying on the sidewalk isn’t particularly helpful.
Are they advocating for abolition of the minimum wage? Can one survive on 1/5th the average salery? Will the combination of inequality and race cause civil unrest?
Sure, a randomly chosen person in a $5,000 GDP/capita economy is unlikely to make $50,000 no matter where you move them, but no-one is proposing to move people at random.
Do you happen to know of a scatter graph for immigration rate vs GDP? It might shed a little light on the matter, though fertility would be a cofounder.
I fail to see the connection between more immigration and improved utility. Notably most of the arguments in this regard tend to assume that people are completely fungible and that third-worlders will magically acquire the characteristics of the average first-wolder the moment they cross the boarder.
I don’t think that’s a necessary implication for wanting to open up immigration. All that’s needed is that new immigrants should do significantly better than they did in their home country, and do some good in the country they’ve moved to.
Notice that your two assumptions nearly contradict each other. After all if the average citizen of the old country was as capable of doing good as the average citizen of the destination, the old country wouldn’t be the kind of place one needs to leave to do significantly better.
Consider the common case of a Latin American man who leaves to become a construction worker in the US.
He’s a basically sensible person who’s willing to work hard, but due to corruption, lack of infrastructure, and probably prejudice in Latin America, he’s extremely poor.
He risks his life to go to America, where he’s still subject to corruption (employers can get away with cheating him of his wages, and there may be payments for his work into Social Security that he will almost certainly never be able to collect) and prejudice, but he’s still better off because he’s hooked into much better infrastructure, probably less corruption, and possibly less prejudice.
Why are these problems so much worse in Latin America? Probably a lot of it has to do with the character of the people there. Thus when he’s in the country he’s likely to do things that incrementally increase the problems he left Latin America to get away from.
If you get irritated by malicious comments like “U.S. people are self-centered, greedy, manipulative, meddlesome, trigger-happy devourers of the world’s resources, entitled policemen of the world’s affairs, and deluded by their overinflated self-importance”, then that should give you a hint of how your odious generalization about Latin Americans is likely to be received.
Many Western societies have seen pretty dramatic productivity-enhancing institutional changes in the last few hundred years that aren’t explicable in terms of changes in genetic makeup. In light of this, your view seems to rely on believing that most currently remaining institutional variation is genetic, whereas this wasn’t the case ~300 years ago. Do you think this is the case?
Hong Kong, Singapore, and South Korea seem to make a pretty strong case for a huge independent effect of institutions.
Who said anything about genetics?
Korea is. China (I assume this is what you mean by Hong Kong and Singapore) is evidence against.
Oops, shouldn’t have assumed you’re talking about genetics :)
Still, if you’re talking about character in a causally neutral sense, it seems that you need to posit character traits that hardly change within a person’s lifetime. Here I admit that the evidence for rapid institutional effects is weaker than the evidence for institutional effects in general.
(Re: Hong Kong, Singapore, no, I do mean those cities specifically. Their economic outcomes differ strikingly from culturally and genetically similar neighbors because of their unique histories.
You seem to assume that everybody in Latin America has the same character, in which case how comes certain people emigrate and other don’t?
You’re leaving out that he left Latin America to get away from those problems, and also that a lot of immigrants want to become real Americans (or whichever country they’re moving to).
But do they understand what caused them.
I’d be more comfortable with an immigration policy that explicitly screened for something like this.
That argument seems to me non-responsive, fallacious, or at least inadequately fleshed out, in three different ways.
Immigrants needn’t be representative of their country of origin, in which case arguments about the average citizen in that country of origin aren’t automatically relevant.
The “average citizen of the old country” needn’t be “as capable of doing good as the average citizen of the destination” for NancyLebovitz’s point to go through; they just need to generate a net gain (however one operationalizes “a net gain”) at the margin.
Something like a fallacy of composition: a priori, “the average citizen of the old country [being] as capable of doing good as the average citizen of the destination” doesn’t automatically imply that “the old country [isn’t] the kind of place one needs to leave to do significantly better”. Given, say, increasing returns to scale & specialization, it’s quite possible that denizens of a tiny country would be outperformed by those of a bigger country, but would nonetheless be quite capable of matching their new neighbours if they moved to that bigger country.
Yes, it was a Baysian not a mathematical argument.
They are unless you have reason to believe the immigrants are above average.
Comparing per-capita GDP with populations suggests we have decreasing returns to scale.
One way to see the problem with Nancy’s argument is to consider the following question: If most people from country X want to move to country Y then wouldn’t it be easier for country Y to simply annex country X? You save on relocation costs and the people are now in country Y.
You do. (For a particular sense of “above average” that’s appropriate here.) The people who choose to leave country A to seek their fortune in country B are going to be (on average) atypical in a bunch of ways.
They will tend to be more optimistic about their prospects in country B and maybe less optimistic about their prospects in country A. It’s not immediately clear what we should expect this to say about them overall, but let’s “change basis” as follows: they will tend to have a higher opinion of how much better they’d do in B than in A, and this seems like it should correlate with actual prospects in B if it’s a healthier country than A.
They will tend to be more proactive, more go-getting. This seems like it should also correlate with productive work.
They will tend to be actually able to get themselves from country A to country B without starving, getting arrested by overzealous police in country A or B or in between, failing to get past border controls, etc. This all seems like it would correlated with effectiveness in getting stuff done. (Both directly, and because their ability to do this will be affected by the resources they have in country A, which for multiple reasons will correlate with their ability to get things done.)
There will probably be differences in their relationships with other people in country A, but I’m not sure which direction the overall effect goes. (Maybe they have looser ties, and that correlates with being less good with people, and that correlates with doing badly; maybe they have good friends and family making them feel well supported and confident, and that correlates with doing well.)
Having got from A to B, they are then going to be strongly motivated to make the trouble and expense worth while, which probably means that whatever their underlying competence they will work harder and more resourcefully in country B than they would have in country A.
So there are lots of reasons to expect people who have emigrated from dysfunctional country A to more-functional country B to be more effective workers in country B than the population average in country A.
(Note: I don’t think this is by any means the dominant reason to agree with Nancy and disagree with Azathoth on this point.)
Something like that was formerly more true. Now I suspect part of the problem is that improvements in transportation have made it “too easy” to immigrate.
I suspect you mean “formerly.”
Thanks, fixed.
Ahem
I don’t quite understand—Bayes is maths.
One could require immigrants to have demonstrable skills. However, if the immigrants are above average then this causes a brain drain for the home country.
This would also stop smaller first world countries like the Japan or the UK becoming overcrowded. OTOH its also completely infeasible in the foreseeable future because people wouldn’t agree to it.
The scatterplot shown here appears to show a strong positive correlation between population and GDP per capita.
[EDITED to add: no, I’m an idiot and misread the plot, which shows a clear correlation between population and total GDP and suggests rather little between population and per capita GDP. Sorry about that. The Gapminder link posted by satt also suggests very little correlation between population and per capita GDP. So the context for my (unchanged) argument below is not “Increasing returns to scale are just one factor; here are a bunch more” but “Increased returns to scale are probably negligible; here are a bunch of things that aren’t”.]
In any case, “increasing returns to scale” were just one example (and I think not the best) of how someone might be more productive on moving from (smaller, poorer, more corrupt, less developed) country A to (larger, richer, less corrupt, more developed) country B. Here, let me list some other specific things that might make someone more productive if they move from (say) Somalia to (say) France.
Better food and healthcare. Our migrant will likely be healthier in country B, and people do more and better work when healthier.
Easier learning. Our migrant may arrive in country B with few really valuable skills, but will find more opportunities than in country A to learn new things.
Better infrastructure. Perhaps our migrant is working on making things; country B has better roads, railways, airports, etc., for shipping the products around for sale. Perhaps s/he is (after taking advantage of those educational opportunities) working on computer software; country B has reliable electricity, internet that doesn’t suck, places to buy computer hardware, etc.
Richer customers. Perhaps our migrant is making food or cleaning houses. People in country B will pay a lot more for this, because they are richer and their time is worth more to them. So, at least as measured in GDP, the same work is more productive-in-dollars in country B than in country A. (Is this a real gain rather than an artefact of imperfect quantification? Maybe. If people in country B are richer and their time is worth more because they are actually doing more valuable things then any given saving in their time is helping the world more.)
Less corruption. Many poor dysfunctional countries have a lot of corruption. This imposes a sort of friction on otherwise-productive activities—one has to spend time and/or money bribing and sweet-talking corrupt officials, and it could have been used for something else. In country B this happens much less.
And what caused these differences between these two countries? (Hint: it’s not magical corruption ray located in Mogadishu.) And how will these traits change as more people move from Somalia to France?
It could be any number of things. Including the one I take it you’re looking for, namely some genetic inferiority on the part of the people in country A. But even if that were the entire cause it could still easily be the case that when someone moves from A to B their productivity (especially if expressed in monetary terms) increases dramatically.
I’m actually not quite sure what point you’re arguing now. A few comments back, though, your claim was that Nancy was (nearly) contradicting herself by expecting immigrants to (1) be productive in their new country even though (2) their old country is the kind of place where it’s really hard to be productive, on the grounds that for #2 to be true the people in the old country must be unproductive people.
It seems to me that for this argument to work you’d need counters to the following points (which have been made and which you haven’t, as it seems to me, given any good counterargument to so far):
There are lots of other ways in which the old country could make productivity harder than the new—e.g., the ones I mention above.
Let me reiterate that these apply even if the old country’s productivity is entirely a matter of permanent, unfixable genetic deficiencies in its people. Suppose the people of country A are substantially stupider and lazier than those of country B; this will lead to all kinds of structural problems in country A; but in country B it may well be that even someone substantially stupider and lazier than the average can still be productive. (Indeed I’m pretty sure many such people are.)
If the differences between A and B do indeed all arise in this way (which, incidentally, I think there are good reasons to think is far from the truth) then yes, if the scale of migration from A to B is large enough then it could make things worse rather than better overall. Given that the empirical evidence I’m aware of strongly suggests that migration to successful countries tends to make them better off, I think the onus is on you if you want to make the case that this actually happens at any credible level of migration.
The people who move from country A to country B may be atypical of the people of country A, in ways that make them more likely overall to be productive in country B.
Your only response to this has been a handwavy dismissal, to the effect that that might have been true once but now immigration is too easy so it isn’t any more. How about some evidence?
Not necessarily, my argument goes through even if it’s memetic.
How about some yourself. Note simply saying that something may happen is not a reason to ignore the prior that it won’t. I responded to your only argument about the prior. Also, look at the way the immigrants are in fact behaving, I believe it involves lots of riots and creating neighborhoods that the police are afraid to go into.
The cost of annexing can be very high. War is hard on people, and we don’t have methods of smoothly and sensibly rearranging national borders.
Also, at that point, country Y has inherited all of country X’s problems.
As far as I can tell, you don’t get everyone (or a very large proportion) of people wanting to leave a country unless there’s a lot of violence or a natural disaster. If the problem is poverty, people would rather have some members of their families emigrate to work, and send money back.
I think “annexing country X” might show some problems with utilitarian thinking—a tendency to abstract away important amounts of detail and the costs of getting from point A to point B. This doesn’t mean utilitarian thinking is always wrong, but these are problems to watch out for.
That’s my point. Having many people from country X emigrate to country Y causes country X to acquire country Y’s problems.
Immigration doesn’t lead to importing country X’s government.
True, but it may very well lead to importing the reason country X has the government it does.
That’s exactly what the FRG did with the GDR.
That was a special case since there the difference was (originally) entirely due to the Soviet occupation. And even then it wasn’t particularly effective, i.e., today the former GDR is doing much worse then the former FRG and (I think) comparably to the neighboring former Soviet satellites that weren’t annexed.
That is a bit better, but even as a Bayesian argument it quietly rests on empirical priors which I find odd (as we’re about to see).
In fact I do. gjm has listed a priori reasons to expect this. More empirically, I already know that epidemiologists talk about a “healthy immigrant effect”, which suggests that immigrants are selected (or indeed self-select) to be healthier & wealthier than the average for their home country. I’ve also seen people bringing up the selectedness of immigrants in arguments about race & IQ, to rebut observations that Third World immigrants tend to do better in their new homes than the estimated mean IQs of their home countries would suggest.
It does, but it is a very weak suggestion. The correlation is not that great to start with, and doesn’t account for other factors, like international differences in the working-age proportion of the population. Economists have tried to take a more systematic approach based on more involved regressions and/or fitting production functions, but the results here seem mixed, varying across industries and the level of aggregation.
(My own prior that increasing returns to scale occur in manufacturing — the class of industries, AFAIK, most important to a country’s development — seems broadly consistent with the available evidence. I may as well add that my point 3 invokes “increasing returns to scale & specialization” simply as an example, and the basic conclusion that there’s “[s]omething like a fallacy of composition” going on stands even if returns to scale are everywhere decreasing.)
I have little to add to what NancyLebovitz & skeptical_lurker have already said.
If I didn’t think I’d have higher utility in a different country than in mine, I wouldn’t want to move to that country in the first place. Granted, my moving there could affect other people’s utility, but it’s not obvious a priori what the sign of the net effect would be. EDIT: Most of the obvious externalities are pecuniary, so their effects on total utility cancel out when considering both sides of each transaction (assuming they have equal marginal utility for money).
I was reading the thread about Neoreaction and remembered this old LW post from five years ago:
When I saw the question posted in the discussion, I thought it had a potential to be a good discussion topic. After all, I reasoned, there are a many thoughtful people on LessWrong who are interested in politics, history, political philosophy. There are a lot of insights to be gained from discussing interesting and difficult questions about society. And there are quite a few insightful neoreactionaries here, eg. Konkvistador and others (some of whom sadly no longer actively participate on LW). And some neoreactionary ideas are interesting and worth a look.
Despite all this, it seems that currently so many of subthreads of that thread basically turned into an unproductive flamewar. Why? Well, politics is the mind-killer, of course. What should have I expected. Nevertheless, I think it could have been avoided. I am not totally against political discussion in some threads. In fact, even many comments in that thread are good. Well, taken individually many comments are quite reasonable, for example, some of them explain a certain position and while you may agree or disagree with a stated position, you can’t say anything bad about the comment itself. However, when aggregated they do not form a productive atmosphere of the thread. While many comments are reasonable, it is suspiciously easy to sort most of them into a two group—pro- and anti- neoreaction with little middle ground who could act as (sort of) judges that could help evaluate the claims of both sides. There is suspiciously little belief updating (even on small issues) going on (maybe it is different among lurkers), which is probably a very important measure of whether discussion was actually productive (I do not claim that all LessWrong discussions are productive. A lot of them aren’t). Many people aren’t arguing in good faith. Some of them even post links to this discussion in other forums as if it is representative of LessWrong as a whole.
I am not calling for censorship or deletion of certain comments. Nor I want discussing controversial issues being prohibited. I am calling for a moment of reflection about mind-killing. For a moment of consideration about whether you yourself aren’t mind-killed, whether you yourself are no longer updating your beliefs honestly, whether you are no longer arguing in a good faith (even if the other side lowered its standards first). I don’t know what ritual could be devised to reinforce this point. Maybe it is true that there is only three stable equilibria points (Vladimir_M, alas, no longer comments on LessWrong). Is it possible to devise something, I don’t know, maybe a ritual or social norm or yet something else that would help keep things in an unstable point of both having the broad scope of questions and the high quality discussion? Or is any such attempt doomed to crumble due to the influence of discussion standards of the outside world?
In addition to that, I think that the question itself was poorly worded. It was way too broad. The questions that are likely to be polarizing would benefit from being much more narrow and much more specific. Maybe this way everyone would be talking about the same thing, as it would be much harder to try to steer the discussion into things that you like to talk about. Maybe this way everyone would have clearer and more concrete idea about what everyone else is talking about, making it easier to reason about the whole situation, easier to weigh the evidence in favour of one or the other position.
I agree it would be good to find a better method to debate politics. Maybe we should have a meta-rule that anyone who starts a political debate must specify rules how the topic should be debated. (So now the burden is on the people who want to debate politics here.)
It seems to me that in political topics most of updating happens between the conversation. It’s not like you say something and the other person is “oh, you totally convinced me, I am changing my mind now”. Instead, you say something, the other person looks at you very suspiciously and walks away. Later they keep thinking about it, maybe google some data, maybe talk with other people, and the next time you meet them, their position is different from the last time.
For example, I have updated, from mildly pro-NR to anti-NR. I admit they have a few good points. But this is generally my experience with political movements: they are often very good at pointing out the obvious flaws of their competitors; the problem is that their own case is usually not much better, only different. I appreciate the few insights, they made me update, and I still keep thinking about some stuff. I just didn’t come to the same conclusion; I separated the stuff that makes sense to me from the stuff that doesn’t. Just like I try to draw good ideas e.g. from religion, without becoming religious. Instead of buying the whole package, I take a few bricks and add them to my model of the world. There are a few bricks in my model now that an outside observer could call “neoreactionary”, although that would probably depend on the exact words I would use to describe them (because they are not unique for NR). The other bricks I have judged separately, and I was unimpressed. That’s where I am now.
There is also this irritating fact that NRs keep associating themselves with LW. I consider that a huge dishonesty and in a way an attack on this community. If people are impressed by LW, this can make them more open towards NR. If people are disgusted by NR, this can make them dislike LW by association. They gain, we lose. It never goes the other way round; no one is going to debate overcoming their cognitive biases just because they fell in love with NR. To put it bluntly, we are used as a recruitment tool for some guy’s cult, and all his shit falls on our heads. Why should we tolerate that? (This, especially #1 should be a required reading for every nerd.) That alone makes me completely unwilling to debate with them, because such debates are then used as further evidence that “LW supports NR”. (As an analogy, imagine how much would you want to have a polite debate with a politician you dislike, if you know that the reason he debates with you is that he can take a photo of you two having a conversation, put it on his personal webpage, and claim that you are one of his supporters, to impress people who know you.) I refuse to ignore this context, because I am strongly convinced that NRs are fully aware of what they are doing here.
So even if we try having rational debates about politics, I would prefer to try them on some other political topics.
I think this is a great suggestion, since it allows different standards for different types of political discussion, as well as giving us a chance to actually observe which set of rules leads to most productive discussion.
Well, I think this is probably true in my experience. On the other hand, since this is an internet forum, no one is forcing them to post their answers immediately. Maybe for most people it takes months to change their position on significant political belief even if they have a lot of evidence that contradict that belief, thus we do not expect that given a person he/she would change their beliefs after a conversation. However, thinking at the margin, there must have been people who were on the fence. There must have been people who quickly jump from one set of beliefs to another one whenever someone posts an interesting essay. Maybe for them a week would have been enough to update? And since this was not a real time conversation, they could post about their update after a week and it was their pride that prevented them from doing so? However, less people seemed be on the fence than I expected, “the distribution of opinions about neoreaction” seemed bimodal. However, now that I write this, I realize that such people would have been less motivated to write their beliefs in the first place, thus they were underrepresented in the total volume of posts in that thread. Thus, it is possible that the impression of bimodality is partially an artefact of that.
It is good to hear that you have found something in that thread that you thought was worth updating on. I also agree that neoreaction is better at finding flaws of other movements (for example, I think that some trends they describe as dangerous are actually dangerous) and providing some intellectual tools for thinking about the world that can be added into one’s toolbox (I am not a neoreactionary, whether those tools accurately describe the world is a different question, to me it seems that they are at least worth thinking about, can they shed some light on things that other intellectual tools neglect?) than providing an alternative course of action, an alternative model of what kind of society is good or an alternative movement that would be worth following (it seems to me that neoreaction is more like, well, reaction to progressivism (in neoreactionary sense of the word) rather than coherent set of goals in itself (it seems to me that the groups that compose neoreaction are as different from each other as either of them is from progressivism)). So, basically, I think my position towards neoreaction is somewhat similar to yours.
This is where my intuition differs from yours. Maybe this is because I have never been to a LW meetup, nor I have ever met another person who reads LW in real life. In addition, I have never met a single neoreactionary in real life. Or maybe I simply don’t know about them, I don’t think I have ever met a single SJW in real life either. I understand that LessWrong consists of real people, but when I think about LessWrong, the mental image that comes to my mind is that of a place, abstract entity and not a community of people. Although I obviously understand that without all these people this place would not exist, the mental image of LessWrong as “a place (maybe cloudlike, maybe vaguely defined) where LW style discussion about LW topics happens (style and topics are most important part of what defines LW to me)” feels more real to me than the mental image of community of people. I do not know much about LW posters beyond what they post here or on other blogs. For example, when I first started reading LessWrong, for quite a long time I thought that Yvain and Gwern were women. Why did I think this? I don’t remember. What I’m trying to say is that I guess that the difference between our intuitions may come from the difference between how we think about these two layers (place, style of discussion, topics vs community of real people). It may be a bias on my part (i.e. I don’t know what kind of thinking leads to an optimal outcome, I am not sure how exactly such optimal outcome would look like) that I neglect the community building aspect of LessWrong, I am not sure. I haven’t disentangled my thoughts about these things in my mind yet, they are very messy. This post is partially an attempt to write down my intuitions about this (as you can see, it is not a very coherent argument), maybe it will help myself to clarify some things.
In addition to that, while an individual identity is relatively well defined (“I am me”), identity of someone who belongs (or does not belong) to a certain group is much less clearly defined and whether someone actively feels belonging to a certain group seems to depend on a situation.
What I am trying to say is that when I see neoreactionaries commenting on LessWrong, I do not perceive them as “them” if they talk in a manner that is close enough to LessWrong style about the topics that are LW topics. In this situation, I do not perceive LWers and LW neoreactionaries as distinct groups in a way that a statement about the attack on the community would make sense. In fact, in this situation, only a small part of my attention is dedicated to identity related thoughts. The situation is different when, e.g. I read someone’s comments (usually outside of LessWrong) attacking LessWrong. In this case the part of my attention that is dedicated to identity related things is much larger. In such situations, I do think of myself as someone who regularly reads LessWrong and finds it a great place with a lot of interesting people who write about their insights, when someone attacks it, my emotions create an urge to defend LessWrong. In such situations much larger part of my attention is dedicated to this, and I do start thinking in terms of who belongs to what group. But unless it is neoreactionaries who are attacking LessWrong, I usually still do not feel (I am just describing what I feel in such situations) that LW neoreactionaries (not neoreactionaries in general) are distinct group. Thus, in my case, it seems that it is conflicts and disagreements that create a sense of identity (even more than vice versa), since, as I have said, I have never participated in an offline LW community. (to be continued in the next comment)
I suspect this is the polarizing effect of politics, not something specific for LW nor specific for neoreaction. We are talking about labels, not ideas. I may agree with half of ideas of some movement, and disagree with other half of ideas, but I usually have a clear opinion about whether I want to identify with a label or not.
My mental image for LW community is more or less “people who have read the Sequences, and in general agree with them”. Yes, I am aware that in recent years many people ignore this stuff, to the degree where mentioning the Sequences is a minor faux pas. (And for a while it was a major faux pas, and some people loudly insisted that telling someone to read the Sequences is a lesswrongeese for “fuck you”. Not sure how much of that attitude actually came from the “Rational”Wiki.) That, in my opinion, is a bad thing, and it sometimes leads to reinventing the wheel in the debates. To put it shortly, it seems to me we have lost the ability to build new things, and became an online debate club. Still a high quality online debate club. Just not what I hoped for at the beginning.
LessWrong was built upon some ideas, and one of them was that “politics is the mindkiller” and that we strive to become more rational, instead of being merely clever arguers. At this moment, neoreactionaries are the group most visibly violating this rule. They strongly contribute to the destruction of the walled garden. Debating them over and over again is privileging a hypothesis; why not choose any other fringe political belief instead, or try creating a new one from scratch, or whatever?
Politics is an advanced topic for a rationalist. Before going there, one should make sure they are able to handle the easier situations first. Also, there should be some kind of feedback, some way of warning people “you have strayed from the path”. Otherwise we will only have clever arguers competing using their verbal skills. When a rationalist sympathetic to neoreaction reads the SSC neoreaction anti-faq, they should be deeply shocked and start questioning their own sanity. They should realize how much they have failed the art of rationality by not realizing most of that on their own. They should update about their own ability to form epistemically correct political opinions. Instead of inventing clever rationalizations for the already written bottom line.
In my opinion, Yvain is the most qualified person for the task of debating politics rationally, and the only obvious improvement would be to somehow find dozen different Yvains coming from different cultural backgrounds, and let them debate with each other. But one doesn’t get there by writing their bottom line first.
Did LW as a group ever have this ability? Going by the archives it seems that there were a small number (less than 10) of posters on LW who could do this. Now that they’re no longer posting regularly, new things are no longer produced here.
A reasonable case could be made that this is how NRx came to be.
If this is where NRx came from, then I am strongly reminded of the story of the dog that evolved into a bacterium. An alternative LW-like community that evolved into an aggresive political movement? Either everyone involved was an advanced hyper-genius or something went terribly wrong somewhere along the way. That’s not to say that something valuable did not result, but “mission drift” would be a very mild phrase.
As far as I can see it evolved into mostly smart people writing dense texts about political philosophy. That’s a bit different :-)
That would describe quite a few political movements, actually—it’s hardly exclusive to NRx.
Nope, political movements and political philosophy belong to different categories.
Some political movements evolve out of political philosophy texts, but not all political philosophy texts evolve into political movements.
I think that at this point it would be fair to say that a movement has developed out of NRx political philosophy.
Show me that movement in actual politics. Is any NRx-er running for office? Do they have an influential PAC? A think tank in Washington, some lobbyists, maybe?
Nah, man. Once you get to that level of politics, you’re already pozzed.
Oh, I think we’re using the phrase “political movement” in different senses. I meant something more like “group of people who define themselves as a group in terms of a relatively stable platform of shared political beliefs, which are sufficiently different from the political beliefs of any other group or movement”. Other examples might be libertarianism, anarcho-primitivism, internet social justice, etc.
I guess this is a non-standard usage, so I’m open to recommendations for a better term.
Yep, looks like we are using different terminology. The distinction between political philosophy and political movement that I drew is precisely the difference between staying in the ideas/information/talking/discussing realm and moving out into the realm of real-world power and power relationships. What matches your definition I’d probably call a line of political thought.
Mencius Moldbug is a political philosopher. Tea Party is a political movement.
Sentiments like this are, in my opinion, a large part of why “politics is the mind-killer.” I am no neoreactionary, but I thought the SSC neoreaction anti-faq was extremely weak. You obviously thought it was extremely strong. We have parsed the same arguments and the same data, yet come out with diametrically opposed conclusions. That’s not how it’s supposed to work. And this is far from a unique occurrence. I frequently find the same article or post being held up as brilliant by people on one side of the political spectrum, and dishonest or idiotic by people on the other side.
It is not merely that people don’t agree on what’s correct, we don’t even agree on what a successful argument looks like.
Well, sometimes that’s exactly how it’s supposed to work.
For example, if you have high confidence in additional information which contradicts the premises of the document in whole or in part, and VB is not confident in that information, then we’d expect you to judge the document less compelling than VB. And if you wished to make a compelling argument that you were justified in that judgment, you could lay out the relevant information.
Or if you’ve performed a more insightful analysis of the document than VB has, such that you’ve identified rhetorical sleight-of-hand in the document that tricks VB into accepting certain lines of reasoning as sound when they actually aren’t, or as supporting certain conclusions when they actually don’t, or something of that nature, here again we’d expect you to judge the document less compelling than VB does, and you could lay out the fallacious reasoning step-by-step if you wished to make a compelling argument that you were justified in that judgment.
Do you believe either of those are the case?
I don’t want to focus on the anti-neoreactionary FAQ, because I don’t want to get this dragged into a debate about neoreaction. In particular I simply don’t know how Viliam_Bur parsed the document, what additional information one of us is privy to that the other is not. My point is that this is a general issue in politics, where one group of people finds a piece compelling, and another group finds a piece terrible.
And note too that this isn’t experienced as something emotional or personal, but rather as a general argument for the truth. In this case, VB thinks neo-reactionaries should be “deeply shocked and start questioning their own sanity.” In other words, he thinks this is basically a settled argument, and implies that people who persist in their neoreaction are basically irrational, crazy or something along those lines. Again, this is a general issue in politics. People generally believe (or at least, talk like they believe) that people who disagree with them politically are clinging to refuted beliefs in the face of overwhelming evidence. I don’t just think this is due to epistemic closure, although that is part of it. I think it’s partly an emotional and cultural thing, where we are moved for pre-rational reasons but our minds represent this to us as truth.
I am certainly not saying I am immune from this, but I don’t have the third-party view on myself. I am not saying I am right and Viliam_Bur is wrong on the case in point. But I do wonder how many neoreactionaries have been deconverted by that FAQ. I suspect the number is very low...
To the extent that you’re making a general point—which, if I’ve understood you correctly, is that human intuitions of truth are significantly influenced by emotional and cultural factors, including political (and more broadly tribal) affiliations—I agree with your general point.
And if I’ve understood you correctly, despite the fact that most of your specific claims in this thread are about a specific ideology and a specific document, you don’t actually want to discuss those things. So I won’t.
I’m happy to discuss specifics, just not about the neo reactionary FAQ. I agree with VB that LW has an unhealthy tendency that every discussion becomes about neo reaction, and I don’t like it.
Instead, how about this article. Jim Edwards is a bright guy, and he clearly intended to persuade with that post. And indeed he has plenty of commenters who think he was making a valuable point. Yet I am at a loss to say what it is. Here he is, claiming to have a graph showing that government spending affects economic growth, yet all that graph shows is changes in government spending. It doesn’t show a correlation, it doesn’t suggest causation, it doesn’t do anything of the sort. Yet some poeople find this persuasive.
When someone says they like dance music (for example), I feel like I’m missing out; they get joy out of something I hate, which in some ways makes them better than me, but fundamentally de gusts us non set disputandum. The older I get, the more I feel like that’s how all persuasion works.
Yup, those charts puzzle me, too (based on about five seconds of analysis, admittedly, but I have a strong preexisting belief that there are many examples of such silliness on the Internet, so I’m strongly inclined to agree that this particular chart is yet another example… which is of course yet another example of the kind of judgment-based-on-non-analytical factors we’re discussing).
How confident are you that this is how all persuasion works?
I don’t know how general this is, but I do think it’s an important factor that I don’t see discussed.
Another point is peer effects. I remember at school my physics teacher used to use proof by intimidation where he would attempt to browbeat and ridicule students into agreeing with him on some subtly incorrect argument. And he wouldn’t just get agreement because he scared people, the force of his personality and the desire to not look foolish would genuinely convince them. And then he’d get cross for real, saying no, you need to stand up for yourself, think through the maths. But if you can’t fully think through the soundness of the arguments, if you are groping around both on the correct and the incorrect answer, then you will be swayed by these social effects. I think a lot of persuasion works like that, but on a more subtle and long-term level.
Yes, I agree.
That’s kinda a general issue in humans and usually goes by the name of Confirmation Bias.
For example, debates about religion or, say, global warming work in exactly the same way.
But I don’t think it’s just confirmation bias. People do get won over by arguments. People do change their minds, convert, etc. And often after changing their mind they become just as passionate for their new cause as they ever were for the old. But what is persuasive and what is logical sometimes seem disjoint to different people.
You are right that these things afflict some areas more than others. Politics and religion are notoriously bad. And I do think a large part of it is that people simply have very different standards for what a successful argument looks like, and that this is almost an aesthetic.
Sure, confirmation bias is a force but it’s not an insurmountable force. It only makes changing one’s beliefs difficult, but not impossible.
I agree and I don’t find this surprising. People are different and that’s fine.
Take the classic “Won’t somebody please think of the children!” argument. I, for example, find it deeply suspect to the extent that it works as an anti-argument for me. But not an inconsiderate number of people can be convinced by this (and, in general, by emotional-appeal strategies).
I guess what kind of people are convinced by what kind of arguments would be an interesting area to research.
This is an interesting question that seems empirically testable—we could ask those people and make a poll. Although there is a difference between “believing that NRs are probably right about most things” and “self-identifying as NR”. I would guess there were many people impressed (but not yet completely convinced) by NR without accepting the label (yet?), who were less impressed after reading the FAQ. So the losses among potential NRs were probably much higher than among already fully convinced NRs.
That’s a warning sign, not a barbed-wire fence patrolled by guards with orders to shoot to kill.
Neoreaction is an interesting line of thought offering unusual—and so valuable—insights. If you don’t want to talk about NRx, well, don’t. If you want to talk about different political beliefs, well, do.
What is “the path”? LW is a diverse community and that’s one of its strengths.
You did mention mindkill, didn’t you? I recommend a look in the mirror. In particular, you seem to be confusing rationality with a particular set of political values.
Political opinions are expressions of values. Values are not epistemically correct or wrong—that’s a category error.
Talking about “neoreaction” (or any other political group) already is a package-deal fallacy. NRs have a set of beliefs. Each of those beliefs individually can be true or false (or disconnected from evidence). These beliefs should be debated individually. It is quite possible that within the set, some beliefs will be true, some will be false, and some will be undefined. Then we can accept the true beliefs, and reject the false beliefs. There is no need to use the word “neoreaction” anywhere in that process.
So, instead of having threads about neoreaction, we (assuming we are going to debate politics) should have threads about each individual belief (only one such thread at a time). Then we should provide evidence for the belief or against the belief. Then we should judge the evidence, and come to a conclusion, unconstrained by identity labels.
The fact that we are not already doing it this way, is for me an evidence on the meta level that we are not ready for having political debates.
Debating beliefs separately, understanding the conjuction fallacy, providing evidence, avoiding labels, tabooing words… this is all rationality 101 stuff. This is “the path” we have already strayed from. If we collectively fail at rationality 101, I don’t trust our ability to debate more complex things.
Value is “I don’t want children to starve”. Political opinion is “we should increase the minimal wage (so the children will not starve)”. There is more than the value; there is also the model of the world saying that “increasing minimal wage will reduce the number of starving children (without significant conflict with other values)”. Other person may share the value, but reject the model. They may instead have a model that “increasing minimal wages increases unemployment, and thus increases the number of starving children”, and therefore have a political opinion “we should remove minimal wage (so the children will not starve)”. Same value, different models, different political opinions.
It seems to me that people usually differ more in their models than in their values. There are probably few people who really want to optimize the world to increase the number of starving children, but there are many people with political opinions contradicting each other. (Believing too quickly that our political opponents have different values is also covered in the Sequences.)
I don’t think it’s quite that simple.
You are arguing for atomicity of beliefs as well as their independence—you are saying they can (and should) stand and fall on their own. I think the situation is more complicated—the beliefs form a network and accepting or rejecting a particular node sends ripples through the whole network.
Beliefs can support and reinforce each other, they can depend on one another. Some foundational beliefs are so important to the whole network that rejecting them collapses the whole thing. Consider e.g. Christianity—a particular network of beliefs. Some can stand or fall on their own—the proliferation of varieties of Christianity attests to that—but some beliefs support large sub-networks and if you tear them down, the rest falls, too. At the root, if you reject the belief in God, debating, for example, the existence of purgatory is silly.
The package-deal fallacy exists and is real, but excessive reductionism is a fallacy, too, and just as real.
Oh, I don’t trust our ability to debate complex things. But debate them we must, because the alternative is much worse. That ability is not a binary flag, by the way.
True, and these should be separated to the extent possible.
I don’t know about that—I’d like to see more evidence. One of the problems is that people may seem to have the same values at the level of costless declarations (everyone is for motherhood and apple pie), but once the same people are forced to make costly trade-offs between things important to them, the real values come out and I am not sure that they would be as similar as they looked before.
I wish I could give this more than one upvote.
Maybe. It seems to me that there could be two systems of political ideas—call them A and B—both of which are pretty credible when taken as wholes, but for which if you take any single proposition from one and examine it in the context of the other, it’s obviously wrong.
(The same thing happens with scientific theories. Key words: “Quine-Duhem thesis”.)
On the other hand, it does also happen that basically-unrelated ideas get bundled together as part of a package deal, and in that case we probably do generally want to try to separate them. So I’m not sure what the best way is to make the tradeoff between splitting and lumping.
(cont.) I guess it is likely true that more people find about neoreaction on LessWrong than vice versa. However, it is not obvious to me that hardly anyone would even join LessWrong to discuss LW topics after being exposed to neoreaction first. I mean, MoreRight recommends its readers to read LessWrong and SlateStarCodex as well. Xenosystems has LW, SSC and OvercomingBias on its blogroll. Radish Magazine’s list of the people they admire includes Eliezer Yudkowsky. Obviously some of those people were LWers themselves, some might post links to LessWrong because they try to make their place more comfortable to LWers who might wander there. But still, I hope that at least some neoreactionaries would come here with an interest in LW topics (cognitive biases, future of humanity and artificial intelligence). I guess that it is probably true that neoreaction gains more members from LW than vice versa. This is the community layer. But there is also the intellectual toolbox layer. And I think that if LessWrong started discussing politics, having a few neoreactionaries (not just any neoreactionaries, but those who think in LW terms, are able to notice cognitive biases) here would probably be beneficial. And I guess that if we are to overcome biases we will have to deal with politics. You see, I fear that by paying attention only to cognitive biases that are easy to test in the lab we are like a proverbial drunk man searching for his keys under a lamp-post. For example, field of cognitive biases researches what happens inside a person’s head, but certain things from political science and economics such as median voter’s theorem, Duverger’s law and Motte-and-Bailey effect (when instead of happening inside a person’s head, it happens to a movement—when different people from the same movement occupy motte and bailey (I think that individual and group motte-and-bailey’s are quite distinct)), seems to be analogous enough so as to be thought of as yet another kind of biases that prevent from optimal decision making at a group level. And if we were to start discussion about things like these, it would be hard to avoid using political examples altogether. By the way, the idea that the list of biases LessWrong usually talks about is not exhaustive enough has already been discussed here.
Yes, definitely. I think that the political topics (especially at the beginning) would have to be much more specific and less related to the questions of identity.
This could just as easily be described, with the opposite connotation, as the movement containing some weakmans*, which makes me think that we need a better way of talking about this phenomenon. ‘Palatability spread’ or ‘presentability spread’? But that isn’t quite right. A hybrid term like ‘mottemans’ and ‘baileymans’ would be the worst thing ever. Perhaps we need a new metaphor, such as the movement being a large object where some parts are closer to you, and some parts are further away, and they all have some unifying qualities, and it is usually more productive to argue against the part that is closer to you rather than the part that is far away, even though focusing on the part that is far away makes it easy to other the whole edifice (weakmanning); and motte-and-baileying is ascribing to the further-away part of your own movement but pretending that you are part of the closer part.
*in the technical sense; their positions may be plenty strong but they are less palatable
Edit: Whoops no one will see this because it’s in an old open thread. Oh well.
What I had in mind was a situation when “a person from outside” talks to a person who “occupies a bailey of the movement” (for the sake of simplicity let’s call them “a movement”, although it doesn’t have to be a movement in a traditional sense). If the former notices that the position of the latter one is weakly supported, then the latter appeals not to the motte position itself, but to the existence of high status people who occupy motte position, e.g. “our movement has a lot of academic researchers on our side” or something along those lines, even though the position of the said person doesn’t necessarily resemble that of the “motte people” beyond a few aspects, therefore “a person from outside” should not criticize their movement. In other words, a criticism against a particular position is interpreted to be a criticism against the whole movement and “motte people”, thus they invoke “a strongman” do deflect the criticism from themselves.
I think you made a very good point. From the inside, if an outsider criticizes a certain position of the movement, it looks as if they attacked a weakman of the movement and since it feels like they attacked a movement itself, an insider of the movement feels that they should present a stronger case for the movement, because allowing an outsider to debate weakmen without having to debate stronger positions could give the said outsider and other observers an impression that these weakmen was what the movement was all about. However, from the said outsider’s perspective it looks like they criticized a particular position of a movement, but then (due to solidarity or something similar) the movement’s strongmen were fielded against them, and from the outsider’s perspective it does look like that the movement pulled a move that looks very similar to a motte-and-bailey.
I think that replying to old comments should be encouraged. Because otherwise if everyone feels that they should reply as quickly as possible (or otherwise not reply at all), they will not think their positions through and post them in a hurry.
Um, this is a horrible idea. The problem is people will make rules that amount to “you’re only allowed to debate this topic if you agree with me”.
One aspect of neoreactionary thought is that it relies on historical narratives instead of focusing on specific claims that could be true or false in a way that can be determined by evidence.
To quote Moldbug:
Given such an idea of how reasoning works, it’s not clear that there an easy solution that allows for agreeing on a social norm to discuss politics.
It isn’t clear to me that this sort of thought should be called “reasoning”, a term which is commonly used for dealing with propositions that do have truth-values, at all.
It seems to me to be more in the vein of “poetry” or “poetry appreciation”.
I don’t think that’s entirely fair to Moldbug. Illustrating patterns and using the human ability for pattern matching does have it’s place in knowledge generation. It’s more than just poetry appreciation.
After reading the quote I thought that he was trying to make an analogy between finding a historical narrative from historical facts and drawing a curve that has the best fit to a given series of data points. Indeed, saying that such curve is “true” or “false” does not make a lot sense, since just because a point lies outside the graph of a function does not mean that this function cannot be a curve of best fit—one cannot decide that from a small number of data points, one needs to measure (in)accuracy of the model over the whole domain. Such analogy would lead to interesting follow-up questions, e.g. how exactly does one measure inaccuracy of a historical narrative?
However, after reading Moldbug’s post I see that he does not try to make such analogy, instead he tries to appeal to intuitive thinking. I think this is not a good argument, since intuition is the ability to acquire knowledge without inference or the use of reason, therefore saying that you used your intuition to arrive at a certain conclusion is basically saying that you used “something else” (similarly to how you cannot build stuff out of nonwood) - this category does not seem specific enough to be a good explanation. Humans are able to find a lot of patterns, some of which are not meaningful. It is an interesting problem how to recognize which patterns are meaningful and which aren’t. But this applies to the whole field of history, not just Moldbug’s ideas.
I don’t see how it does this any more than any other political philosophy.
It’s not true for someone who does get his beliefs by thinking about issues individually. Whether or not you call such a person having a political philosophy is another matter.
If you could have perfect control of your own mind, what would you do with it?
(I realize the question is a bit vague. Please try and answer anyway.)
The same stuff I normally do, except with less Akrasia and procrastinating, and more rapid research and self-correction.
So, basically, “Be a functional human being”, rather than “sit around trying to do cool things, but cycle through Facebook and Lesswrong and a couple other sites all day instead”.
Learn every human language (and invent one that is more pleasant, more practical and more efficient than them all), free up bytes by deleting my entire MP3 collection and playing all my music inside my head (I can already do this to some degree), go to bookstores and consume dozens of books at a sitting by flipping the pages and remembering every word, catch up with my pending movies by watching them all at the same time, periodically look at HD photographs of the night sky to scan for differences that might point to supernovas or incoming asteroids, and donate my spare processing power to run both LHC@home and SETI@home in my head.
Have you already invested effort into language invention?
Yes, but so far the results have been too Eurocentric.
Pain and “ugh field” asymbolia. That would be pretty awesome...
Ugh field asymbolia would be the state in which you perceive the sensations attached to “ugh fields” but do the things anyway.
Set up the subconscious to cache its processes in memory palaces, so its work can be reviewed much like a debug file.
Hmm. Can you elaborate on how that would look? it sounds like you wan tot view your own source code, but I don’t see why memory palaces are the best format (unless I don’t understand how memory palaces can be used ).
This is primarily inspired by Blink by Malcolm Gladwell combined with the cliche math teacher insisting that you show your work instead of jumping to the solution.
The goal would be to inspect your own snap judgements to understand the conclusions reached while screening for biases. So not exactly the source code but a print out of every variable the code writes, if the metaphor holds up.
A common example: you step out your door, and know it is going to rain today. You know it is going to rain because of a lifetime of experience in the region combined with the perceived cloud cover, humidity, air pressure all give you the strong hunch it will rain. However if asked to defend the hunch consciously, you would have a hard time even noticing the changes in humidity and air pressure.
Why in form of memory palaces instead of mentally visualized printouts?
The memory palace was just the first idea for storing large amounts of unconnected information.
I have no experience with memory palaces.
Achieve enlightenment.
I know my question was vague, but would you elaborate on what that means?
Well, here I meant the entirely traditional concepts of moksha or nirvana. The relevant abilities include escape from the bonds of the material world and ability to see things as they truly are :-)
Definitely take over the world. Study enough psychology to be able to guess other people’s passwords, then hack into the NSA, then download their secret blackmail stash. I now have near-complete control over almost anyone who’s been online for any significant amount of time including politicians.
My university network has password requirements strict enough that when I had to make a new one, I generated a random 8-digit string and it disallowed it because it looked too much like a real word or phrase (I have no idea what), and I had to pick a string it generated for me that was “random enough”. The NSA probably doesn’t have much looser standards for its employees’ passwords than my university has for its students’, so I expect your advanced psychology skills approaching the level of mindreading would just tell you who’s annoyed they weren’t allowed to use “password”, and who’s got an even longer string of random digits than usual.
You don’t need to study psychology for that. The lists of common passwords are freely available along with tools to apply them.
You don’t imagine they rely on text passwords, do you..?
I’m reasonably sure there’s somewhere along the chain that I’d be able to figure out how to get in at 100% control of my own brain. Maybe not specifically passwords, but if I hacked some top employees emails, I should be able to build on that. Or add myself to the top-secret clearance list and do it that way.
If the NSA used common passwords, they’d already be hacked many times over. I’m talking about getting their secure passwords, which should be a lot easier than doing it now. Social engineering when you can think many times faster than the person you’re talking to would be a game-changer. The method doesn’t matter, specifically, especially on a hypothetical.
Are you taking the position that with a perfect brain, it would still be impractical to get into NSA databases? What about multiple brains, then?
Proper security (nowadays) usually depends on a mix of three things—what you know (passwords), what you have (physical tokens), and what you are (biometrics).
Not sure which capabilities do you think a “perfect” brain will have. You still won’t become an X-man.
Do you think it’s possible to get into the NSA at some point along the security chain with only “what you know (passwords)”? Just curious.
I would assume that that means no more cognitive biases, perfect memory, no logical errors, etc. The increase in processing speed that would come from no longer having to deal with the corresponding mental garbage would no doubt be quite impressive. Furthermore, if you have “perfect control” as specified by the OP, you ought to be able to simulate people just by forcing your mind to act like theirs. It doesn’t seem like too much of a stretch that you might be able to guess some NSA workers’ passwords while simulating them with near-perfect fidelity.
Depends on what do you mean by “get into the NSA”, of course, but I’d rate your chances as very low, especially post-Snowden.
To start with, I assume their sensitive information is air-gapped.
Not sure about that. You still have the same underlying biology, you don’t get to optimize your neuron topology, and neurons are slow.
No, that wouldn’t work because you don’t have enough information about their mind in order to simulate it.
Guessing passwords is not the hard part :-)
Attacking the NSA is very high risk and probably still a dumb move.
I’d crank up the pleasure center of my brain until the handle broke off.
Why hasn’t anyone mentioned this yet? I’m sad to see that others would sacrifice greatly just to be more capable or powerful.
Hopefully, self-deception is to blame; others want to want to improve themselves instead, but if they were able to sample bliss, they wouldn’t trade it for anything.
Speaking of Social Justice Warriors (SJW’s) versus a man familiar to many of us:
http://www.bloombergview.com/articles/2014-11-21/economics-is-a-dismal-science-for-women
http://www.overcomingbias.com/2014/11/hanson-loves-moose-caca.html
http://www.unz.com/isteve/noah-smith-tries-to-sic-shirtstorm-mob-on-poor-robin-hanson/
Am I the only one to whom ‘what’s wrong with raping someone if they don’t get injured, traumatized, pregnant, nor get STDs’ sounds a lot like ‘what’s wrong with driving at 100 km/h while drunk, sleep-deprived and talking on the phone if you don’t have any accidents’?
You obviously missed the point completely. Hanson’s thought experiment wasn’t claiming there would be nothing wrong with committing that type of rape, his point was that it would be traumatic in the same way being cuckolded is traumatic. And yet committing that type of rape is illegal, while cuckolding is not even a misdemeanor. His point wasn’t to denigrate the psychological harm of rape, it was to investigate the roots of the difference in the way these harms are treated.
It’s not the same thing, rape is still a violation of bodily integrity even if there’s no health damage of memories.
If a sterile man without STDs drugs a women to be unconscious and then rapes her, that’s still wrong and still should be punished strongly by law. The sacred value of a woman’s control over her own body is still violated. The fact that Hanson doesn’t address that sacred value is what makes his post a bit creepy.
It isn’t just a matter of “sacred values”. It’s a matter of the consequences of making the statement.
“What’s wrong with doing if ?” will predictably have the effect, on the margin, of encouraging people to do even when don’t actually hold. We can predict this for reasons closely analogous to why knowing about biases can hurt people: Arming people with more rationalizations for bad things that they already were tempted to do will generally make them worse, not better.
Conducting motivated search for conditions under which something normally very harmful can be justified as barely non-harmful is the sort of thing someone would do, in conversation, if they wanted to negotiate down the badness of a specific act.
“What I did isn’t real reckless driving. In real reckless driving — the maximally bad sort — the driver has to be driving too fast, while drunk, sleep-deprived, and talking on the phone. Me, I was only sleep-deprived. So stop treating me like my drugged-out ass ran over a dozen schoolkids or something.”
(See actual political discussions of “real rape”.)
Sacred values is a term out of modern decision theory. Putting quotes around it is like putting quotes around cognitive bias.
I don’t think that’s a strong argument. It’s quite useful to play out scenarios of “is X still a bad idea if we change Y” to understand why we think X is a bad idea. It’s how you do reductionist analysis. You reduce something into separate parts to see which of those parts is the real issue.
If I say: “Stealing is bad but there are cases where a person has to steal to avoid starvation.”, that’s a permissible statement. We don’t ban that kind of analysis just because stealing is generally bad.
I think it’s quite foolish to believe that a societal debate about what rape happens to be is bad when your goal is to reduce rape. Tabooing that discussion prevents people to speak in polite company openly about issues of consent and as a result a lot of people don’t think deeply about those issues and make bad decisions.
It’s silly to try to raise rape awareness while at the same time wanting to prevent the topic from getting discussed.
Which is extremely idiotic and mostly seems to consist of feminists attempting to get away with further and further expanding the definition of “rape” while keeping the word’s connotations the same.
Would you accept the same argument for cuckoldry violating the sacredness value of the marriage? If so then wasn’t Hanson comparing two sacredness violations? If not how do you decide which sacredness values to accept?
No, Hanson didn’t speak about sacred values at all. If one wants to make the argument that marriage is sacred and violations should be punished then the logical conclusion are laws that criminalize adultery. Hanson is not in favor of those.
Yes, a little bit.
However, one difference is that in the rape example, some people in the audience are more likely to see themselves in the role of the rapist (“Is he saying that I should be allowed to get away with that, if it really caused no harm?”) whereas others are more likely to see themselves in the role of the rape victim (“Is he saying that someone should be allowed to do that to me, if they can explain away my objections?”).
Summary:
1. Some economists (including Robin Hanson) said some things using rape as an example to illustrate abstract theoretical points, which pissed a lot of people off to whom rape is not an abstract matter. One of them (Steve Landsburg) also participated in sexual harassment of an activist. Some other economists (including Lawrence Summers) claimed that women are (relatively) bad at math and science.
Some yet other economists actually performed a study and found that economics has “a persistent sex gap in promotion that cannot be readily explained by productivity differences” and that “the average female economist—unlike the average female physicist or mathematician—is likely to have a better publication record than her male peers”, and thus that economics as an institution is biased against women practitioners.
2. Robin Hanson vigorously agrees that economics as an institution is biased against women practitioners, but doesn’t think his use of rape as a theoretical example has anything to do with that, and certainly that he didn’t intend to minimize rape. Just the opposite: “Just as people who accuse others of being like Hitler do not usually intend to praise Hitler, people who compare other harms to rape usually intend to emphasize how big are those other harms, not how small is rape.” He cites professional racist Steve Sailer in his defense, because that’s really going to change anyone’s minds.
3. Steve Sailer calls a bunch of people vile names. Yawn.
My conclusion: Robin Hanson is probably not personally responsible for the observed phenomenon of economics as an institution being biased against women practitioners. But the level of instrumental rationality and/or good rhetoric demonstrated here is not super great. If you want to convince people that you are not a bigot, why cite a bigot’s defense of you as his ally?
Second, it isn’t clear to me that the “badness” of using rape as a theoretical example has anything to do with minimizing it. Rather, it has to do with choosing to offhandedly mention something really awful which is linked to the victimization of a fraction of your students or readers. Being constantly squicked by gratuitous scary examples that selectively target you is not great for the concentration.
In a book about management that my partner S was reading recently, there was an analogy where managers are compared to ship captains. A page later, bad managers are described as “whipping” their underlings — that is, using punishment as an incentive. Now, S is from the West Indies, and the proximity of “ship captain” and “whipping” immediately reminded her of the slave trade, and gave her a pretty heavy-duty squick reaction. Does this mean the authors of the book intended to turn people off of being managers through stereotype threat or something? No. Does it mean they even intended to remind anyone of the slave trade? No. Does it mean that these analogies cumulatively had the effect of distracting S from the point the authors were trying to make? Yes.
If the authors were challenged on this, would they resort to citing professional racists in their defense? Probably not; they’d probably say something like “oops, wow, that was totally unintentional and really embarrassing.”
This is so obviously and egregiously false that it makes me question your bona fides.
Okay, how about “Steve Landsburg justified and participated in calling someone a lot of sexual slurs in response to that person’s testimony before Congress”?
(I’ll admit that “sexual harassment” is a pretty vague term and thus ill-chosen. Landsburg’s behavior was still detestable and vile by either progressive or conservative standards of acceptable public conduct: he managed to be both a frothing sexist and an ungentlemanly boor.)
Still false. Easily verifiably, objectively, false.
At best, you have a reckless disregard for the truth. I’m out.
EDIT: Just to be clear for others reading this, the post has never been deleted, so anyone can check it right here. Landsburg does not call Fluke either a slut or a prostitute. He calls her an “extortionist with an overweening sense of entitlement” which is no compliment, but not a sexual slur. You will note that even Noah Smith, in his vile hit piece, does not accuse Landsburg of sexual slurs. Instead he says that Landsburg ‘seemed to call pro-contraception activist Sandra Fluke a “prostitute”’ (italics mine). Why this “seemed”? Because if you read the post, you’ll note that Landsburg explicitly stated that calling Fluke a prostitute was wrong! Yet people who I can only assume engaged in deliberate misreading used it to accuse him of calling her a prostitute, and Smith now piggybacks on those accusations to say “seemed” about a statement that he knows is a tawdry lie—but by using that word, he can claim he never said anything false himself.
It’s disgusting, and Smith, and you, should be ashamed of yourselves.
Stretching terms like “sexual harassment” to apply them to more people doesn’t do good. It weakens them when they are used in the proper context. There no reason to call everything that’s not acceptable public conduct “sexual harassment”.
As I said, vague and ill-chosen.
Still, the problem remains. If you lie down with dogs, you get up with fleas. Hanson’s choice to associate himself — and his audience — with people who do nasty and rude things in public and with people who espouse and practice harming others for their views is unfortunate. It is especially unfortunate in this context; it thoroughly undermines his apparent attempt to dissociate himself from some of the same.
I think vague is a cop out. The term does have a clear meaning. It doesn’t exist that you can throw it out as a slur. The fact that you incorrectly slur other people to complain about them engaging in slurring has it’s irony.
I think falsely accusing someone of sexual harassment is detestable and vile.
I wonder if Eliezer will now attempt to disassociate himself from Hanson like he has the NRx’s.
I was writing a Markov text generator yesterday, and happened to have a bunch of corpora made up of Less Wrong comments lying around from a previous toy project. This quickly resulted in the Automated Gwern Comment Generator, and then the Automated sixes_and_sevens Comment Generator.
Anyone who’s ever messed around with text generated from simple Markov processes (or taken the time to read the content of some of their spam messages) will be familiar with the hilarious, and sometimes strangely lucid, garbage they come out with. Here is AutoGwern:
(I should point out that I’m picking on Gwern because his contribution to LW means I have an especially massive text file with his name on it.)
Here’s some AutoMe:
I have become weirdly fascinated by these. Although they ditch any comprehensible meaning, they preserve distinctive tics in writing style and vocabulary, and in doing so, they preserve elements of tone and sentiment. Without any meaningful content to try and parse, it’s a lot easier to observe what the written style feels like.
On a less introspective note, it’s also interesting to note how dumping out ~300 words maintains characteristics of Less Wrong posts, like edits at the end, or spontaneous patches of Markov-generated rot13.
Also Yvain could happen when I feel like my brain compels me to be the money pump.
Well that’s a fanfiction I haven’t read before.
This should be generated for every user on their “overview” page. :D
I did this a while back using cobe: dumped in
gwern.net
, all my IRC logs, LW comments etc. It seems that the key is to set the depth fairly high to get more meaningful language out, but then it works eerily well: with some curation, it really did sound like me! (As long as you were reading quickly and weren’t trying to extract real meaning out of it. I wonder if a recurrent neural network could do even better?) You can see some at http://pastebin.com/tPGL300J—I particularly likeCouple of random thoughts about cryonics:
It would actually be better to have cryonics legally recognized as a burial ritual than as a cadaver experimentation. In that way it can be performed on someone who hasn’t formally signed a will, granting their body as an anatomical gift to the cryonic service provider. Sure, ideally it should be considered a medical procedure on a living person in a critical condition, but passing such legislation is next to impossible in the foreseeable future, whereas the former sounds quite feasible.
The stabilization procedure should be recognized as an acceptable form of active euthanasia. This is probably the shortest way to get to work with not yet brain-dead humans, and it would allow people to trade couple of months or years of rather painful live for better chances at living again.
Insulin injections should probably be a part of the stabilization protocol (especially in the previous case). According to “Glucose affects the severity of hypoxic-ischemic brain injury in newborn pigs” by LeBlanc MH et al, hypoglycemic brains sustain hypoxia much better than normally. That totally makes sense: oxygen is mainly consumed for glycolysis, so if there’s nothing to oxidize, oxygen consumption will decrease.
Some of the major problems of cryonics can probably be solved by preventing water from expanding upon freezing. According to 1 and 2, ice is denser than water at about 30 kbar. That is a bit technically complicated, but I would speculate that with this trick we could have reversible freezing in wood frogs right now.
That would destroy cryonics companies who make money via insurance that depends on people legally dying.
What do you mean exactly? If I understood it right then vitrification is done to prevent ice crystals from forming. Do you mean something different?
Vitrification also amounts to antifreeze poisoning. Quite a big deal if you’re counting on full-body revival (which, granted, I don’t consider feasible in the first place—cryonics is, in the best case, an information backup as far as I’m concerned).
That’s the whole point: if we can if we can prevent water from expanding by freezing and keeping the sample under high pressure, thus making crystal formation harmless (probably), we can use less cryoprotectant. I don’t know if it’s possible to get rid of it completely, so I mentioned wood frogs, that already have all the mechanisms necessary to survive slightly below the freezing temperature. It’s just their cryoprotectant isn’t good enough to go any colder, but it’s not so poisonous either. Also, they’re small, so it’s easier to find high pressure units to fit them in—they’re perfect model organisms for cryonics research.
As of now, cryonics is at best an information backup indeed, but I see no reason why we should be content with that. Yes, we will probably eventually invent advance nanomachinery, as well as whole brain simulation and scanning, but that’s too many unknowns in the equation. We could do much better than that.
The insurance companies would have to alter their contracts. Ideally, whatever legislation classifies cryonics as a medical procedure would include a clause that for the purposes of any contract written before then, it counts as death.
Wouldn’t it just shift to health insurance in this case? But generally, yes, recognizing cryonic patients as alive has a lot of legal ramifications. On the other hand, it provides a much better protection against unfreezing: just like with the patients in a persistent vegetative state, someone authorized has to actively make a decision to kill them, as opposed to no legal protection at all. I’m not sure which of these is the net positive. Besides, that would challenge the current definition of death, which currently basically boils down to “we positively can do nothing to bring the patient back from this state”. Including potential technologies in the definition is a rather big perspective change, that can also have consequences for vegetative patients as well.
As ZankerH mentioned below, vitrification leads to cryoprotectant poisoning, which is a sufficiently big problem to prevent us from experimenting with unfreezing even in small organisms. If the function of the cryoprotectant can be fully or partially replaced by keeping the sample under the high pressure, that problem is mostly solved. That doesn’t prevent crystals from forming, but unlike normal ice, these crystals take less volume than the water they were made of, so they shouldn’t damage the cells. In addition, amorphous solids aren’t guaranteed to be stable, and can undergo slow crystallization. I’m not sure how pig of a problem that is for cryonics, but in case of going directly to ice-IX, that’s definitely not a problem anymore.
Health insurance doesn’t automatically pay for everything.
A bit but not that much. A company that can’t afford to freeze would still shut down. Nobody is foreced to pay to keep the machines on in every case.
This year’s edition of Stoic Week begins next monday. It’s a seven day long introduction to the basic ideas of Stoicism combined with an attempt to measure its effects.
I have signed up for this, and also registered for the meeting in London on the 29th. Has anyone else?
I am at the event in London right now. Anyone else?
Anyone know of, or have links to research into the question of whether the near-unlimited access to global media via the Internet, etc. has a net negative effect on people’s self esteem?
Couple examples...
Facebook
My brother quit Facebook saying, “I know these people’s lives are not nearly interesting as their feed’s make them out to be. Every time I hang out with these people… and most of them sit around trying to decide what to watch on Netflix six nights out of seven. It’s annoying.”
Excepting my cynical brother, FB seems to dupe a lot of people—I’ve heard it anecdotally and it seems like a reasonable hypothesis—into believing the lives of their friends are much better and more interesting than their own. I hear people say, “People these attractive pictures at fun events while I’m at home eating reheated Noodles & Co in my sweatpants...”
The Internet and Perceptions of Beauty
I had a discussion about this a female friend recently as to whether she’d considered female perceptions of beauty had become way out of whack since the Internet.
Thinking about it… In 1930, you might be the most beautiful women anyone had ever seen in your lifetime. And because people were disconnected, there would be thousands, or even millions, of “Most-Beautiful-Evers”.
In 1980, there’s going be much fewer MBEs due to TV and print media.
In 2014, there is a short list of MBEs due to...well...Buzzfeed alone, and everyone else is feeling comparatively less attractive as a result.
It seems to be the sort of “big fish in little pond” scenario everyone goes through as they progress in life. Except accelerated by technology.
Is there more info on this anyone is aware of?
(Note: Anticipating criticism of using females in the second example… Sorry. Feel free to insert male interchangeably if it suits you.)
There is TONS of mostly feminist literature on how movies/internet/Barbie/etc. teach women their bodies and themselves are inadequate and so lower their self-esteem. It is a very popular feminist trope.
Sure. I should have been clearer. I guess I’m more curious about the degree to which this exists because of the internet. Assuming it’s true that self-esteem takes a negative hit (at least initially) due to realizing you’re a small fish in a huge pond, does the availability of information compound that effect to a level where it is exponentially consequential?
Ultimately, I think it’s a very natural and spiritual process to go through—to realize you are, in fact, not the center of the universe. But it’s been a more gradual process in our history. Nowadays, it’s like saturation in the idea that “I suck. I’m boring. I’m ugly” 24⁄7 at the age of 10 or before.
Could you, please, debuzzwordify your question and add some meaning to it?
I’ll try to be more clear. No promises.
Let’s assume there is a social comparison element to self-esteem. Whether it be strength, intelligence, tennis ability, driving skills, beauty—People reference where they stand in any given dimension among the total population to gauge how they feel about themselves.
If you are in the 99th percentile of a good trait based on the portion of the population you’re aware of, you generally feel good about identifying with that. Given awareness of a large enough population, you might fall from 99th to the 50th percentile. Or to the 10th. Your trait was unchanged, but your awareness of it’s rank among the whole population changed.
Is it possible this exponential leap in the sheer number of people we can (or are almost forced to) compare ourselves because of the internet corresponds directly to the hit our self-esteem takes? If so, how severe are the consequences?
I’ve read some about ‘Facebook Depression’, where one of the components is the envy and disillusionment that comes from seeing how everyone else is doing cool things while you’re sitting at home. I was just wondering about people’s thoughts about the Internet writ large.
First, people differentiate between the online world and the meatspace world. If I am the prettiest girl at school, I may be aware that Kim Kardashian’s butt almost broke the internet and I (hopefully) realize that my butt does not have the same capabilities, but I am still the prettiest girl at my school. “Population” is different and here the relevant metric is closeness to you. Comparing yourself to people you meet every day is rather different from comparing yourself to pop star pictures on a screen.
Second, this whole self-esteem thing is driven by System 1 and System 1 is pretty bad at large numbers. In fact, I suspect that that System 1 counts like this: one, two, three, Miller’s number, ~12, ~30, Dunbar’s number, many (see e.g. this). In this progression the jump from 1 million to 10 million doesn’t happen—both are “many”.
Of course. But I’m not sure they are able to fully separate the two. Internet has some effect, right?
True. And people will always enjoy benefits from being the biggest fish in their particular pond, whatever population defines that.
A person may believe they are in the 99th percentile by assuming the population they are in is representative of the global population. Even if this isn’t true, there will be zero loss of self-esteem without sufficient spread of information, since said person may never become aware of the objective reality that they are in fact in, say, the bottom 20%.
Why?
It certainly has, we are only discussing its magnitude.
For example, you interact with the former but you do not with the latter.
Your language left it unclear to me whether you thought the differentiation to be total.
Can you explain how that effects anything? (In fact, I’ve seen cases to the contrary of what you seems to be suggesting all the time. People often idolize pop stars they’ve never met and think of them as flawless...due, in some part, to the fact they don’t get to see them up close in their unedited regular-ness.)
I think I understand what you’re talking about.
I didn’t get internet access until I was almost in my 20s. So I grew up with certain talents where friends/family would consistently tell me that I was the best at what I did. Nowadays, you can go to online discussion boards where people who are the best of the best in field X congregate and see just how “average” you are in that bigger pond.
Though I was good enough to get into specialized high school/colleges for that, I chose not to go that route. I’m guessing that the same sort of seeing how average I was in that larger pond where everyone is at the top of their game would have happened anyway had I gone to those specialized schools.
This is why people should move to tumblr, where the people know they’re boring and ugly and celebrate that fact :-)
(I’m joking, but in my experience the LW-tumblr space is very accepting of and open about defects like that. Despite the insistence of calling everyone cute.)
I’m not if it works with physical attractiveness, but in case of intellectual adequacy, I’m just not letting any internal doubts in my competence to interfere with external confidence. Even if I suspect that I’m not as smart as people around me, I still act exactly as if I am.
I think the main damage may actually come from female perception of male sexiness, (and conversely), e.g., a girl being disappointed in her boyfriend because he isn’t Brad Pitt.
Edit: fixed.
Female perceptions of beauty were way out of whack before the internet. The internet may be making things worse, but I think there’s a natural drift towards more extreme standards. If the ideal is to be strikingly thin, then the demand will be to be ever thinner.
Male standards have gotten more extreme as a body builder aesthetic has partially taken hold.
In the early 90s there was a TV show called Time Trax, about an Übermensch law enforcement officer from 200 years in the future, sent back to apprehend fugitives from his own time, who were carrying out futuristic crimes in the 20th century.
I re-watched a couple of episodes recently. Something that stood out, beyond the Windows Movie Maker level special effects, was how non-ripped the main character was.
Thinness is a very superficial standard. It’s something you can use to judge people who are fake and who hide their emotions. You don’t need more than a picture to see whether someone is thin.
It harder to visually distinguish whether someone is depressed. The average person won’t identify the physical signs for it, especially not via seeing an image on the internet.
Yesterday I was dancing a while Salsa with a girl who’s a beginner but dances for a few months. She’s thin but she doesn’t really relax. Even after a few songs of dancing her hands are still cold because they aren’t well vascularized. I find a girl who weighs a bit more but who’s in touch with her body more beautiful.
Height is also a very interested topic. Plenty of girls think they would be more beautiful if they would be 1,80m tall instead of 1,60m tall because models are tall. On the catwalk being tall matters. On the other hand I think a majority of guys prefer a 1,60m tall girl over a 1,80m tall girl. Shorter girls get more OkCupid messages than taller girls.
Wy is that the “main” damage? I’d agree mmale appearance standards have also changed, but generic western society values women on their appearance more than men so you’d expect the psychological impat to be larger.
Sorry, I was revising the comment, mistake corrected.
On the other hand, no matter how weird your kink is, the Internet will let you find other people who share it.
Advice/help needed: how do I study math by doing lots of exercises when there’s nobody there to clue me in when I get stuck?
It’s a stupid problem, but because of it I’ve been stuck on embarrassingly simple math since forever, when (considering all the textbooks and resources I have and the length of time I’ve had it as a goal) I should have been years ahead of my peers. Instead, I’m many years behind. (Truth be told, when performance is tested I’m about the same as my peers. But that’s because my peers and I have only struggled for a passing grade. That’s not what my standard of knowledge is. I want to learn everything as thoroughly as possible, to exhaust the textbook as a source of info; I usually do this by writing down the entire textbook, or at least every non-filler info.)
There is a great disparity between the level of math I’ve been acquainted with during my education, and the level of math at which I can actually do all the exercises effortlessly. In theory by now I’m well into undergraduate calculus and linear algebra. In practice I need to finish a precalculus exercise book (tried and couldn’t). While I’m learning math, I constantly oscillate between boredom (“I’m too old for this shit” ; “I’ve seen this proof tens of times before”) and the feeling of getting stuck on a simple matter because of a momentary lack of algebraic insight (“I could solve this in an instant if only I could get rid of that radical”). I’ve searched for places online where I could get my “homework” questions answered, but they all have rather stringent rules that I must follow to get help, and they’d probably ban me if I abused the forums in question.
This problem has snowballed too much by now. I kept postponing learning calculus (for which I’ve had the intuitions since before 11th grade when they began teaching it to us) and therefore all of physics (which I’d terribly love to learn in-depth), as well as other fields of math or other disciplines entirely (because my priority list was already topped by something else).
I’ve considered tutoring, but it’s fairly expensive, and my (or my tutor’s) schedule wouldn’t allow me to get as much tutoring as I would need to—given that I sometimes only have time to study during the night.
Do any LessWrongers have resources for me to get my questions answered? Especially considering that, at least at the beginning until I get the hang of it, I will be posting loads of these. Tens to hundreds in my estimation.
Sorry to jump from object-level to meta-level here but it seems pretty clear that the problem here is not just about math. Your subjective assessments of how difficult these topics are is inconsistent with how well you report you are doing at them. And you’re attaching emotions of shame and panic (“problem has snowballed”) to observations that should just be objective descriptions of where you are now. Get these issues figured out first (unless you’re in some educational setting with its own deadlines). Math isn’t going anywhere; it will still be there when you’re in a place where doing it won’t cause you distress.
I can see how it would sound to an outside observer now that you point it out, but in my situation at least I have trouble buying into the idea that math isn’t going anywhere. The problem really is urgent; there are loads of fields I want to study that build upon math (and then upon each other), and it just isn’t feasible that I can further postpone deep, lasting learning of basic math concepts any further, until after I’m in the “right mindset” for it. There just isn’t time, and my neuroplasticity won’t get any better with age. It’ll take me at least a decade to reach the level I desire in all these fields. Not to mention that I’ve long since been having trouble with motivation, or else I could have been done with this specific math in about 2011-2012. I’m not doing well at these topics (despite evaluating them as easy) because I spend less than a few hours per month on them.
How old are you? I think the peak is around 30 years of age.
I’m 21. I thought it began to decline after early/mid 20s. It’ll definitely take me longer than that to learn just a few prerequisites thoroughly.
In my opinion, what you probably need is some mix of tutoring and therapy—I don’t know if it exists, and if there is a word for it—someone who would guide you through the specific problem, but also through your thought processes, to discover what it is you are doing wrong. Not just what are you doing wrong mathematically, but also what are you doing wrong psychologically. This assumes you would speak your thoughs aloud while solving the problem.
The psychological level is probably more important than the mathematical level, because once the problems with thinking are fixed, you can continue to study maths on your own. But this probably cannot be done without doing the specific mathematical problems, because it’s your thought patterns about those math problems that need to be examined and fixed.
I happen to have a background in math and psychology and teaching, so if you would like to try a free Skype lesson or two, send me an e-mail to “viliam AT bur.sk”. Don’t worry about wasting my time, since it was my idea. (Worst case: we will waste two hours of time and see that it doesn’t work this way. Best case: your problem with maths is fixed, I get an interesting professional experience.)
Thanks for the offer! Yes, this sounds interesting. One of the things I’ve tried to get out of my math tutoring experience was to see how the teacher looked at the problem before beginning to actually solve it and see if it works out. But they never actually thought out loud for me to understand their thought process; also, I often can’t tell what mental resources someone else uses when solving a difficult/tricky math problem. (Experience? Sudden insight? Additional notions I don’t yet have?)
Find somebody else who is in the same situation and study together.
I’m wondering whether you’re expecting some math to be effortless when you’re actually still at the stage of trying one thing and another, and need to be less focused on how you want the process to feel.
Effortless, no; however, some of it is sufficiently familiar by now that I don’t think there is additional value in rehearsing the same material. (Example: I first encountered derivatives on Khan Academy, before they officially taught them to us; then in 11th grade as an intro to calculus; then on several tests and exams; then in freshman year of college. I don’t think I need to be given the intuition on derivatives one more time, and if I took a test in derivatives I’d ace it without much effort.) I’m not expecting new notions to be very easy to learn—moderately challenging and definitely not impossible, but not easy.
Have you looked into Physics Forums? They have Homework & Coursework Questions subforum.
Yes. Recently made an account there, started off by asking a homework question. Wouldn’t want to spam it, though.
Maybe you should post here about one or two or the problems you’re stuck on.
I’m wondering whether you’re demanding too much facility of yourself before you go on, but this is only a guess.
Have you looked into Khan Academy?
Only if the LW Study Hall organized threads like that; I’ve been considering joining it, but never got around to it. Otherwise, I’d feel it would be a waste of people’s time. There are lots of exercises I might get stuck on, and it’s usually not as if any particular one of them represents a key insight to me; the answer/solution/hint to any given exercise is low-value to me, but the answer to all or most of them is high-value.
Been on Khan Academy since 2012, I think. Earned a Sun Badge and everything. (That type of performance didn’t constitute my real challenge, though; I’ve been struggling for these two years to earn the modest Good Habits badge, to no avail to this day.)
In general having a goal to exhaut textbooks by copying them is stupid. That’s not what they are for. Get rid of that strategy.
Most forums have rules that prevent you from asking questions without deeply thinking about those questions yourself.
On math.stackexchange you are allowed to ask all maths questions provided you search before for similar questions and put in the effort to write a decent question.
Asking decent questions is a skill. Learn it and people will answer your questions.
Suspected as much… but I am not sure what strategy to replace it with. I definitely plan on doing all the exercises, and sometimes if I also write down some of the theory/proofs it helps with recalling them later. I’m guessing good rules of thumb are: stick to the essentials; use common sense; review at appropriate intervals.
I’m fine at asking a question that doubtlessly won’t get removed. A question. I’m not sure that asking 20 (good, rule-abiding, thought-out) questions per week in a given forum would last me very long. That’s about how often I might get stuck if I do some exercises every day.
On stackexchange nobody has a problem with someone asking a lot of question provided they are good, rule-abiding and thought-out.
You can also answer questions of other people. It might even be better than textbook problem because someone well tell you when you are wrong via comments.
Have you tried using WolframAlpha? If you get the pro subscription (~$66 per year - ~$45 a year for students) you gain access to their step by step solutions and unlimited problem generator. I am currently studying a Masters of Biostatistics (which has a heavy calculus section) and found this service invaluable.
Examples http://i59.tinypic.com/2yuniv9.png http://i60.tinypic.com/c5i7n.png
You should go on to more advanced topics without being fluent in less advanced topics. You can always later go back to the exercises that stumped you. The typical student taking calculus without being fluent in adding fractions is a mistake, but that don’t over-correct that mistake.
A riddle for Lesswrong: what exactly is the virtue of Bissonomy?
When I read the article, I got the feeling that there were enough clues to extrapolate a solution in the same way that EY extrapolated the Dementors’ ‘true natures’. That this was a solvable riddle. I’ve got my suspicions, but I’d like to hear what you guys can come up with.
Bissonomy is the virtue of stability.
Specifically, bissonomy is the virtue of knowing (-nomy) how to attach yourself (“bissys” is the name for filaments by which certain bivalves attach themselves to rocks and other substrate) to some stable object.
Arguments:
The name. It basically tells you outright (and “bisso” is Portuguese for “bissys”, just in case).
She was turned into oysters—bivalves—which is a big hint.
She was punished for throwing a mole. What do moles do? They dig! They undermine and clearly, the virtue of stability couldn’t be seen undermining anything.
Two children is another hint at stability as families with two children neither increase nor decrease the population—they keep it stable.
She was forgotten—for the only thing constant in the world is change.
Sounds like a noodle virtue to me, or at least it uses the same basic idea for humour.
But if you want to keep in the spirit of EY’s dementors, the article writer does some wacky reasoning to end up roughly at the virtue of dietary restriction.
Terry Pratchett’s noodle implements generally make sense if you work them out. I doubt this is any different.
A-ha! That makes sense! Also, it’s actually an important virtue! People judge you on it!
You say that this feels like a riddle to you, but I would prefer to call it a koan. I think the best hint that we have to the nature of Bissonomy lies in the vagueness of its descriptions, as this is the sort of irony that Pratchett is famed for, and this sort of self-answering riddle has a Buddhist feel to it in turn. This also seems like a useful starting point for another reason: there are only so many potential virtues that Bissonomy can plausibly be, and the standard Western ones are already accounted for.
I suspect Bissonomy is the emotional acceptance of both that which is known to be true and that which is uncertain or unknowable. Ignorance, as we all know, is purported to be bliss, and that sounds like biss. However, Bissonomy is not the embrace of ignorance, nor does it bring bliss. Rather, it is the acceptance of that which is known to be unknowable, and it brings inner-peace.
Bissonomy and Tubso seem to be connected. Both virtues were forgotten due to their rarity. Explaining what one was should hopefully tell us something about the other. But it’s even harder to say things about Tubso than about Bissonomy; the only thing that we know about Tubso is that its name is absurd. In a world where nominative determinism exists, however, this might be the only hint we need. I submit that Tubso is the virtue of absurdism, which surely has a place in Discworld. This fits with the koan framework, and establishes the desired connection between the two lost virtues; a koan’s answer is absurd and hints at strange knowledge, but truly understanding a koan requires the recognition that one can never understand it fully, if at all.
Of necessity, this theory is largely speculative. We may never know the true answer to this question, and that’s okay. And that, in turn, is Bissonomy.
This sounds somewhat like a specialized form of the very Christian value of Resignation, specifically resignation towards the Ineffability of God and his Mysterious Ways, and the seemingly chaotic creation.
Then again God often plays a Dao-like “empty center that lets the wheel turn” in these sorts of doctrines.
You say that this feels like a riddle to you, but I would prefer to call it a koan. I think the best hint that we have to the nature of Bissonomy lies in the vagueness of its descriptions, and there is a Buddhist feel to this. There is also the sort of irony that Pratchett is famous for. Also, this seems like a useful starting point for another reason: there are only so many potential virtues that Bissonomy can plausibly be, and all the standard Western ones are already accounted for.
I suspect Bissonomy is the emotional acceptance of both that which is known to be true and that which is uncertain or unknowable. Ignorance, as we all know, is purported to be bliss, and that sounds like biss. However, Bissonomy is not the embrace of ignorance, nor does it bring bliss. Rather, it is the acceptance of that which is known to be unknowable, and it brings inner-peace.
Bissonomy and Tubso seem to be connected. Both virtues were forgotten due to their rarity. Explaining what one was should hopefully tell us something about the other. But it’s even harder to say things about Tubso than about Bissonomy; the only thing that we know about Tubso is that its name is absurd. In a world where nominative determinism exists, however, this might be the only hint we need. I submit that Tubso is the virtue of absurdism, which surely has a place in Discworld. This fits with the koan framework, and establishes the desired connection between the two lost virtues; a koan’s answer is absurd and hints at strange knowledge, but truly understanding a koan requires the recognition that one can never understand it fully, if at all.
Of necessity, this theory is largely speculative. We may never know the true answer to this question, and that’s okay. And that, in turn, is Bissonomy.
You say that this feels like a riddle to you, but I would prefer to call it a koan. I think the best hint that we have to the nature of Bissonomy lies in the vagueness of its descriptions, and there is a Buddhist feel to this. There is also the sort of irony that Pratchett is famous for. Also, this seems like a useful starting point for another reason: there are only so many potential virtues that Bissonomy can plausibly be, and all the standard Western ones are already accounted for.
I suspect Bissonomy is the emotional acceptance of both that which is known to be true and that which is uncertain or unknowable. Ignorance, as we all know, is purported to be bliss, and that sounds like biss. However, Bissonomy is not the embrace of ignorance, nor does it bring bliss. Rather, it is the acceptance of that which is known to be unknowable, and it brings inner-peace.
Bissonomy and Tubso seem to be connected. Both virtues were forgotten due to their rarity. Explaining what one was should hopefully tell us something about the other. But it’s even harder to say things about Tubso than about Bissonomy, the only thing that we know about Tubso is that its name is absurd. In a world where nominative determinism exists, however, this might be the only hint we need. I submit that Tubso is the virtue of absurdism, which surely has a place in Discworld. This fits with the koan framework, and establishes the desired connection between the two lost virtues; a koan’s answer is absurd and hints at strange knowledge, but truly understanding a koan requires the recognition that one can never understand it fully, if at all.
Of necessity, this theory is largely speculative. We may never know the true answer to this question, and that’s okay. And that, in turn, is Bissonomy.
This is rather disorganized, unfortunately. Don’t read ahead unless you’re willing to be frustrated by near stream-of-consciousness writing.
http://english.stackexchange.com/questions/116456/meaning-of-onomy-ology-and-ography
I strongly suspect that the virtue of Bissonomy is the virtue of acknowledging that which you do not or can not understand, also known as epistemic humility. (Yes, faith is its sister virtue—don’t ask how they’re both able to be virtues at once despite their symmetrical character, as that would mortally offend both of them.) The vague and unspecified nature of Bissonomy is the best clue we have as to its nature, there is a Buddhist feel to it. This has enough of a satiric bite to it that I can imagine the author actually writing it.
Biss sounds like bliss, which as we all know is the result of ignorance. But Bissonomy is not the intellectual embrace of ignorance, although it might sound like it (haha). Rather, Bissonomy is the emotional acceptance/knowledge that our map will never fully match the territory, and that some parts of the territory are not mappable at all (eg uncertainties of quantum measurements).
The main reason to be skeptical of this idea, other than a severe lack of evidence for it (perhaps this allows it to be reconciled with faith, though :p), is that it fails to explain the virtue of Tubso, and it seems to me like understanding one of the lost virtues might require also understanding the other.
However, I have a suspicion that the virtue of Tubso might be something absurd like “not throwing shellfish at the shadows of deities”, or perhaps Tubso is absurdism itself. The word certainly sounds silly enough for it. If we adopt the perspective that Bissonomy and Tubso are the “lost Buddhist” virtues, in my opinion this makes quite a lot of sense. A hint of absurdism paired with peaceful acceptance, just like a koan paired with its answer.
I think adopting this perspective seems like a good idea. There are only so many potential virtues, after all. And although I’ve not read enough of Pratchett’s work, I would bet he’s used non-Western culture as a source of inspiration before.
Another reason to be skeptical of my interpretation is that I ironically exhibit a lack of Bissonomy myself in so hastily jumping to the conclusion that it is possible to deduce what Bissonomy means, despite that Prachett provides little evidence around it. But perhaps this is simply another aspect of Prachett’s joke—forcing the reader to be hypocritical in order to understand the virtue of Bissonomy would be like a half-serious chastisement/reward for spoiling the fun of leaving the virtue unknown, similar to how trying very hard to find a true answer to a koan is a bad but necessary step in allowing the koan to change the way your mind thinks.
Ultimately, we may never know. And that’s okay. And that’s Bissonomy.
http://english.stackexchange.com/questions/116456/meaning-of-onomy-ology-and-ography
I suspect that the virtue of Bissonomy is the virtue of acknowledging that which you do not or can not understand, also known as epistemic humility. (Faith is its sister virtue—don’t ask how they’re both able to be virtues at once despite their symmetrical nature, that would mortally offend both of them!) The vague and unspecified nature of Bissonomy is the best clue we have as to its nature, there is a Buddhist feel to it. This has enough of a satiric bite to it that I can imagine the author actually writing it.
Biss sounds like bliss, which as we all know is the result of ignorance. But Bissonomy is not the intellectual embrace of ignorance, although it might sound like it (haha). Rather, Bissonomy is the emotional acceptance/knowledge that our map will never fully match the territory, and that some parts of the territory are not mappable at all (eg uncertainties of quantum measurements).
The main reason to be skeptical of this idea, other than a severe lack of evidence for it (perhaps this allows it to be reconciled with faith), is that it fails to explain the virtue of Tubso, and it seems to me like understanding one might require also understanding the other.
However, I have a suspicion that the virtue of Tubso might be something absurd like “not throwing shellfish at the shadows of deities”, or perhaps Tubso is absurdism itself. The word certainly sounds silly enough for it. If we adopt the perspective that Bissonomy and Tubso are the “lost Buddhist” virtues, in my opinion this makes quite a lot of sense. A hint of absurdism paired with peaceful acceptance, just like a koan paired with its answer.
Another reason to be skeptical of my interpretation is that I ironically exhibit a lack of Bissonomy myself in so hastily jumping to the conclusion that it is possible to deduce what Bissonomy means, despite that Prachett provides little evidence around it. But perhaps this is simply another aspect of Prachett’s joke—forcing the reader to be hypocritical in order to understand the virtue of Bissonomy would be like a half-serious chastisement/reward for spoiling the fun of leaving the virtue unknown, similar to how trying very hard to find a true answer to a koan is a bad but necessary step in allowing the koan to change the way your mind thinks.
Ultimately, we may never know. And that’s okay. And that’s Bissonomy.
What do you think? Tell us in a week at most, please?
http://english.stackexchange.com/questions/116456/meaning-of-onomy-ology-and-ography
I’m not very familiar with the series. However, I suspect that the virtue of Bissonomy is the virtue of acknowledging that which you do not understand, also known as epistemic humility. (Faith is its sister virtue—don’t ask how they’re both able to be virtues despite their symmetrical nature, that would mortally offend both of them!) The vague and unspecified nature of Bissonomy is its own clue, there is a Buddhist feel to it. This has enough of a satiric bite to it that I can imagine the author actually writing it.
Biss sounds like bliss, which as we all know is the result of ignorance. But Bissonomy is not the intellectual embrace of ignorance, although it might sound like it (haha). Rather, Bissonomy is the emotional acceptance/knowledge that our map will never fully match the territory, and that some parts of the territory are not mappable at all (eg uncertainties of quantum measurements).
The main reason to be skeptical of this idea, other than a severe lack of evidence for it (perhaps this allows it to be reconciled with faith), is that it fails to explain the virtue of Tubso, and it seems to me like understanding one might require also understanding the other.
However, I have a suspicion that the virtue of Tubso might be something absurd like “not throwing shellfish at the shadows of deities”, or perhaps Tubso is absurdism itself. The word certainly sounds silly enough for it. If we adopt the perspective that Bissonomy and Tubso are the “lost Buddhist” virtues, in my opinion this makes quite a lot of sense. A hint of absurdism paired with peaceful acceptance, just like a koan paired with its answer.
Another reason to be skeptical of my interpretation is that I unfortunately exhibit a supreme lack of Bissonomy myself in so hastily jumping to the conclusion that it is possible to deduce what Bissonomy means, despite that Prachett provides little evidence around it. But perhaps this is simply another aspect of Prachett’s joke—forcing the reader to be hypocritical in order to understand the virtue of Bissonomy would be like a half-serious chastisement for spoiling the fun.
Ultimately, we may never know. And that’s okay. And that’s Bissonomy.
Does the ten year old child provide an actuarial model for superlongevity?
According to the actuarial tables:
http://www.ssa.gov/oact/STATS/table4c6.html
A ten year old boy has a probability of survival in that year of 0.999918. After that, his probability of survival in a given year decreases with every additional year.
If you could lock in the ten year old’s probability of survival per year after the age of 10, mathematically a population of such individuals would have a “half life” of ~ 8000 years. In other words, if you had a population of 1,000 such individuals with their annual probability of survival fixed at .999918, about half of them could survive for 8,000 years or so.
Of course this sort of calculation doesn’t mean anything without an empirical demonstration; the data have to come in over 8,000 years. That shows the problem with claims of life extension breaktrhoughs within current human life expectancies. You can’t measure the results any faster than the rate at which humans already happen to live, and unmodified humans can regularly live longer than other mammals any way.
A half life of ~8000 given current-day levels of accidental and other death. If we get our act together enough to get rid of the problem of aging, I would assume that we would continue to get rid of other sources of death as well, which would make the actuarial model less useful.
In the real world, the probability of a ten year old’s death also reflects the fact that in developed countries, children live in sheltered conditions.
According to the CDC, the leading causes for death for children aged 5-9 (in 2012 in the United States) are:
Unintentional injury
Malignant neoplasm (aka cancer)
Congenital disorders
Homicide
Heart disease
If we solved aging, it seems likely we could eliminate or significantly reduce deaths from cancer, congenital disorders and heart disease.
Once we look at the 10-14 age bracket or above, suicide makes it into the top five causes of death until age ~50 and above.
We can also look at the leading causes of unintentional injury. For the 5-9 age bracket, we have
Motor vehicle accidents
Drowning
Fire/Burns
Unintentional suffocation
Other land transport injuries
Traffic accidents seem likely to be solvable to a large degree with self-driving car technology. Not as sure about the others. It’s worth noting that the primary cause of unintentional injury deaths for adults is unintentional poisoning. This was surprising to me; I would guess it’s mostly due to drug use.
Food poisoning is more common and serious than most people I’ve talked about it with have expected.
“Unintentional poisoning” is 36k deaths, according to Adele’s link. Food poisoning is only 3k deaths (5k according to an older article). I think most of the accidental poisoning deaths are drug overdoses. Half (broken link) of those are legal drugs. 2⁄3 of the legal ones are opiates, 1⁄3 benzos.
I am not convinced that food poisoning is even included in accidental poisoning deaths. Wonder has deaths classified by ICD codes. About half of accidental poisonings are listed as X44 other/unknown. My guess is that this means an overdose on street drugs.
Less Wrong survey asked what is your favourite Less Wrong post. Slate Star Codex survey asks what is your favourite SSC post. Naturally, the next question is:
What are your favourite Overcomingbias posts? What posts did you find especially insightful or informative? What posts changed the way you think about something? What posts did you find thought-provoking even if you disagree with the ideas expressed in them (perhaps similarly to how Bryan Caplan thinks about Robin Hanson’s ideas)? What good posts do you think should be better known? What posts would you recommend to others?
http://www.overcomingbias.com/2009/11/bad-emulation-advance.html
http://www.overcomingbias.com/2007/05/policy_tugowar.html
http://www.overcomingbias.com/2009/12/tiptoe-or-dash-to-future.html
http://www.overcomingbias.com/2010/11/fertility-the-big-problem.html
http://www.overcomingbias.com/2010/01/enable-raiders.html
http://www.overcomingbias.com/2014/07/firm-inefficiency.html
http://www.overcomingbias.com/2008/07/refuge-markets.html
No general procedure for bug checks will do.
Now, I won’t just assert that, I’ll prove it to you.
I will prove that although you might work till you drop,
you cannot tell if computation will stop.
A poetic proof of the Halting Problem’s undecidability by one Geoffrey Pullum, continued here. Enjoy!
Ummm… are things like this OB post really supposed to be considered even remotely about Rationality? It’s basically just Robin Hanson using a few anecdotes to weave himself a short narrative designed to justify his misanthropy and make himself feel better about holding political positions others consider immoral.
Why are you asking that question?
Because it was linked under “Recent on Rationality Blogs”, I figured I’d give it a whirl, and was pretty surprised to find such low-grade reasoning on such a list.
Even a generally rational person may write an irrational article once in a time. (I am speaking generally here; I haven’t read the specific article yet.) To create a list containing only rational articles, someone would have to check each one of them individually, and vote.
Adding such functionality to LW software would be too much work. But maybe it could be done indirectly. We could do the voting on some other website (for example Reddit), and import only the positively-voted links to LW. But this would need a group of people to add new articles, read them, and vote.
“Once in a time” is quite an understatement of the frequency of completely preposterous posts on OB.
That’s a stupid argument.
There are multiple valid criticism of the way things are labeled but if you want to make that criticism it would be worthwhile to think a bit deeper about the structures and why they are arranged the way they are.
Tenure.
BTW, did anyone else get an audio track on that page? I don’t see any visible embedded media, but some sort of music track started a few seconds after loading the page, and stopped the moment I closed the window. Same thing when I reopened the page. I downloaded the HTML, but I didn’t see a cause of it there. A few minutes later it no longer happened, although the HTML of the page was identical to the previous download.
ETA: This may be the explanation. Some sort of advertising thing. The song described there is what I heard.
If my karma takes a hit is there an easy way that I can find out what is being downvoted? I can’t self correct if I don’t know what is disliked.
Self-correction on the basis of twitches of the hive mind is not necessarily a great idea.
Haha.
Yeah. I’ve been posting a lot recently and my overall karma is increasing, but my percentage of positive karma is steadily decreasing. I suppose that if I’m not upsetting some people, I’m not doing a good job, even among a community that I by and large respect.
Why?? I occasionally hear that repeated, but it sounds like a cheap excuse to act like a dick, or to retroactively brush off when people point out when you said something wrong in public. It calls to mind the image of a lazy teen spouting out every random stupid idea that goes through his mind and considering that the essence of being a Brave Independent Thinker.
(this is not targeted at you, Capla, I haven’t been paying special attention to your posts)
Because if you’re not upsetting some people, you are not impacting the status quo in any meaningful way.
What’s so great about impacting the status quo? That doesn’t seem like something worth aiming for. I mean, yeah, sure, most ways of making the world a better place impact the status quo; but most ways of making the world a better place involve making noise at one point of the other, that doesn’t mean that making noise is some great thing we should aim for.
Things that make the world (or lesswrong, or your family, etc.) a worse place are more likely to make people upset than things that make the world a better place. There are also more ways to make things worse than to make things better.
Depends on your value system, of course. But impacting the status quo is a synonym for “making a difference” and if you don’t ever make a difference, well...
Please continue. If you don’t ever make a difference, then what?
Then I’m not sure what your terminal values are beyond surviving.
I think there may be a communication failure here. While most desirable changes are themselves changes to the status quo, the phrase “changing the status quo” generally has the connotation of moving away from an undesirable state, instead of moving toward a desirable state.
For a concrete example, if I wanted to eradicate malaria, I would say “I want to eradicate malaria,” not “I want to impact the status quo” or “I want to make a difference,” even though both types of statements are true. The goal is to make a specific difference, not to make a difference.
Yes, looks like a communication problem. The phrase “changing the status quo” has no such connotation for me, at all. I would be happy to call eradicating malaria changing the status quo.
Maybe that is true in many cases, but even so, it still is a bad thing to optimize for. The outside view says that most of the time, having your percentage of positive karma steadily decreasing means the quality of your comments are getting worse. If you want to be controversial and still be taken seriously, you need to signal your competence in less controversial areas.
Impacting the status quo is a fine thing to optimize for. Negative karma is a stupid thing to optimize for.
However I believe that here we are not talking about optimizing, but rather about warning signs.
No, it’s not. It is much easier to “impact the status quo” by making things worse than by making things better.
Optimization implies that you picked a direction. Optimizing impact means optimizing not just magnitude, but magnitude of a specific sign.
I can hazard a guess at your downvotes. You’re flooding the forum with lots of short posts a large part of which can be answered by a few minutes with Google, and some fraction of the rest would disappear if only you were to think about the matter for a few minutes...
You can look at your recent posts/comments and see what looks negative.
Thanks.
Found it.
Is there a way to see what comments (not articles) I have downvoted? Or get a summary of how many downvotes and upvotes I’ve made?
On your profile page, click on “DISLIKED.”
[Cross-posted from So8res’s, recent guide to MIRI’s research]
Just thought add links to these other “guides”:
“Atoms of Neural computation”: List of promising research directions for neuro-inspired AI (IOW, tries to answer the question “Deep Learning is just regression, so what could we possibly do next?”)
“Physical Principles for Scalable Neural Recording”: List of promising research directions for developing tools to do live recording of the brain (a separate issue from connectomics).
Is this a case of multiple discovery?[1] And might something similar happen with AGI? Here are 4 projects who have concurrently developed very similar looking models:
(1) University of Toronto: Unifying Visual-Semantic Embeddings with Multimodal Neural Language Models
(2) Baidu/UCLA: Explain Images with Multimodal Recurrent Neural Networks
(3) Google: A Neural Image Caption Generator
(4) Stanford: Deep Visual-Semantic Alignments for Generating Image Descriptions
[1] The concept of multiple discovery is the hypothesis that most scientific discoveries and inventions are made independently and more or less simultaneously by multiple scientists and inventors.
How meaningful is the “independent” criterion given the heavy overlaps in works cited and what I imagine must be a fairly recent academic MRCA among all the researchers involved?
http://selenite.livejournal.com/282731.html
A discussion of constraining an AI by building in detailed contracts and obedience to laws against theft and criminal behavior.
I don’t think this is obviously correct (if nothing else, parts of a complex set of rules can interact unpredictably), but these are the tools that humans have developed for dealing with semi-malign natural intelligences, so we should at least take a look at them.
Laws are useful to prevent behavior where you know what kind of behavior you want to forbid. They are often possible to game and need judging to interpret what happens in edge cases.
Copying laws and contracts might not be the complete solution, but there’s a lot of stored experience in the legal system which could be worth studying.
We know some of the behavior we want to forbid—we definitely don’t want the AI to take our atoms. Well, almost definitely.
Would it be a good outcome if the F?AI made really good simulations of us in a well-done simulated environment so that it was possible to have a lot more people?
Some medical interventions probably include taking atoms. Even something as simply as disposing of bodily secretions could be problematic.
[Meta]
I noticed that this recent lw post showed up on the “recent comments” side-bar, but that it doesn’t show up in the list of “discussion” posts. Is this just me? Do other people have this show up in “discussion”? (Also, this is not the first time that I’ve noticed that there are posts in the side-bar that I can’t find in Discussion).
It’s in Main where the best written and most rationality-relevant articles go. Discussion is for lower quality and less relevant posts. I wonder if missing Main is common and whether something should be done about that.
It took me several months before I discovered that not everything in Main shows up when you click on “Main”—you need to first click on “Main” and then click on “New”. It took me even longer before I figured out that there’s an RSS feed for Main / New so I could find out when something gets added there.
For me, at least, anything in Main that wasn’t also promoted had almost no visibility for several months. Which is a pity because a lot of the best stuff goes there.
If it’s in Main, why does the article’s Karma Bubble look like it’s in Discussion? (i.e. it’s not a filled in with green).
Not everything in Main is promoted; in fact, most of it isn’t. Only promoted articles get the green bubbles.
Interesting
Is there any remotely feasible way for us to contain a superintelligence aside from us also becoming superintelligences?
That’s what MIRI are trying to work out, right?
Prevent it from becoming a super-intelligence in the first place. You can’t guarantee boxing a fully self-improving AI but if you run an AI on a Macintosh 2 you could probably keep it contained.
That doesn’t answer the question.
It does, the answer given is “no”.
You could put it in a box with no gatekeeper or other way to interact with the outside world. It would be completely pointless and probably unethical, but you could do it.
Isn’t this the uncomputable utilon question?
Just here to remind you to notice when you are confused. If you don’t, this ( https://www.youtube.com/watch?v=7BOWOMPUbvE ) WILL happen to you and everyone WILL laugh. And you wouldn’t want that.
I keep wondering when the cryonics community will attract the attention of female social justice warriors (SJW’s) because of the perception that wealthy white guys dominate cryonics, even though plenty of middle class men have signed up for cryopreservation as well by using life insurance as the funding mechanism. The stereotype should push the SJW’s buttons about inequality, patriarchy and the lack of diversity.
Instead these sorts of women have ignored cryonics so far instead of trying to meddle with it to transform it according to their standards of “social justice.” If anything, cryonics acts like “female Kryptonite.”
I’ve also noticed the absence of another sort of women, namely adventuresses. If people believe that lonely white guys with money sign up for cryonics, you would expect to see more young, reasonably attractive women showing up at public gatherings of cryonicists to try to find men they can exploit for a share of the wealth.
So what kind of tipping point in the public’s view of cryonics would have to happen to make SJW’s and adventuresses notice cryonics as a field for social activism or financial exploitation?
In general if your beliefs lead you to expect something that doesn’t happen, it’s time to question your beliefs. Reality isn’t wrong. Beliefs must pay rent.
I think you’d get replies if you didn’t pepper it so much with needless political tribal signaling. We get it; you read Steve Sailer.
The way you wrote the question is horrible, but here are some things to consider:
most people haven’t heard about cryonics;
most of those who heard about it believe it doesn’t work;
there are many things way more expensive than cryonics.
Wow. And a million times, wow.
One must really, really, very deeply hate women for that to be the expectation that comes to mind.
Search your favourite search engine for “hostile wife phenomenon”.
You don’t see many men in yoga classes either.
When it reaches the point that most of them have heard of it. I know we’ve had at least one SJW come by and rant about how talking about cryonics reflected white privilege.