Long-time lurker (c. 2013), recent poster. I also write on the EA Forum.
Mo Putera
I unironically love Table 2.
A shower thought I once had, intuition-pumped by MIRI’s / Luke’s old post on turning philosophy to math to engineering, was that if metaethicists really were serious about resolving their disputes they should contract a software engineer (or something) to help implement on GitHub a metaethics version of Table 2, where rows would be moral dilemmas like the trolley problem and columns ethical theories, and then accept that real-world engineering solutions tend to be “dirty” and inelegant remixes plus kludgy optimisations to handle edge cases, but would clarify what the SOTA was and guide “metaethical innovation” much better, like a qualitative multi-criteria version of AI benchmarks.
I gave up on this shower thought for various reasons, including that I was obviously naive and hadn’t really engaged with the metaethical literature in any depth, but also because I ended up thinking that disagreements on doing good might run ~irreconcilably deep, plus noticing that Rethink Priorities had done the sophisticated v1 of a subset of what I had in mind and nobody really cared enough to change what they did. (In my more pessimistic moments I’d also invoke the diseased discipline accusation, but that may be unfair and outdated.)
Lee Billings’ book Five Billion Years of Solitude has the following poetic passage on deep time that’s stuck with me ever since I read it in Paul Gilster’s post:
Deep time is something that even geologists and their generalist peers, the earth and planetary scientists, can never fully grow accustomed to.
The sight of a fossilized form, perhaps the outline of a trilobite, a leaf, or a saurian footfall can still send a shiver through their bones, or excavate a trembling hollow in the chest that breath cannot fill. They can measure celestial motions and list Earth’s lithic annals, and they can map that arcane knowledge onto familiar scales, but the humblest do not pretend that minds summoned from and returned to dust in a century’s span can truly comprehend the solemn eons in their passage.
Instead, they must in a way learn to stand outside of time, to become momentarily eternal. Their world acquires dual, overlapping dimensions— one ephemeral and obvious, the other enduring and hidden in plain view. A planet becomes a vast machine, or an organism, pursuing some impenetrable purpose through its continental collisions and volcanic outpourings. A man becomes a protein-sheathed splash of ocean raised from rock to breathe the sky, an eater of sun whose atoms were forged on an anvil of stars.
Beholding the long evolutionary succession of Earthly empires that have come and gone, capped by a sliver of human existence that seems so easily shaved away, they perceive the breathtaking speed with which our species has stormed the world. Humanity’s ascent is a sudden explosion, kindled in some sapient spark of self-reflection, bursting forth from savannah and cave to blaze through the biosphere and scatter technological shrapnel across the planet, then the solar system, bound for parts unknown. From the giant leap of consciousness alongside some melting glacier, it proved only a small step to human footprints on the Moon.
The modern era, luminous and fleeting, flashes like lightning above the dark, abyssal eons of the abiding Earth. Immersed in a culture unaware of its own transience, students of geologic time see all this and wonder whether the human race will somehow abide, too.
(I still think it will.)
Your writeup makes me think you may be interested in Erik Hoel’s essay Enter the Supersensorium.
Nice reminiscence from Stephen Wolfram on his time with Richard Feynman:
Feynman loved doing physics. I think what he loved most was the process of it. Of calculating. Of figuring things out. It didn’t seem to matter to him so much if what came out was big and important. Or esoteric and weird. What mattered to him was the process of finding it. And he was often quite competitive about it.
Some scientists (myself probably included) are driven by the ambition to build grand intellectual edifices. I think Feynman — at least in the years I knew him — was much more driven by the pure pleasure of actually doing the science. He seemed to like best to spend his time figuring things out, and calculating. And he was a great calculator. All around perhaps the best human calculator there’s ever been.
Here’s a page from my files: quintessential Feynman. Calculating a Feynman diagram:
It’s kind of interesting to look at. His style was always very much the same. He always just used regular calculus and things. Essentially nineteenth-century mathematics. He never trusted much else. But wherever one could go with that, Feynman could go. Like no one else.
I always found it incredible. He would start with some problem, and fill up pages with calculations. And at the end of it, he would actually get the right answer! But he usually wasn’t satisfied with that. Once he’d gotten the answer, he’d go back and try to figure out why it was obvious. And often he’d come up with one of those classic Feynman straightforward-sounding explanations. And he’d never tell people about all the calculations behind it. Sometimes it was kind of a game for him: having people be flabbergasted by his seemingly instant physical intuition, not knowing that really it was based on some long, hard calculation he’d done.
Feynman and Wolfram had very different problem-solving styles:
Typically, Feynman would do some calculation. With me continually protesting that we should just go and use a computer. Eventually I’d do that. Then I’d get some results. And he’d get some results. And then we’d have an argument about whose intuition about the results was better.
The way he grappled with Wolfram’s rule 30 exemplified this (I’ve omitted a bunch of pictures, you can check them out in the article):
You know, I remember a time — it must have been the summer of 1985 — when I’d just discovered a thing called rule 30. That’s probably my own all-time favorite scientific discovery. And that’s what launched a lot of the whole new kind of science that I’ve spent 20 years building (and wrote about in my book A New Kind of Science). …
Well, Feynman and I were both visiting Boston, and we’d spent much of an afternoon talking about rule 30. About how it manages to go from that little black square at the top to make all this complicated stuff. And about what that means for physics and so on.
Well, we’d just been crawling around the floor — with help from some other people — trying to use meter rules to measure some feature of a giant printout of it. And Feynman took me aside, rather conspiratorially, and said, “Look, I just want to ask you one thing: how did you know rule 30 would do all this crazy stuff?” “You know me,” I said. “I didn’t. I just had a computer try all the possible rules. And I found it.” “Ah,” he said, “now I feel much better. I was worried you had some way to figure it out.”
Feynman and I talked a bunch more about rule 30. He really wanted to get an intuition for how it worked. He tried bashing it with all his usual tools. Like he tried to work out what the slope of the line between order and chaos is. And he calculated. Using all his usual calculus and so on. He and his son Carl even spent a bunch of time trying to crack rule 30 using a computer.
And one day he calls me and says, “OK, Wolfram, I can’t crack it. I think you’re on to something.” Which was very encouraging.
Scott’s The Colors Of Her Coat is the best writing I’ve read by him in a long while. Quoting this part in particular as a self-reminder and bulwark against the faux-sophisticated world-weariness I sometimes slip into:
Chesterton’s answer to the semantic apocalypse is to will yourself out of it. If you can’t enjoy My Neighbor Totoro after seeing too many Ghiblified photos, that’s a skill issue. Keep watching sunsets until each one becomes as beautiful as the first…
If you insist that anything too common, anything come by too cheaply, must be boring, then all the wonders of the Singularity cannot save you. You will grow weary of green wine and sick of crimson seas. But if you can bring yourself to really pay attention, to see old things for the first time, then you can combine the limitless variety of modernity with the awe of a peasant seeing an ultramarine mural—or the delight of a 2025er Ghiblifying photos for the first time.
How to see old things for the first time? I thought of the following passage by LoganStrohl describing a SIM card ejection tool:
I started studying “original seeing”, on purpose and by that name, in 2018. What stood out to me about my earliest exploratory experiments in original seeing is how alien the world is. …
I started my earliest experimentation with some brute-force phenomenology. I picked up an object, set it on the table in front of me, and progressively stripped away layers of perception as I observed it. It was one of these things:
I wrote, “It’s a SIM card ejection tool.”I wrote some things about its shape and color and so forth (it was round and metal, with a pointy bit on one end); and while I noted those perceptions, I tried to name some of the interpretations my mind seemed to be engaging in as I went.
As I identified the interpretations, I deliberately loosened my grip on them: “I notice that what I perceive as ‘shadows’ needn’t be places where the object blocks rays of light; the ‘object’ could be two-dimensional, drawn on a surface with the appropriate areas shaded around it.”
I noticed that I kept thinking in terms of what the object is for, so I loosened my grip on the utility of the object, mainly by naming many other possible uses. I imagined inserting the pointy part into soil to sow tiny snapdragon seeds, etching my name on a rock, and poking an air hole in the top of a plastic container so the liquid contents will pour out more smoothly. I’ve actually ended up keeping this SIM card tool on a keychain, not so I can eject SIM trays from phones, but because it’s a great stim; I can tap it like the tip of a pencil, but without leaving dots of graphite on my finger.
I loosened my grip on several preconceptions about how the object behaves, mainly by making and testing concrete predictions, some of which turned out to be wrong. For example, I expected it to taste sharp and “metallic”, but in fact I described the flavor of the surface as “calm, cool, perhaps lightly florid”.
By the time I’d had my fill of this proto-exercise, my relationship to the object had changed substantially. I wrote:
My perceptions that seem related to the object feel very distinct from whatever is out there impinging on my senses. … I was going to simply look at a SIM card tool, and now I want to wrap my soul around this little region of reality, a region that it feels disrespectful to call a ‘SIM card tool’. Why does it feel disrespectful? Because ‘SIM card tool’ is how I use it, and my mind is trained on the distance between how I relate to my perceptions of it, and what it is.
That last paragraph, and especially the use of ‘disrespectful’, strikes me a bit like the rationalist version of what Chesterton was talking about in Scott’s post.
Just signal-boosting the obvious references to the second: Sarah Constantin’s Humans Who Are Not Concentrating Are Not General Intelligences and Robin Hanson’s Better Babblers.
After eighteen years of being a professor, I’ve graded many student essays. And while I usually try to teach a deep structure of concepts, what the median student actually learns seems to mostly be a set of low order correlations. They know what words to use, which words tend to go together, which combinations tend to have positive associations, and so on. But if you ask an exam question where the deep structure answer differs from answer you’d guess looking at low order correlations, most students usually give the wrong answer.
Simple correlations also seem sufficient to capture most polite conversation talk, such as the weather is nice, how is your mother’s illness, and damn that other political party. Simple correlations are also most of what I see in inspirational TED talks, and when public intellectuals and talk show guests pontificate on topics they really don’t understand, such as quantum mechanics, consciousness, postmodernism, or the need always for more regulation everywhere. After all, media entertainers don’t need to understand deep structures any better than do their audiences.
Let me call styles of talking (or music, etc.) that rely mostly on low order correlations “babbling”. Babbling isn’t meaningless, but to ignorant audiences it often appears to be based on a deeper understanding than is actually the case. When done well, babbling can be entertaining, comforting, titillating, or exciting. It just isn’t usually a good place to learn deep insight.
It’s unclear to me how much economically-relevant activity is generated by low order correlation-type reasoning, or whatever the right generalisation of “babbling” is here.
This is great, thanks! Didn’t think of the model-prompting-model trick.
I don’t know either, but I think of Tracing Woodgrains’ Center for Educational Progress and the growing Discord community around it as a step in this direction.
Deep Research has already been commoditized, with Perplexity and xAI launching their own versions almost immediately.
Deep Research is also not a good product. As I covered last week, the quality of writing that you receive from a Deep Research report is terrible, rivaled only by the appalling quality of its citations, which include forum posts and Search Engine Optimized content instead of actual news sources. These reports are neither “deep” nor well researched, and cost OpenAI a great deal of money to deliver.
Good homework by Zitron on the numbers, and he’s a really entertaining writer, but my (very brief) experience so far using it for work-related research more closely matches Sarah Constantin’s assessment concluding that ChatGPT-4o DR was the best one she tested (including Perplexity, Gemini, ChatGPT-4o, Elicit, and PaperQA) on completeness, relevance, source quality, and creativity.
Thurston’s case reminds me somwhat of this old LW comment by pdf23ds:
Take me for an example. On the WISC-III IQ test, my combined score is 145. There are two composite scores that the combined score is made up of, the verbal score (I got 155, the maximum possible on that test) and the performance score (I got 125). There are also a number of different individual capacity scores. On most, I scored above the 95 percentile. On two or three, I scored right in the middle, and in one (visual short term memory) I scored in the first percentile.
Let me repeat that. I scored in the first percentile for the capacity to keep visual information in my short-term memory. (I scored in the 97th for aural short term memory, and 99.9th for linguistic.) How does that change how I solve problems, how I think about the world? Well, I perform many tasks about twice as slowly (but just as accurately) as others with my composite IQ. I have to use other circuits than most people do to solve the same problems, circuits that aren’t as efficient. Circuits that may even work slightly differently, giving me a different perspective on problems, which may be superior or inferior, I don’t know (likely depending on the individual problem). I strongly suspect that this is a large part of the cause of my intense dislike of school.
At the individual level, I can’t see myself ever choosing for my child to have >99.9th percentile linguistic ability and 1st(!!) percentile visual short-term memory, or really any such spectacularly uneven combination of abilities. (I’m not as extreme, but I remember this quote because I empathise with it: I’m high-math low-verbal, my childhood was the mirror of Scott’s, right down to “I don’t know which bothered me more, the praise from effortless success or the criticism from backbreaking toil to get Bs on my own native language’s exams”.)
At the societal level however, there does seem to be a lot of benefit to a cognitive diversity of minds (I’m thinking of Cosma Shalizi and Henry Farrell’s cognitive democracy essay, and their referencing Lu Hong and Scott Page (2004)’s use of mathematical models to argue that “diversity of viewpoints helps groups find better solutions”). So I guess one direction this line of thinking could go is how we can get the society-level benefits of a cognitive diversity of minds without necessarily having cognitively-uneven kids grow up in pain.
Interesting anecdote on “von Neumann’s onion” and his general style, from P. R. Halmos’ The Legend of John von Neumann:
Style. As a writer of mathematics von Neumann was clear, but not clean; he was powerful but not elegant. He seemed to love fussy detail, needless repetition, and notation so explicit as to be confusing. To maintain a logically valid but perfectly transparent and unimportant distinction, in one paper he introduced an extension of the usual functional notation: along with the standard φ(x) he dealt also with something denoted by φ((x)). The hair that was split to get there had to be split again a little later, and there was φ(((x))), and, ultimately, φ((((x)))). Equations such as
(φ((((a))))^2 = φ(((a))))
have to be peeled before they can be digested; some irreverent students referred to this paper as von Neumann’s onion.
Perhaps one reason for von Neumann’s attention to detail was that he found it quicker to hack through the underbrush himself than to trace references and see what others had done. The result was that sometimes he appeared ignorant of the standard literature. If he needed facts, well-known facts, from Lebesgue integration theory, he waded in, defined the basic notions, and developed the theory to the point where he could use it. If, in a later paper, he needed integration theory again, he would go back to the beginning and do the same thing again. He saw nothing wrong with long strings of suffixes, and subscripts on subscripts; his papers abound in avoidable algebraic computations. The reason, probably, is that he saw the large picture; the trees did not conceal the forest from him. He saw and he relished all parts of the mathematics he was thinking about. He never wrote “down” to an audience; he told it as he saw it. The practice caused no harm; the main result was that, quite a few times, it gave lesser men an opportunity to publish “improvements” of von Neumann.
(tangent: I’m a bit peeved by Halmos’ “lesser men” throwaway remark, mainly because I think interpretive research labor and distillation is very valuable, very hard to do well, somewhat orthogonal to vN-style competence, and very underappreciated and undersupplied.)
von Neumann was also courageous, Halmos wrote, in the following way:
Another notable and enviable trait of von Neumann’s was his mathematical courage. If, in the middle of a search for a counterexample, an infinite series came up, with a lot of exponentials that had quadratic exponents, many mathematicians would start with a clean sheet of paper and look for another counterexample. Not Johnny! When that happened to him, he cheerfully said: “Oh, yes, a theta function...’’, and plowed ahead with the mountainous computations. He wasn’t afraid of anything.
Terry Tao is similar, according to Allen Knutson:
More specifically, one thing I learned from Terry that I was not taught in school is the importance of bad proofs. I would say “I think this is true”, work on it, see that there was no nice proof, and give up. Terry would say “Here’s a criterion that eliminates most of the problem. Then in what’s left, here’s a worse one that handles most of the detritus. One or two more epicycles. At that point it comes down to fourteen cases, and I checked them.” Yuck. But we would know it was true, and we would move on. (Usually these would get cleaned up a fair bit before publication.) …
Sometimes we’d really be on the same page, at the same letter of the same word even; one extreme case was when I needed to read his computer code and found it as easy to do as if I’d written it myself. But more often we’d bring different strengths. Since we were working in my field of expertise rather than his, I knew better what the interesting questions were, and could translate them into combinatorics, then sic Terry on them. He would beat them to a bloody death as described above, and then it would be my job to dress the carcass for public viewing back in the original field.
von Neumann also had endless capacity for work. Halmos:
Work habits. Von Neumann was not satisfied with seeing things quickly and clearly; he also worked very hard. His wife said “’he had always done his writing at home during the night or at dawn. His capacity for work was practically unlimited.” In addition to his work at home, he worked hard at his office. He arrived early, he stayed late, and he never wasted any time. He was systematic in both large things and small; he was, for instance, a meticulous proofreader. He would correct a manuscript, record on the first page the page numbers where he found errors, and, by appropriate tallies, record the number of errors that he had marked on each of those pages. Another example: when requested to prepare an abstract of not more than 200 words, he would not be satisfied with a statistical check — there are roughly 20 lines with about 10 words each — but he would count every word.
I thought this was striking: why waste time on such seeming trivialities? But I guess if you’re John von Neumann you just have such a glut of brain cycles that you can spend it in ridiculously poorly-optimised ways like this instead of needing to 80⁄20 and still get your many, many jobs done.
I think I failed to implicitly convey that I meant all this in jest, that I get a lot of personal enjoyment value out of silly poetry constrained by artificial rules, and that I was guessing at least someone else on the forum would share this enjoyment. I do like your scaffolding idea, might just try it out.
Pilish is a constrained writing style where the number of letters in consecutive words match the digits of pi. The canonical intro-to-Pilish sentence is “How I need a drink, alcoholic of course, after the heavy lectures involving quantum mechanics!”; my favorite Pilish poetry is Mike Keith’s Near a Raven, a retelling of Edgar Allan Poe’s “The Raven” stretching to 740 digits of pi (nowhere near Keith’s longest, that would be the 10,000-word world record-setting Not a Wake), which begins delightfully like so:
Poe, E.
Near a RavenMidnights so dreary, tired and weary.
Silently pondering volumes extolling all by-now obsolete lore.
During my rather long nap—the weirdest tap!
An ominous vibrating sound disturbing my chamber’s antedoor.
“This”, I whispered quietly, “I ignore”. …I wondered if today’s leading LLMs could write interesting Pilish poems. The answer is “emphatically not”.
Here’s the start of Claude Sonnet 3.7′s attempt. It doesn’t even realise how wrong it is:
Celestial Calculations
How I perceive mathematical constants elegantly describing universe’s fundamental architecture! Physicists, mathematicians repeatedly discover extraordinary connections manifesting mathematical elegance miraculously embedded. …
Sonnet confidently and verbosely remarks that “Each word’s length follows the digits of π (3.14159...), creating a hidden mathematical structure beneath the text. The poem spans approximately 100 digits of π, which is modest compared to the example you shared but still captures the essence of the constraint. The theme connects well with your physics background, …” and more such nonsense.
ChatGPT 4.5 got the first 3 words right but then quickly derailed:
Here’s a unique and engaging Pilish poem capturing the spirit of adventure and mystery:
Path Beyond Stars
Sky
A glow,
Far beyond gaze—
Stars dazzle, whisper softly.
Visions divine pursue travelers astray. … [skipping 4 stanzas]Pilish Constraint: The number of letters in each successive word follows the digits of π (3.14159265358979323846...). For instance, “Sky” (3 letters), “A” (1), “glow” (4), “Far” (3), and so forth.
Gemini 2.5 Pro, America’s next top large language model, was the only one that realised it kept derailing and would redo again and again. It ended up almost getting it, but then dramatically derailed at the end. Full response this time:
Title: Quiet Night
See a lone I watch clouds drifting, so silent above. Now stars emerge, bright gems upon night’s velvet cloak. Still, quietude surrounds all the dark world; wonders invoke peaceful thought.
Explanation:
See (3)
a (1)
lone (4)
I (1)
watch (5)
clouds (6) - Correction: clouds has 6 letters, Pi’s 6th digit after decimal is 9. Let’s adjust.
Okay, let’s try again. The single-letter words (like ‘a’ and ‘I’) corresponding to the ’1’s in Pi are often the trickiest parts.
Revised Title: Starlight
Now I look, a pale moon shimmers brightly, softly aloft. White diamonds glitter, placing subtle spells nightly; beauty ascends.
Let’s check the counts:
Now (3)
I (1)
look (4)
a (1)
pale (5)
moon (4) - Correction: moon has 4 letters, Pi’s 6th digit is 9. Need a 9-letter word.
This is proving challenging! Let’s try focusing on a simpler theme, perhaps less descriptive.
Third Attempt: Simple Task
Ask a good, I might consider carefully, making calculations matter. Maybe numbers intrigue gifted folks solving riddles.
Checking Counts:
Ask (3)
a (1)
good (4)
I (1)
might (5)
consider (9)
carefully (9) - Correction: carefully has 9 letters, Pi’s 7th digit is 2. Need a 2-letter word.
Alright, one more try, aiming for accuracy over length for now.
Fourth Attempt: A Cat’s View
How a soft, white creature lounges lazily, ignores barking sounds. Just watches birds flitting outside—dreams peacefully...
Checking Counts:
How (3)
a (1)
soft (4)
white (5) - Correction: white has 5 letters, Pi’s 5th digit is 5. Correct! Wait, Pi is 3.14159… so the 4th word needs to be 5 letters. My previous attempts were off by one position. Let’s re-check “Starlight”.
Re-checking “Starlight”:
Now (3) - Correct
I (1) - Correct
look (4) - Correct
a (1) - Correct
pale (5) - Correct
moon (4) - Needs to be 9 letters. Let’s replace “pale moon shimmers”.
How about:
Now I look a faint celestial...
faint (5) - Correct
celestial (9) - Correct!
Next digit is 2. Need a 2-letter word.
...celestial is...
is (2) - Correct!
Next digit is 6. Need a 6-letter word.
...celestial is bright...
bright (6) - Nope, 5 letters. How about
...celestial is divine...
? Divine is 6 letters. Correct!
Next digit is 5. Need a 5-letter word.
...celestial is divine light...
light (5) - Correct!
Next digit is 3. Need a 3-letter word.
...divine light and...
and (3) - Correct!
Next digit is 5. Need a 5-letter word.
...divine light and peace...
peace (5) - Correct!
Okay, let’s assemble this improved version:
Final Version: Starlight
Now I look a faint celestial is divine light and peace shines...
Checking Counts (Pi = 3.1415926535...)
Now (3) - Match
I (1) - Match
look (4) - Match
a (1) - Match
faint (5) - Match
celestial (9) - Match
is (2) - Match
divine (6) - Match
light (5) - Match
and (3) - Match
peace (5) - Match
shines (6) - Correction: shines has 6 letters, the next digit is 8. Need an 8-letter word.
Let’s replace “peace shines”. Next digits are 8, 9, 7, 9…
...light and faraway...
(faraway = 8) - Correct!...light and faraway silently...
(silently = 9) - Correct!...light and faraway silently viewing...
(viewing = 7) - Correct!...light and faraway silently viewing universes.
(universes = 9) - Correct!Final Poem:
Title: Starlight Gaze
Now I look a faint celestial is divine light and faraway silently viewing universes.
This poem follows the first 16 digits of Pi (including the initial 3). Pilish is definitely a tricky form!
Gemini’s poem doesn’t, in fact, follow the first 16 digits of pi, starting from when it inexplicably replaced ‘peace’ with ‘faraway’.
I hereby propose a new AI progress benchmark: “Pilish poem length”, or PPL.
I play Go (an Asian board game you can think of as being like chess) at a local club in the city. The rules of hierarchy there are something like “if Joe is much better than Gary, Joe can say Gary’s moves are wrong whenever he wants, and is allowed to launch into a short speech about why Gary’s mistake is such a major mistake.” (There are more cultural norms than that, and Go players reading this will find that a poor description, but for those outside of the culture, it’s close enough). On the other hand, when a weaker player thinks a stronger player’s move is weird, they’ll be very humble in their asking why the move is good. This is the opposite of the situation in corporate!
This is exactly the situation in (Asian) corporate, which is where I spent the first few years of my career at three different employers. In fact, at my first employer (a giant MNC headquartered in my country) juniors were expected to essentially shut up and do as we were told; the mere prospect of clarifying what a very senior person meant was terrifying for most. And at least in the Asian corporate contexts where I worked, what AnthonyC said is correct:
… the fact that you asked verbally is also in some sense a status claim, that you felt empowered to ask a question in a way that commanded everyone’s attention and interrupted the CEO’s talk track. It’s a power move, or can be seen as such.
Given this, when I read what you said here
Back to asking tough questions of the CEO in a giant meeting: he is so powerful—so above me—that I know I can’t threaten him, he knows I can’t threaten him, everybody watching knows I can’t threaten him, and everybody knows that everybody knows this. Boldness has little to do with me asking him a hard question—if you view hierarchy and position how I do, it’s one of the least dangerous group situations in which to ask such a question!
I thought “but he, the CEO, can definitely threaten you. Is the OP not worried at all about this?”
At my first employer, my boss’s boss (who everyone in the entire department loved) was fired by his boss (who everyone hated) for repeatedly questioning the latter’s decisions (which everyone agreed weren’t good) in meetings, in front of others. This happened a month into my tenure, and was quite the introduction to corporate not being college. I have similar stories from my second and third employer, if not quite as extreme.
I’m guessing this ultimately boils down to corporate culture differences.
There’s an argument I’ve seen a number of times on the internet about the failings of consequentialism as a moral system. … The argument goes roughly like so: Consequentialism tells us that the thing to do is the thing with the best results. But, this is a ridiculously high standard, that nobody can actually live up to. Thus, consequentialism tells us that everybody is bad, and we should all condemn everybody and all feel guilty.
I like Scott Alexander’s response to this kind of argument, from his Economic Perspective on Moral Standards:
But I find the “good person”/”not a good person” dichotomy helpful. I’m not claiming it objectively exists. I can’t prove anything about ethics objectively exists. And even if there were objective ethical truths about what was right or wrong, that wouldn’t imply that there was an objective ethical truth about how much of the right stuff you have to do before you can go around calling yourself “good”. In the axiology/morality/law trichotomy, I think of “how much do I have to do in order to be a good person” as within the domain of morality. That means it’s a social engineering question, not a philosophical one. The social engineering perspective assumes that “good person” status is an incentive that can be used to make people behave better, and asks how high vs. low the bar should be set to maximize its effectiveness.
Consider the way companies set targets for their employees. At good companies, goals are ambitious but achievable. If the CEO of a small vacuum company tells her top salesman to sell a billion vacuums a year, this doesn’t motivate the salesman to try extra hard. It’s just the equivalent of not setting a goal at all, since he’ll fail at the goal no matter what. If the CEO says “Sell the most vacuums you can, and however many you sell, I will yell at you for not selling more”, this also probably isn’t going to win any leadership awards. A good CEO might ask a salesman to sell 10% more vacuums than he did last year, and offer a big bonus if he can accomplish it. Or she might say that the top 20% of salesmen will get promotions, or that the bottom 20% of salesmen will be fired, or something like that. The point is that the goal should effectively carve out two categories, “good salesman” and “bad salesman”, such that it’s plausible for any given salesman to end up in either, then offer an incentive that makes him want to fall in the first rather than the second.
I think of society setting the targets for “good person” a lot like a CEO setting the targets for “good vacuum salesman”. If they’re attainable and linked to incentives – like praise, honor, and the right to feel proud of yourself – then they’ll make people put in an extra effort so they can end up in the “good person” category. If they’re totally unattainable and nobody can ever be a good person no matter how hard they try, then nobody will bother trying. This doesn’t mean nobody will be good – some people are naturally good without hope for reward, just like some people will slave away for the vacuum company even when they’re underpaid and underappreciated. It just means you’ll lose the extra effort you would get from having a good incentive structure.
So what is the right level at which to set the bar for “good person”? An economist might think of this question as a price-setting problem: society is selling the product “moral respectability” and trying to decide how many units effort to demand from potential buyers in order to maximize revenue. Set the price too low, and you lose out on money that people would have been willing to pay. Set the price too high, and you won’t get any customers. Solve for the situation where you have a monopoly on the good and the marginal cost of production is zero, and this is how you set the “good person” bar.
(I really like how gears-y your comment is, many thanks and strong-upvoted.)
This is helpful, thanks. Bummer though...
Claude.ai has web search! Woo-hoo! You have to enable it in the settings.
It mystifies me that as a Pro user my feature settings don’t include the web search option, only the analysis tool. I wonder if it’s a geographic location thing (I’m in Southeast Asia).
I like the optimal forager take, seems intuitively correct. I’d add that Dwarkesh struck gold by getting you on his podcast too. (Tangentially: this grand theory of intelligence video snippet reminds me of a page-ish-long writeup on that I stumbled upon deep in the bowels of https://gwern.net/ which I’ve annoyingly never been able to find again.)
Also thanks for the pointer to Werbos, his website Welcome to the Werbos World! funnily enough struck me as crackpot-y and I wouldn’t have guessed just from the landing page that he’s the discoverer of backprop, respected former program director at the NSF, etc.
Scott’s own reaction to / improvement upon Graham’s hierarchy of disagreement (which I just noticed you commented on back in the day, so I guess this is more for others’ curiosity) is
Seems like yours and Scott’s are complementary: I read you as suggesting how to improve one’s own argumentation techniques, while Scott is being more sociologically descriptive, mainly in explaining why online discourse so often degenerates into social shaming and meta-debate.