Yes. But it doesn’t matter because all (well, I guess, most) of the authors agree that AI Alignment is needed and understand deeply so because they have read the Sequences.
Many of these people have even read the same Sequences.
Maybe not all of them, and maybe a part is just trusting the community. But the community shares parts that are not pre-paradigmatic. Sure, the solution is open, unfortunately, very much so. And on that I agree. But when people outside of the community complain, that is very much like creationists coming to a biology class—except they can’t be kicked out so easily, because alignment is a niche science that suddenly made it into the spotlight.
The rest of this will be about how the community here does do the things you demand. Yudkowsky specifically. Beating the same points as you.
Be ready and willing to talk and listen, on levels so basic that without context they would seem condescending. “I know the basics, stop talking down to me” is a bad excuse when the basics are still not known.
Yes, that was the case when the Sequences were written and discussed ten years back.
Draw diagrams. Draw cartoons. Draw flowcharts with boxes and arrows. The cheesier, the more “obvious”, the better.
If you’re smart, you probably skip some steps when solving problems. That’s fine, but don’t skip writing them down! A skipped step will confuse somebody. Maybe that “somebody” needed to hear your idea.
Read The Sense of Style by Steven Pinker. You can skip chapter 6 and the appendices, but read the rest. Know the rules of “good writing”. Then make different tradeoffs, sacrificing beauty for clarity. Even when the result is “overwritten” or “repetitive”.
Link (pages that contain your baseline assumptions) liberally. When you see one linked, read it at least once.
You mean, as in the Sequence?
Use cheesy “A therefore B” formal-logic syllogisms, even if you’re not at “that low a level of abstraction”. Believe me, it still makes everything clearer.
Repeat your main points. Summarize your main points. Use section headers and bullet points and numbered lists. Color-highlight and number-subscript the same word when it’s used in different contexts (“apple1 is not apple2”).
Dito.
Do all of this to the point of self-parody. Then maybe, just maybe, someone will understand you.
Except all of that was done, and some people understood, but unfortunately, those people complaining now didn’t.
Luckily, Eliezer has written enough about it that you can just ask ChatGPT exactly that:
EY’s Views on Consciousness and Ethics
Model: GPT-4
User
summarise EY’s views on consciousness and on ethics
ChatGPT
EY, or Eliezer Yudkowsky, is a prominent artificial intelligence (AI) researcher and writer known for his work on rationality, AI alignment, and friendly AI. While I can’t provide an exhaustive account of his views, I can give a brief summary of some of his key ideas on consciousness and ethics.
Consciousness:
Reductionism: Yudkowsky approaches consciousness from a reductionist perspective, which means he believes that consciousness can be understood and explained in terms of simpler, lower-level components.
No “magic”: He argues against the idea that there is some sort of “magic” or inherently inexplicable aspect to consciousness. Instead, he sees it as a natural phenomenon that can be understood through scientific inquiry.
Importance of AI: Yudkowsky believes that understanding consciousness is crucial for the development of AI, particularly AI alignment, as it will help us build artificial intelligences that can comprehend and interact with human values and emotions.
Ethics:
Meta-ethics: Yudkowsky’s approach to ethics is rooted in meta-ethics, which focuses on understanding the underlying nature and structure of ethics itself.
Reflective equilibrium: He advocates for the concept of “reflective equilibrium,” in which we arrive at ethical conclusions by balancing and adjusting our intuitions, principles, and beliefs to achieve a coherent and stable ethical framework.
Coherent extrapolated volition (CEV): Yudkowsky has proposed the concept of CEV as a way to align AI systems with human values. CEV involves using AI to model the values that humanity would converge upon if given enough time and resources to reflect, discuss, and reason together.
Friendly AI: Yudkowsky emphasizes the importance of developing AI that is both intelligent and aligned with human values. He believes that ensuring AI safety and value alignment is a critical challenge for humanity.
In conclusion, Eliezer Yudkowsky’s views on consciousness emphasize reductionism and the importance of understanding it for AI development. His ethical approach is grounded in meta-ethics, with a focus on reflective equilibrium and coherent extrapolated volition for developing friendly AI.
Neither of those is an actual theory. Saying consciousness of is reductionistic, somehow, doesn’t tell you how it works, any more than saying it is emergent, somehow. Saying the true ethics are what you get out CEV also doesn’t specify anything, because there is no gears level specification of CEV.
To me, it seems you are engaging with the ChatGPT summary. You can find more about in Metaethics, e.g.
Eliezer Yudkowsky wrote a Sequence about metaethics, the Metaethics sequence, which Yudkowsky worried failed to convey his central point (this post by Luke tried to clarify); he approached the same problem again from a different angle in Highly Advanced Epistemology 101 for Beginners. From a standard philosophical standpoint, Yudkowsky’s philosophy is closest to Frank Jackson’s moral functionalism / analytic descriptivism; Yudkowsky could be loosely characterized as moral cognitivist—someone who believes moral sentences are either true or false—but not a moral realist—thus denying that moral sentences refer to facts about the world. Yudkowsky believes that moral cognition in any single human is at least potentially about a subject matter that is ‘logical’ in the sense that its semantics can be pinned down by axioms, and hence that moral cognition can bear truth-values; also that human beings both using similar words like “morality” can be talking about highly overlapping subject matter; but not that all possible minds would find the truths about this subject matter to be psychologically compelling.
That there is no gears level specification is exactly the problem he points out! We don’t know how specify human values and I think he makes a good case pointing that out—and that it is needed for alignment.
That there is no gears level specification is exactly the problem he points out! We don’t know how specify human values
And we don’t know that human values exist as a coherent object either. So his metaethics is “ethics is X” where X is undefined and possibly non existent.
I am not asking because I want to know. I was asking because I wanted you to think about those sequences,and what they are actually saying , and how clear they are. Which you didn’t.
Why would it matter what an individual commenter says about the clarity of the Sequences? I think a better measure would be what a large number of readers think about how clear they are. We could do a poll but I think there is already a measure: The votes. But these don’t measure clarity. More something like how useful people found them. And maybe that is a better measure? Another metric would be the increase in the number of readers while the Sequences were published. By that measure, esp. given the niche subject, they seem to be of excellent quality.
But just to check I read one high (130) and one low-vote (21) post from the Metaethics sequence and I think they are clear and readable.
I don’t think these are mutually exclusive? The Sequences are long and some of the posts were better than others. Also, what is considered “clear” can depend on one’s background. All authors have to make some assumptions about the audience’s knowledge. (E.g., at minimum, what language do they speak?) When Eliezer guessed wrong, or was read by those outside his intended audience, they might not be able to fill in the gaps and clarity suffers—for them, but not for everyone.
Yes. But it doesn’t matter because all (well, I guess, most) of the authors agree that AI Alignment is needed and understand deeply so because they have read the Sequences.
Maybe not all of them, and maybe a part is just trusting the community. But the community shares parts that are not pre-paradigmatic. Sure, the solution is open, unfortunately, very much so. And on that I agree. But when people outside of the community complain, that is very much like creationists coming to a biology class—except they can’t be kicked out so easily, because alignment is a niche science that suddenly made it into the spotlight.
The rest of this will be about how the community here does do the things you demand. Yudkowsky specifically. Beating the same points as you.
See: The Sequences. The original ones. Those published as a book. Nicely ordered and curated. There are reading groups about it.
Yes, that was the case when the Sequences were written and discussed ten years back.
You mean like
(from here)or
(from here)or this
(from here), oh, that’s not cheesy enough, how about (from here)or
(from here, or The Cartoon Guide to Löb’s Theorem)Or many other.
You mean Inferential Distance (list of 50 posts)?
You mean Human’s Guide to Words (a 25 post sequence from 2008)?
You mean, as in the Sequence?
You mean as on Arbital? Like the Introduction to Bayes’ Rule?
Dito.
Except all of that was done, and some people understood, but unfortunately, those people complaining now didn’t.
My point wasn’t “the community never does the above things”, it was “each member of the community should do these things, and more often”.
EDIT: You are correct that I should’ve mentioned more of the prior-art on this, though, especially Human’s Guide To Words.
Also, maybe LW should bring back “Sequences reruns” for that inferential distance post.
Can you summarise EY’s views on consciousness and on ethics?
Luckily, Eliezer has written enough about it that you can just ask ChatGPT exactly that:
EY’s Views on Consciousness and Ethics
Model: GPT-4
User summarise EY’s views on consciousness and on ethics
ChatGPT EY, or Eliezer Yudkowsky, is a prominent artificial intelligence (AI) researcher and writer known for his work on rationality, AI alignment, and friendly AI. While I can’t provide an exhaustive account of his views, I can give a brief summary of some of his key ideas on consciousness and ethics.
Consciousness:
Reductionism: Yudkowsky approaches consciousness from a reductionist perspective, which means he believes that consciousness can be understood and explained in terms of simpler, lower-level components. No “magic”: He argues against the idea that there is some sort of “magic” or inherently inexplicable aspect to consciousness. Instead, he sees it as a natural phenomenon that can be understood through scientific inquiry. Importance of AI: Yudkowsky believes that understanding consciousness is crucial for the development of AI, particularly AI alignment, as it will help us build artificial intelligences that can comprehend and interact with human values and emotions. Ethics:
Meta-ethics: Yudkowsky’s approach to ethics is rooted in meta-ethics, which focuses on understanding the underlying nature and structure of ethics itself. Reflective equilibrium: He advocates for the concept of “reflective equilibrium,” in which we arrive at ethical conclusions by balancing and adjusting our intuitions, principles, and beliefs to achieve a coherent and stable ethical framework. Coherent extrapolated volition (CEV): Yudkowsky has proposed the concept of CEV as a way to align AI systems with human values. CEV involves using AI to model the values that humanity would converge upon if given enough time and resources to reflect, discuss, and reason together. Friendly AI: Yudkowsky emphasizes the importance of developing AI that is both intelligent and aligned with human values. He believes that ensuring AI safety and value alignment is a critical challenge for humanity. In conclusion, Eliezer Yudkowsky’s views on consciousness emphasize reductionism and the importance of understanding it for AI development. His ethical approach is grounded in meta-ethics, with a focus on reflective equilibrium and coherent extrapolated volition for developing friendly AI.
Neither of those is an actual theory. Saying consciousness of is reductionistic, somehow, doesn’t tell you how it works, any more than saying it is emergent, somehow. Saying the true ethics are what you get out CEV also doesn’t specify anything, because there is no gears level specification of CEV.
To me, it seems you are engaging with the ChatGPT summary. You can find more about in Metaethics, e.g.
That there is no gears level specification is exactly the problem he points out! We don’t know how specify human values and I think he makes a good case pointing that out—and that it is needed for alignment.
And we don’t know that human values exist as a coherent object either. So his metaethics is “ethics is X” where X is undefined and possibly non existent.
He doesn’t say “ethics is X” and I disagree that this is a summary that advances the conversation.
For any value of X?
I am not asking because I want to know. I was asking because I wanted you to think about those sequences,and what they are actually saying , and how clear they are. Which you didn’t.
Why would it matter what an individual commenter says about the clarity of the Sequences? I think a better measure would be what a large number of readers think about how clear they are. We could do a poll but I think there is already a measure: The votes. But these don’t measure clarity. More something like how useful people found them. And maybe that is a better measure? Another metric would be the increase in the number of readers while the Sequences were published. By that measure, esp. given the niche subject, they seem to be of excellent quality.
But just to check I read one high (130) and one low-vote (21) post from the Metaethics sequence and I think they are clear and readable.
Yes, lots of people think the sequences are great. Lots of people also complain about EY’s lack of clarity. So something has to give.
The fact that it seems to be hugely difficult for even favourably inclined people to distill his arguments is evidence in favour of unclarity.
I don’t think these are mutually exclusive? The Sequences are long and some of the posts were better than others. Also, what is considered “clear” can depend on one’s background. All authors have to make some assumptions about the audience’s knowledge. (E.g., at minimum, what language do they speak?) When Eliezer guessed wrong, or was read by those outside his intended audience, they might not be able to fill in the gaps and clarity suffers—for them, but not for everyone.
I agree that it is evidence to that end.
Semirelated: is this referring to me specifically? If not, who/else?
I am sorry for this quip. It was more directed at the people on Twitter, but they are not a useful sample to react to.
Ah, ok!