If sperm whales were sapient, and had their own languages how recently would humans have noticed? We wouldn’t be able to hear much of their speech. Without the ability for advanced tool use and without agriculture, I think it would be rather hard for us to notice. I don’t think this would be discovered any earlier than the 20th century. Do we know that we aren’t in this situation?
Walker Vargas
Walker Vargas’s Shortform
Do they think it’s a hardware/cost issue? Or do they think that “true” intelligence is beyond our abilities?
This is also a plausible route for spreading awareness of AI safety issues to the left. The downside is that it might make AI safety a “leftest” issue if a conservative analogy is not introduced at the same time.
I think of it as deferring to future me vs. deferring to someone else.
Another consideration is how much money someone has to hand. If someone only make $1,000 a month, they may choice $25 shoes that will last a year over $100 shoes that will last 5 years. Essentially, it is the complimentary idea of economy of scale.
Personhood is a legal category and an assumed moral category that policies can point to. Usually, the rules being argued about are about the acceptability of killing something. The category is used differently depending on the moral framework, but it is usually assumed to point at the same objects. Therefore disagreements are interpreted as mistakes.
Personally, I have my doubts on there being an exact point in development that you can point to where a human becomes a person. If there is it might be weeks after birth.
If I remember right, it was in the context of there not being any universally compelling arguments. A paperclip maximizer would just ignore the tablet. It doesn’t care what the “right” thing is. Humans also probably don’t care about the cosmic tablet either. That sort of thing isn’t what “morality” is references. The argue is more of a trick to get people recognize that than a formal argument.
I think the point is that people try to point to things like God’s will in order to appear like they have a source of authority. Eliezer is trying to lead them to conclude that any such tablet being authoritative just by nature is absurd and only seems right because they expect the tablet to agree with them. Another method is asking why the tablet says what it does. Asking if God’s decrees are arbitrary or if there is a good reason, ask why not just follow those reasons.
While I see a lot of concern about the big one. I think the whole AI environment being unaligned is the more likely but not any better outcome. A society that is doing really well by some metrics that just happen to be the wrong ones. I thinking of idea of freedom of contract that was popular at the beginning of the 20th century and how hard it was to dig ourselves out of that hole.
Highly positive outcomes are assumed to be more particular and complex than highly bad outcomes. Another assumption I think is common is that a utility of a maximally good life is lower than the magnitude of the utility of a maximally bad life. Is there a life good enough that you would take a bet of a 50% chance of that life and a 50% chance of the worst life of torture?
I don’t think the fundamental ought works as a default position. Partly because there will always be a possibility of being wrong about what that fundamental ought is no matter how long it looks. So the real choice is about how sure it should be before it starts acting on it’s best known option.
The right side can’t be NULL, because that’d make the expect value of both actions NULL. To do meaningful math with these possibilities there has to be a way of comparing utilities across the scenarios.
No, if you are contributing to a preexisting discussion, there should be some older work you can cite. For example, you learned about the theory of path semantics from something that wasn’t written by you. Cite that source.
I don’t think that matrix is right. I think it describes a different scenario. Suppose an AI’s Utility function is defined referentially as being equal to some unknown function written on a letter on Mt. Everest. It also has a given utility function that it has little reason to think is correlated with the real one. Then it would be vary important to find out want that true function is. Than the expected value of any action would be NULL if that letter doesn’t exist.
But an AI that only assigns a probability that that scenario is the case might still have most of its expected value tied to following its current utility function. Well given some way of comparing them. Without that there’s no way to weigh up the choice.
I just had a thought. If Mary was presented with a red, a blue, and a green tile on a white background could she identify which was which without additional visual context clues like comparing them to her nails? If not, I would expect a p-zombie to have the same issue implying that that failure isn’t to do with consciousness.
Depending on who you are talking to for-profit corporations is a good analogy for what is meant by “misaligned”. You can then point out that those same organizations are likely to make AI with profit maximization in mind, and might skimp out on moral restraint in favor of being superhumanly good at PR.
Use that comparison with the wrong person and they’ll call you a communist.
I want to add that the AI probably does not know it is misaligned for a while.
This sounds similar to the replication crisis, in terms of the incentivization issues.
Under unification wouldn’t it make sense to consider ourselves to be every instance of our mind state? So there’s no fact of the matter of what your surroundings are like until they effect your mind state. Similarly, every past and future that is compatible with your current mind state happened and will happen respectively.
In the ending where humanity gets modified, some people commit suicide. The captain thinks it doesn’t make sense to choose complete erasure over modification.