No, I’m insisting that no realistic AGI at all is a Magic Genie which can be instructed in high-level English. If it were, all I would have to say is, “Do what I mean!” and Bob’s your uncle. But since that cannot happen without solving Natural Language Processing as a separate problem before constructing an AGI, the AGI
I was actually agreeing with you that NLP needs to be solved separately if you want to instruct it in English. The rhetoric about magic isn’t helpful.
agent has a utility function coded as program code in a programming language—which makes desirable behavior quite improbable.
I don’t see why that would follow, and in fact I argued against it.
knowing is quite different from caring.
I know.
What we could do in this domain is solve natural-language learning and processing separately from AGI, and then couple that to a well-worked-out infrastructure of normative uncertainty, and then, after making absolutely sure that the AI’s concept-learning via the hard-wired natural-language processing library matches the way human minds represent concepts computationally, use a large corpus of natural-language text to try to teach the AI what sort of things human beings want.
That’s not what I was saying. I was saying an AI with a motivation to understand .NL correctly would research whatever human value was relevant.
We have some idea of a safe goal function for the AGI (it’s essentially a longer-winded version of “Do what I mean, but taking the interests of all into account equally, and considering what I really meaneven under reflection as more knowledge and intelligence are added”), the question is how to actually program that
That’s kind of what I was saying.
If being devious to humans is instrumentally rational, an instrumentally rational AI agent will do it.
Non sequitur. In general, what is an instrumental goal will vary with final goals, and epistemic rationality is a matter of final goals. Omohundran drives are unusual in not having the property of varying with final goals.
I was actually agreeing with you that NLP needs to be solved separately if you want to instruct it in English. The rhetoric about magic isn’t helpful.
I don’t see why that would follow, and in fact I argued against it.
I know.
That’s not what I was saying. I was saying an AI with a motivation to understand .NL correctly would research whatever human value was relevant.
That’s kind of what I was saying.
Non sequitur. In general, what is an instrumental goal will vary with final goals, and epistemic rationality is a matter of final goals. Omohundran drives are unusual in not having the property of varying with final goals.