A working intelligence must entangle the search for desirable outcomes with the search for attainable outcomes on pretty much every level.
“Entangle” how? Surely not by confusing or conflating the concepts…?
A prediction search process that only covers questions of fact, with no regard for questions of desirability (or undesirability) (in sum, questions of importance), will effectively be undirected. It will waste hours trying to figure out minutia about things that don’t matter.
What is a “prediction search process”…?
I think… explicitly distinguishing the conditionals from the conjunctives in this little look forward of mine is more work than I’m willing to do today.
Fair enough, but… well, look, you seem to be reading some very complicated things into my comment. All I want to know is:
What, actually, are you claiming? (Are you claiming anything at all? I assume so, since you asked for “refutations”.)
I listed three kinds of claims you might be making. Are you, indeed, making any of these three sorts of claims? If yes, which one(s)? And what are the claims exactly?
There’s no need to go off on any tangents about “prediction search process” or “undirected human intelligence” or anything like that. Really, I’m asking very straightforward questions here!
This is a paraphrasing of what intelligence is. If you can implement a search for useful predictions that generalizes over some relatively broad domain of things that can be predicted over, that’s AI-complete. That will be an AI. Is this not a common idea?
I am not conflating desirability with expectation. I will always speak of them in the same breath because they are entangled, not just for the technical reasons I expounded, but for deep decision-theoretic reasons that the field has only started to really get a grasp recently. There are many important situations where people/agents/crowds have to make a decision about what they will believe and what they want to be true simultaniously, because the beliefs/protocols/actions are a direct logical consequence of the desires. For instance, we attribute value to money because we want money to have value. If The Market comes to see some financial system as primarily parasitic or defective, and if they are good LDT agents, that system’s currency will not be accepted by The Market after that point. The truth- whether it will be valued- will change because there are situations in which it the truth is a direct consequence of desires.
Which is not especially relevant.
I’m sorry, I’ve already explained the intent of the post to you. You didn’t find the explanation satisfactorally reductive?.. I don’t really know how to go any further there. I’m not sure how to go about justifying like, a conversation style. I understand where you’re coming from. Personally I don’t find the way of engaging that you’re looking for to be fun or productive. You want me to drag you every step of the way.. That’s what it feels like, anyway. I can’t be the one to do that. I only have time to give a summary.
If that’s not interesting to people, if I haven’t motivated a deeper exploration or if it’s not evident to enough people that this would be a useful framework for discussion, well, okay. Maybe I have alienated people who know this area well enoug to confirm or refute, or maybe this isn’t the right medium for that.
While I think you are not wrong about the entanglement of intellectual exploration and truth-value, I do think you did not really explain the intent of the post. You only really said half a sentence about it, and that one did seem pretty weird to me:
...and instead of saying “it wont happen because X will happen, it’s pointless thinking about” we should just start talking about X and how to prevent it. We probably have enough time to do something.
This seems to indicate that your goal is to get the people on this site to start working towards the future you described, which seems like a valid and fine goal. However, at this point I am not particularly convinced that working on the future you described is tractable, something I can influence particularly much, or something I should care about. It sure sounds pretty cool, but there are a lot of visions for the future that sound pretty cool.
You want me to drag you every step of the way.. That’s what it feels like, anyway. I can’t be the one to do that. I only have time to give a summary.
Don’t have time?! You typed a way longer comment than you would have needed to type if you had just answered my questions!
I’m sorry, I’ve already explained the intent of the post to you. You didn’t find the explanation satisfactorally reductive?
You really didn’t explain it, though. You said “you might be misreading the intent of the post”… and then didn’t follow that up with a statement of what the intent of the post was.
Again, I’m asking very simple questions. You seem to be avoiding answering them. I’m not sure why. It seems like it would take you very little effort to do so, much less effort than making the comments you are making.
“Entangle” how? Surely not by confusing or conflating the concepts…?
What is a “prediction search process”…?
Fair enough, but… well, look, you seem to be reading some very complicated things into my comment. All I want to know is:
What, actually, are you claiming? (Are you claiming anything at all? I assume so, since you asked for “refutations”.)
I listed three kinds of claims you might be making. Are you, indeed, making any of these three sorts of claims? If yes, which one(s)? And what are the claims exactly?
There’s no need to go off on any tangents about “prediction search process” or “undirected human intelligence” or anything like that. Really, I’m asking very straightforward questions here!
This is a paraphrasing of what intelligence is. If you can implement a search for useful predictions that generalizes over some relatively broad domain of things that can be predicted over, that’s AI-complete. That will be an AI. Is this not a common idea?
I am not conflating desirability with expectation. I will always speak of them in the same breath because they are entangled, not just for the technical reasons I expounded, but for deep decision-theoretic reasons that the field has only started to really get a grasp recently. There are many important situations where people/agents/crowds have to make a decision about what they will believe and what they want to be true simultaniously, because the beliefs/protocols/actions are a direct logical consequence of the desires. For instance, we attribute value to money because we want money to have value. If The Market comes to see some financial system as primarily parasitic or defective, and if they are good LDT agents, that system’s currency will not be accepted by The Market after that point. The truth- whether it will be valued- will change because there are situations in which it the truth is a direct consequence of desires.
Which is not especially relevant.
I’m sorry, I’ve already explained the intent of the post to you. You didn’t find the explanation satisfactorally reductive?.. I don’t really know how to go any further there. I’m not sure how to go about justifying like, a conversation style. I understand where you’re coming from. Personally I don’t find the way of engaging that you’re looking for to be fun or productive. You want me to drag you every step of the way.. That’s what it feels like, anyway. I can’t be the one to do that. I only have time to give a summary.
If that’s not interesting to people, if I haven’t motivated a deeper exploration or if it’s not evident to enough people that this would be a useful framework for discussion, well, okay. Maybe I have alienated people who know this area well enoug to confirm or refute, or maybe this isn’t the right medium for that.
While I think you are not wrong about the entanglement of intellectual exploration and truth-value, I do think you did not really explain the intent of the post. You only really said half a sentence about it, and that one did seem pretty weird to me:
This seems to indicate that your goal is to get the people on this site to start working towards the future you described, which seems like a valid and fine goal. However, at this point I am not particularly convinced that working on the future you described is tractable, something I can influence particularly much, or something I should care about. It sure sounds pretty cool, but there are a lot of visions for the future that sound pretty cool.
Don’t have time?! You typed a way longer comment than you would have needed to type if you had just answered my questions!
You really didn’t explain it, though. You said “you might be misreading the intent of the post”… and then didn’t follow that up with a statement of what the intent of the post was.
Again, I’m asking very simple questions. You seem to be avoiding answering them. I’m not sure why. It seems like it would take you very little effort to do so, much less effort than making the comments you are making.