This is a paraphrasing of what intelligence is. If you can implement a search for useful predictions that generalizes over some relatively broad domain of things that can be predicted over, that’s AI-complete. That will be an AI. Is this not a common idea?
I am not conflating desirability with expectation. I will always speak of them in the same breath because they are entangled, not just for the technical reasons I expounded, but for deep decision-theoretic reasons that the field has only started to really get a grasp recently. There are many important situations where people/agents/crowds have to make a decision about what they will believe and what they want to be true simultaniously, because the beliefs/protocols/actions are a direct logical consequence of the desires. For instance, we attribute value to money because we want money to have value. If The Market comes to see some financial system as primarily parasitic or defective, and if they are good LDT agents, that system’s currency will not be accepted by The Market after that point. The truth- whether it will be valued- will change because there are situations in which it the truth is a direct consequence of desires.
Which is not especially relevant.
I’m sorry, I’ve already explained the intent of the post to you. You didn’t find the explanation satisfactorally reductive?.. I don’t really know how to go any further there. I’m not sure how to go about justifying like, a conversation style. I understand where you’re coming from. Personally I don’t find the way of engaging that you’re looking for to be fun or productive. You want me to drag you every step of the way.. That’s what it feels like, anyway. I can’t be the one to do that. I only have time to give a summary.
If that’s not interesting to people, if I haven’t motivated a deeper exploration or if it’s not evident to enough people that this would be a useful framework for discussion, well, okay. Maybe I have alienated people who know this area well enoug to confirm or refute, or maybe this isn’t the right medium for that.
While I think you are not wrong about the entanglement of intellectual exploration and truth-value, I do think you did not really explain the intent of the post. You only really said half a sentence about it, and that one did seem pretty weird to me:
...and instead of saying “it wont happen because X will happen, it’s pointless thinking about” we should just start talking about X and how to prevent it. We probably have enough time to do something.
This seems to indicate that your goal is to get the people on this site to start working towards the future you described, which seems like a valid and fine goal. However, at this point I am not particularly convinced that working on the future you described is tractable, something I can influence particularly much, or something I should care about. It sure sounds pretty cool, but there are a lot of visions for the future that sound pretty cool.
You want me to drag you every step of the way.. That’s what it feels like, anyway. I can’t be the one to do that. I only have time to give a summary.
Don’t have time?! You typed a way longer comment than you would have needed to type if you had just answered my questions!
I’m sorry, I’ve already explained the intent of the post to you. You didn’t find the explanation satisfactorally reductive?
You really didn’t explain it, though. You said “you might be misreading the intent of the post”… and then didn’t follow that up with a statement of what the intent of the post was.
Again, I’m asking very simple questions. You seem to be avoiding answering them. I’m not sure why. It seems like it would take you very little effort to do so, much less effort than making the comments you are making.
This is a paraphrasing of what intelligence is. If you can implement a search for useful predictions that generalizes over some relatively broad domain of things that can be predicted over, that’s AI-complete. That will be an AI. Is this not a common idea?
I am not conflating desirability with expectation. I will always speak of them in the same breath because they are entangled, not just for the technical reasons I expounded, but for deep decision-theoretic reasons that the field has only started to really get a grasp recently. There are many important situations where people/agents/crowds have to make a decision about what they will believe and what they want to be true simultaniously, because the beliefs/protocols/actions are a direct logical consequence of the desires. For instance, we attribute value to money because we want money to have value. If The Market comes to see some financial system as primarily parasitic or defective, and if they are good LDT agents, that system’s currency will not be accepted by The Market after that point. The truth- whether it will be valued- will change because there are situations in which it the truth is a direct consequence of desires.
Which is not especially relevant.
I’m sorry, I’ve already explained the intent of the post to you. You didn’t find the explanation satisfactorally reductive?.. I don’t really know how to go any further there. I’m not sure how to go about justifying like, a conversation style. I understand where you’re coming from. Personally I don’t find the way of engaging that you’re looking for to be fun or productive. You want me to drag you every step of the way.. That’s what it feels like, anyway. I can’t be the one to do that. I only have time to give a summary.
If that’s not interesting to people, if I haven’t motivated a deeper exploration or if it’s not evident to enough people that this would be a useful framework for discussion, well, okay. Maybe I have alienated people who know this area well enoug to confirm or refute, or maybe this isn’t the right medium for that.
While I think you are not wrong about the entanglement of intellectual exploration and truth-value, I do think you did not really explain the intent of the post. You only really said half a sentence about it, and that one did seem pretty weird to me:
This seems to indicate that your goal is to get the people on this site to start working towards the future you described, which seems like a valid and fine goal. However, at this point I am not particularly convinced that working on the future you described is tractable, something I can influence particularly much, or something I should care about. It sure sounds pretty cool, but there are a lot of visions for the future that sound pretty cool.
Don’t have time?! You typed a way longer comment than you would have needed to type if you had just answered my questions!
You really didn’t explain it, though. You said “you might be misreading the intent of the post”… and then didn’t follow that up with a statement of what the intent of the post was.
Again, I’m asking very simple questions. You seem to be avoiding answering them. I’m not sure why. It seems like it would take you very little effort to do so, much less effort than making the comments you are making.