How do you know that in the future we wont be able to use advanced computer modeling to predict many parts of a brains and bodies function based on what we have frozen and correct it with advanced nanotechnology?
That does seem to rather violate the principle of information decay. It is not possible to derive exact information from a decayed state, for the same reason that you cannot determine which exact formula has a derivative of “3”.
It is not possible to derive exact information from a decayed state
That’s true in the most general situation, when there is no prior information available. But a brain is not a random chunk of matter, it’s a highly particular one, with certain patterns and regularities. So it’s not implausible that a superintelligence could restore even a moderately damaged brain.
For a real example, think of image restoration of natural scenes. A photograph is not a random matrix of pixels, it belongs to a very small subset of all possible images, and that knowledge allows seemingly “impossible” tasks of focusing, enhacement and all that.
So it’s not implausible that a superintelligence could restore even a moderately damaged brain.
Intelligence is not magic. Information that no longer exists cannot be reinvented with fidelity.
For a real example, think of image restoration of natural scenes. A photograph is not a random matrix of pixels, it belongs to a very small subset of all possible images, and that knowledge allows seemingly “impossible” tasks of focusing, enhacement and all that.
And yet still-shots are limited to their original resolution; anything further is artistic rendering and not a valid reconstruction of the original. It is possible to “enhance” the resolution of a still-shot of a video feed of a person’s face. It is not possible to “enhance” the resolution of a single still-shot of a person’s face.
Memories are tricky things. We do not, now, know exactly how high the fidelity needs to be to sustain a person’s “actual” psyche. If information-theoretically significant portions are missing, no amount of genius can resolve that.
For the same reason you cannot extrapolate from the number 3 to the function f(x) whose derivative of x is “3”.
You can integrate it with initial conditions though and just like we can use our prefrontal cortex to predict the probable initial conditions of events(albeit inaccurately occasionally), a powerful computer may be able to predict our complex mental pathways based on known past events with high fidelity. I’m not saying that you wont need the initial conditions to integrate the function, I just think AGI would have less trouble with it than you assume. I think you have a good point about the principle though and I will take informative decay into my perceived utility of cryonics in the future.
a powerful computer may be able to predict our complex mental pathways based on known past events with high fidelity
“known past events”—unless those past events are full-brain scans of the past, then all you’re going to get is a reduction of the scope of the configuration space and not the exact function.
“A powerful computer” != “magic”. No matter how smart you are, fifteen tons of mass moving at five thousand miles per second will still contain the same amount of kinetic energy. No amount of cleverness can extract information that has been decayed.
This is information-theoretically proven.
I just think AGI would have less trouble with it than you assume.
The question at hand is, can a personality be reconstructed from partial data by a sufficiently clever process? We have analogues to this question. Compress and decompress the same mp3 file a hundred times or so. Then see if you can find an algorithm that can restore the lost fidelity.
According to information-theoretic physics, information once lost cannot be retrieved. It’s simply gone. New information can be derived (at the cost of destroying more information than it creates; this is entropy) -- but that will always be approximations. This then leads to a second, corollary question which you seem to be asserting is “where the magic happens”: can a sufficiently-clever process extrapolate from historical records a personality which is of sufficient fidelity to qualify as that same person?
I’ve had this conversation before, and I ended it then as I will end my contribution now: how could that result be shown to be the proper one? I do not want someone who is “like me” to be uploaded. I want me to be uploaded. That means an information-theoretically-complete scan of me. Not approximations. Yes, this is not a black-and-white picture. Measurements are always approximations. The point is, without those measurements in a complete state, there isn’t a way to determine what those measurements “ought to be”.
That does seem to rather violate the principle of information decay. It is not possible to derive exact information from a decayed state, for the same reason that you cannot determine which exact formula has a derivative of “3”.
That’s true in the most general situation, when there is no prior information available. But a brain is not a random chunk of matter, it’s a highly particular one, with certain patterns and regularities. So it’s not implausible that a superintelligence could restore even a moderately damaged brain.
For a real example, think of image restoration of natural scenes. A photograph is not a random matrix of pixels, it belongs to a very small subset of all possible images, and that knowledge allows seemingly “impossible” tasks of focusing, enhacement and all that.
Intelligence is not magic. Information that no longer exists cannot be reinvented with fidelity.
And yet still-shots are limited to their original resolution; anything further is artistic rendering and not a valid reconstruction of the original. It is possible to “enhance” the resolution of a still-shot of a video feed of a person’s face. It is not possible to “enhance” the resolution of a single still-shot of a person’s face.
Memories are tricky things. We do not, now, know exactly how high the fidelity needs to be to sustain a person’s “actual” psyche. If information-theoretically significant portions are missing, no amount of genius can resolve that.
For the same reason you cannot extrapolate from the number 3 to the function f(x) whose derivative of x is “3”.
You can integrate it with initial conditions though and just like we can use our prefrontal cortex to predict the probable initial conditions of events(albeit inaccurately occasionally), a powerful computer may be able to predict our complex mental pathways based on known past events with high fidelity. I’m not saying that you wont need the initial conditions to integrate the function, I just think AGI would have less trouble with it than you assume. I think you have a good point about the principle though and I will take informative decay into my perceived utility of cryonics in the future.
“known past events”—unless those past events are full-brain scans of the past, then all you’re going to get is a reduction of the scope of the configuration space and not the exact function.
“A powerful computer” != “magic”. No matter how smart you are, fifteen tons of mass moving at five thousand miles per second will still contain the same amount of kinetic energy. No amount of cleverness can extract information that has been decayed.
This is information-theoretically proven.
The question at hand is, can a personality be reconstructed from partial data by a sufficiently clever process? We have analogues to this question. Compress and decompress the same mp3 file a hundred times or so. Then see if you can find an algorithm that can restore the lost fidelity.
According to information-theoretic physics, information once lost cannot be retrieved. It’s simply gone. New information can be derived (at the cost of destroying more information than it creates; this is entropy) -- but that will always be approximations. This then leads to a second, corollary question which you seem to be asserting is “where the magic happens”: can a sufficiently-clever process extrapolate from historical records a personality which is of sufficient fidelity to qualify as that same person?
I’ve had this conversation before, and I ended it then as I will end my contribution now: how could that result be shown to be the proper one? I do not want someone who is “like me” to be uploaded. I want me to be uploaded. That means an information-theoretically-complete scan of me. Not approximations. Yes, this is not a black-and-white picture. Measurements are always approximations. The point is, without those measurements in a complete state, there isn’t a way to determine what those measurements “ought to be”.
Just thought I’d say: excellently well put.