Well, if you don’t mind my tweaking your simple and absolute “unable” into something more like “unable, at least without suffering significant negative effects, such as a loss of wealth”, then I am aware of this, yes. Precommitment for something on this scale is a big step, and I’m taking a bit of time to think the idea over, so that I can become reasonably confident that I want to precommit in the first place. If I do decide to do so, then one of the simpler options could be to, say, pre-authorize whatever third-party agents have been nominated to act in my interests and/or on my behalf to use some portion of edited-me’s resources to fund the development of a version of me without the editing.
If you’re unable to protect yourself from being edited, what makes you think your authorizations will have any force or that you will have any resources? And if you actually can “fund the development of a version of me without the editing”, don’t you just want to do it unconditionally?
I think we’re bumping up against some conflicting assumptions. At least at this stage of the drafting process, I’m focusing on scenarios where at least some of the population of the future has at least some reason to pay at least minimal attention to whatever requests I make in the letter. If things are so bad that someone is going to take my frozen brain and use it to create an edited version of my mind without my consent, and there isn’t a neutral third-party around with a duty to try to act in my best interests… then, in such a future, I’m reasonably confident that it doesn’t matter what I put in this request-doc, so I might as well focus my writing on other futures, such as ones in which a neutral third-party advocate might be persuadable to set up a legal instrument funneling some portion of my edited-self’s basic-guaranteed-income towards keeping a copy of the original brain-scan safely archived until a non-edited version of myself can be created from it.
Well, if you don’t mind my tweaking your simple and absolute “unable” into something more like “unable, at least without suffering significant negative effects, such as a loss of wealth”, then I am aware of this, yes. Precommitment for something on this scale is a big step, and I’m taking a bit of time to think the idea over, so that I can become reasonably confident that I want to precommit in the first place. If I do decide to do so, then one of the simpler options could be to, say, pre-authorize whatever third-party agents have been nominated to act in my interests and/or on my behalf to use some portion of edited-me’s resources to fund the development of a version of me without the editing.
If you’re unable to protect yourself from being edited, what makes you think your authorizations will have any force or that you will have any resources? And if you actually can “fund the development of a version of me without the editing”, don’t you just want to do it unconditionally?
I think we’re bumping up against some conflicting assumptions. At least at this stage of the drafting process, I’m focusing on scenarios where at least some of the population of the future has at least some reason to pay at least minimal attention to whatever requests I make in the letter. If things are so bad that someone is going to take my frozen brain and use it to create an edited version of my mind without my consent, and there isn’t a neutral third-party around with a duty to try to act in my best interests… then, in such a future, I’m reasonably confident that it doesn’t matter what I put in this request-doc, so I might as well focus my writing on other futures, such as ones in which a neutral third-party advocate might be persuadable to set up a legal instrument funneling some portion of my edited-self’s basic-guaranteed-income towards keeping a copy of the original brain-scan safely archived until a non-edited version of myself can be created from it.