It is not Kurzweil, it is rather Aubrey De Grey’s pretty good execution of a “weak inside view” analysis. Longevity escape velocity, as he calls it. I have my own minor complaint about the “accelerated via the a Singularity” part, which is markedly not part of De Grey’s analysis (assuming that the OP really wrote the above paragraph with De Grey in mind and not someone else).
Yes I was hinting at longevity escape velocity, which doesn’t depend on a Singularity per se, but a Singularity—meaning a vast increase in effective intelligence and computation—undoubtedly has a high probability of achieving longevity escape velocity, amongst other things.
I happen to believe the odds of achieving that escape velocity anytime soon without a vast increase in effective intelligence and computation are very low.
Ahh I can see why you might take objection to that—Singularity can be a vague term. Nonetheless, it does describe a range of future scenarios which I do believe there is sufficient evidence to warrant as quite possible, without considering specifics—it’s actually good in this case to use a term that wraps a diverse set of related futures together.
It is not Kurzweil, it is rather Aubrey De Grey’s pretty good execution of a “weak inside view” analysis. Longevity escape velocity, as he calls it. I have my own minor complaint about the “accelerated via the a Singularity” part, which is markedly not part of De Grey’s analysis (assuming that the OP really wrote the above paragraph with De Grey in mind and not someone else).
Yes I was hinting at longevity escape velocity, which doesn’t depend on a Singularity per se, but a Singularity—meaning a vast increase in effective intelligence and computation—undoubtedly has a high probability of achieving longevity escape velocity, amongst other things.
I happen to believe the odds of achieving that escape velocity anytime soon without a vast increase in effective intelligence and computation are very low.
That.
Ahh I can see why you might take objection to that—Singularity can be a vague term. Nonetheless, it does describe a range of future scenarios which I do believe there is sufficient evidence to warrant as quite possible, without considering specifics—it’s actually good in this case to use a term that wraps a diverse set of related futures together.
I wouldn’t have used it, because the term “singularity” makes me think of a sign saying “STOP: No humans beyond this point.”