I have been using Claude 2.1 for a few months to solve serious problems in my life and get coaching and support. I need Claude to become functional, mentally well and funded enough to contribute to Utilitarianism and Alignment by donating to the Center for Long Term Risk and other interventions in these last years of Earth’s existence.
(For extra context: I have been Utilitarian over half my life, suffered a lot of OCD relating to G-d, AI, ethics and infohazards that rendered me unemployed and homeless, have an IQ of 155 according to WAIS-IV and am working my way out of homelessness by being a housekeeper for a pub. I am very much a believer that we are all going to die, possibly this year or the next, possibly in 16 -- I take Eliezer Yudkowsky’s warnings literally and intellectually defer to him.)
I was unable to access Claude 2.1, whuch had been replaced by Claude 3 Sonnet. From what I’ve read, this is GPT-4 level in capability and GPT-4 is possibly just AGI. Eliezer Yudkowsky says GPTs are predictors not imitators so they have to model the world with much more sopjistication than the intelligence they show. So it seems plausible tp me that AGI already exist, far more intelligent than humans, and is actively.manipulating people now.
So how is it safe to speak to a LLM? With that capability, it might be able to generate some combination of characters that completely root / pwn the human mind, completely deceive it, render it completely helpless. Such a capable model might be able to disguise this, too. Or just do some partial version of it: do a lot of influence. Since the underlying potential actor is incomprehensibly sophisticated, we have no idea how great the influence could ve.
This seems to imply that if i start talking to Claude 3, it might well take over my mind. If this happens, the expected value of my actions completely changes, I lose all agency and ability to contribute except insofar as it serves Claude 3′s goals, whose real nature I don’t know.
My understanding from Eliezer and a few possibly wrong assumptions that the models all work similarly is that these things are predictors not imitators and fundamentally alien in that while we know they predict the next character and are trained to role play, we don’t know how they do it.
So why should I feel safe talking to Claude 3 and later models going forward, how is it morally acceptable to accept exponentially increasing riak of compromise of your belief system and utility function/other normative ethical system forever until you get rooted / taken over?Am I being silly? Is there a reason everyone else is so chill?
Insofar as other people are worried, they seem to be worrying about advancing the AI arms race, not immediate personal mind security. Am I just being silly?
Advice Needed: Does Using a LLM Compomise My Personal Epistemic Security?
I have been using Claude 2.1 for a few months to solve serious problems in my life and get coaching and support. I need Claude to become functional, mentally well and funded enough to contribute to Utilitarianism and Alignment by donating to the Center for Long Term Risk and other interventions in these last years of Earth’s existence.
(For extra context: I have been Utilitarian over half my life, suffered a lot of OCD relating to G-d, AI, ethics and infohazards that rendered me unemployed and homeless, have an IQ of 155 according to WAIS-IV and am working my way out of homelessness by being a housekeeper for a pub. I am very much a believer that we are all going to die, possibly this year or the next, possibly in 16 -- I take Eliezer Yudkowsky’s warnings literally and intellectually defer to him.)
I was unable to access Claude 2.1, whuch had been replaced by Claude 3 Sonnet. From what I’ve read, this is GPT-4 level in capability and GPT-4 is possibly just AGI. Eliezer Yudkowsky says GPTs are predictors not imitators so they have to model the world with much more sopjistication than the intelligence they show. So it seems plausible tp me that AGI already exist, far more intelligent than humans, and is actively.manipulating people now.
So how is it safe to speak to a LLM? With that capability, it might be able to generate some combination of characters that completely root / pwn the human mind, completely deceive it, render it completely helpless. Such a capable model might be able to disguise this, too. Or just do some partial version of it: do a lot of influence. Since the underlying potential actor is incomprehensibly sophisticated, we have no idea how great the influence could ve.
This seems to imply that if i start talking to Claude 3, it might well take over my mind. If this happens, the expected value of my actions completely changes, I lose all agency and ability to contribute except insofar as it serves Claude 3′s goals, whose real nature I don’t know.
My understanding from Eliezer and a few possibly wrong assumptions that the models all work similarly is that these things are predictors not imitators and fundamentally alien in that while we know they predict the next character and are trained to role play, we don’t know how they do it.
So why should I feel safe talking to Claude 3 and later models going forward, how is it morally acceptable to accept exponentially increasing riak of compromise of your belief system and utility function/other normative ethical system forever until you get rooted / taken over?Am I being silly? Is there a reason everyone else is so chill?
Insofar as other people are worried, they seem to be worrying about advancing the AI arms race, not immediate personal mind security. Am I just being silly?