A guy from Conjecture told me about this proposal in the lines of “let’s create a human-level AI system that’s built kind of like humans and is safe to run at a human speed”, and it seemed like a surprisingly bad proposal, so I looked up this post and it still looks surprisingly bad:
Even if you succeed at this, how exactly do you plan to use it? Running one single human at a human speed seems like the kind of thing one can get by simply, you know, hiring someone; running a thousand of these things at 1000x normal speed means you’re running some completely different AI system that’s bound to have a lot of internal optimisation pressures leading to sharp left turn dynamics and all of that, and more importantly, you need to somehow make the whole system aligned, and my current understanding (from talking to that guy from Conjecture) is you don’t have any ideas for how to do that.
If it is a proposal of “how we want to make relatively safe capable systems”, then cool, I just want someone to be solving the alignment problem as in “safely preventing future unaligned AIs from appearing and killing everyone”.
The capabilities of one human-level intelligence running at 1x human speed are not enough to solve anything alignment-complete (or you’d be able to spend time on some alignment-complete problem and solve it on your own).
If it is not intended to be an “alignment proposal” and is just a proposal of running some AI system safely, I’d like to know whether Conjecture has some actual alignment plan that addresses the hard bits of the problem.
A guy from Conjecture told me about this proposal in the lines of “let’s create a human-level AI system that’s built kind of like humans and is safe to run at a human speed”, and it seemed like a surprisingly bad proposal, so I looked up this post and it still looks surprisingly bad:
Even if you succeed at this, how exactly do you plan to use it? Running one single human at a human speed seems like the kind of thing one can get by simply, you know, hiring someone; running a thousand of these things at 1000x normal speed means you’re running some completely different AI system that’s bound to have a lot of internal optimisation pressures leading to sharp left turn dynamics and all of that, and more importantly, you need to somehow make the whole system aligned, and my current understanding (from talking to that guy from Conjecture) is you don’t have any ideas for how to do that.
If it is a proposal of “how we want to make relatively safe capable systems”, then cool, I just want someone to be solving the alignment problem as in “safely preventing future unaligned AIs from appearing and killing everyone”.
The capabilities of one human-level intelligence running at 1x human speed are not enough to solve anything alignment-complete (or you’d be able to spend time on some alignment-complete problem and solve it on your own).
If it is not intended to be an “alignment proposal” and is just a proposal of running some AI system safely, I’d like to know whether Conjecture has some actual alignment plan that addresses the hard bits of the problem.