This post doesn’t make me actually optimistic about conjeture actually pulling this off, because for that I would have to see details but it does at least look like you understand why this is hard and why the easy versions like just telling gpt5 to imitate a nice human won’t work.
And I like that this actually looks like a plan.
Now maybe it will turn out to not be a good plan but at least is better than openAI’s plan of
“well figure out from trial and error how to make the Magic safe somehow”.
This post doesn’t make me actually optimistic about conjeture actually pulling this off, because for that I would have to see details but it does at least look like you understand why this is hard and why the easy versions like just telling gpt5 to imitate a nice human won’t work. And I like that this actually looks like a plan. Now maybe it will turn out to not be a good plan but at least is better than openAI’s plan of
“well figure out from trial and error how to make the Magic safe somehow”.