Keep in mind that currently gpt-4 is using the open agency/CAIS method of alignment. The only thing that matters is the output. So it doesn’t matter yet.
Also keep in mind philosophy doesn’t matter—we can just try it multiple ways and judge based on the data. Well, normally we could—in this case the millions of dollars a training run makes that currently infeasible.
Philosophically what you are saying makes sense.
Keep in mind that currently gpt-4 is using the open agency/CAIS method of alignment. The only thing that matters is the output. So it doesn’t matter yet.
Also keep in mind philosophy doesn’t matter—we can just try it multiple ways and judge based on the data. Well, normally we could—in this case the millions of dollars a training run makes that currently infeasible.