if the original model learned complex, power-seeking behaviors that doesn’t help it do well on the training data
The problem with power-seeking behavior is that it helps to do well in quite broad range of tasks.
As of right now, I don’t think that LLMs are trained to be power seeking and deceptive.
Power-seeking is likely if the model is directly maximizing rewards, but LLMs are not quite doing this.
The problem with power-seeking behavior is that it helps to do well in quite broad range of tasks.
As of right now, I don’t think that LLMs are trained to be power seeking and deceptive.
Power-seeking is likely if the model is directly maximizing rewards, but LLMs are not quite doing this.