Hmm. I guess that might be okay? as long as you don’t do really intense planning, the model shouldn’t be any more misaligned than a human, so it then boils down to training kindness by example and figuring out game dynamics. https://www.youtube.com/watch?v=ENpdhwYoF5g. more braindump of safety content I always want to recommend in every damn conversation here on my shortform
Hmm. I guess that might be okay? as long as you don’t do really intense planning, the model shouldn’t be any more misaligned than a human, so it then boils down to training kindness by example and figuring out game dynamics. https://www.youtube.com/watch?v=ENpdhwYoF5g. more braindump of safety content I always want to recommend in every damn conversation here on my shortform