I am using Wikipedia’s definition: “Ensuring that emergent goals match the specified goals for the system is known as inner alignment.”
Inner alignment is definitely a problem. In the case you described, the emergent goal was long term (ensure I remember the answer to 1+1), and I remain wondering whether by default short term specified goals do or do not lead to strange long term goals like in your example.
I am using Wikipedia’s definition: “Ensuring that emergent goals match the specified goals for the system is known as inner alignment.”
Inner alignment is definitely a problem. In the case you described, the emergent goal was long term (ensure I remember the answer to 1+1), and I remain wondering whether by default short term specified goals do or do not lead to strange long term goals like in your example.