Evolution was working within tight computational efficiency limits (the human brain burns roughly 1⁄6 of our total calories), using a evolutionary algorithm rather than gradient descent training scheme which is significantly less efficient, and we’re now running the human brain well outside it’s training distribution (there were no condoms on the Savannah) — nevertheless, the human population is 8 billion and counting, and we dominate basically every terrestrial ecosystem on the planet. I think some people overplay how much inner alignment failure there is between human instincts and human genetic fitness.
So:
Use a model large enough to learn what you’re trying to teach it
Use stochastic gradient descent
Ask your AI to monitor for inner alignment problems (we do know Doritos are bad for us)
Retrain if you find yourself far enough outside your training distribution that inner alignment issues are becoming a problem
Evolution was working within tight computational efficiency limits (the human brain burns roughly 1⁄6 of our total calories), using a evolutionary algorithm rather than gradient descent training scheme which is significantly less efficient, and we’re now running the human brain well outside it’s training distribution (there were no condoms on the Savannah) — nevertheless, the human population is 8 billion and counting, and we dominate basically every terrestrial ecosystem on the planet. I think some people overplay how much inner alignment failure there is between human instincts and human genetic fitness.
So:
Use a model large enough to learn what you’re trying to teach it
Use stochastic gradient descent
Ask your AI to monitor for inner alignment problems (we do know Doritos are bad for us)
Retrain if you find yourself far enough outside your training distribution that inner alignment issues are becoming a problem