Or is that sentence meant to indicate that an instance running after training might figure out how to hack the computer running it so it can actually change it’s own weights?
I was thinking of a scenario where OpenAI deliberately gives it access to its own weights to see if it can self improve.
I agree that it would be more like to just speed up normal ML research.
I quit YouTube a few years ago and it was probably the single best decision I’ve ever made.
However I also found that I naturally substitute it with something else. For example, I subsequently became addictived to Reddit. I quit Reddit and substituted for Hackernews and LessWrong. When I quit those I substituted for checking Slack, Email and Discord.
Thankfully being addicted to Slack does seem to be substantially less harmful than YouTube.
I’ve found the app OneSec very useful for reducing addictions. It’s an app blocker that doesn’t actually block, it just delays you opening the page, so you’re much less likely to delete it in a moment of weakness.