I spent 10-20 hours looking into the issue and 1-2 hours trying to make a forecast, whereas for transformative AI I’ve spent thousands of hours thinking about the domain and hundreds on forecasting.
I wonder if the actual wall time invested actually makes a difference here. If your model is flawed, if you’re suffering from some bias, simply brute forcing effort won’t yield any returns.
I think this is a fallacious assumptions, unless rigorously proven otherwise.
That seems in odds with what optimization theory dictates—in the limit of compute (or data even) the representations should converge to the optimal ones. instrumental convergence too. I don’t get why any model trained on Othello-related tasks wouldn’t converge to such a (useful) representation.
IMHO this point is a bit overlooked. Perhaps it might be worth investigating why simply playing Othello isn’t enough? Has it to do with randomly initialized priors? I feel this could be very important, especially from a Mech Inter viewpoint—you could have different (maybe incomplete) heuristics or representations yielding the same loss. Kinda reminds me of the EAI paper which hinted that different learning rates (often) achieve the same loss but converge on different attention patterns and representations.
Perhaps there’s some variable here that we’re not considering/evaluating closely enough...