A policeman sees a drunk man searching for something under a streetlight and asks what the drunk has lost. He says he lost his keys and they both look under the streetlight together. After a few minutes the policeman asks if he is sure he lost them here, and the drunk replies, no, and that he lost them in the park. The policeman asks why he is searching here, and the drunk replies, “this is where the light is”.
I’ve always been sympathetic to the drunk in this story. If the key is in the light, there is a chance of finding it. If it is in the dark, he’s not going to find it anyway so there isn’t much point in looking there.
Given the current state of alignment research, I think it’s fair to say that we don’t know where the answer will come from. I support The Plan and I hope research continues on it. But if I had to guess, alignment will not be solved via getting a bunch of physicists thinking about agent foundations. It will be solved by someone who doesn’t know better making a discovery they “wasn’t supposed to work”.
On an interesting side here a fun story about experts repeatedly failing to make an obvious-in-hindsight discovery because they “knew better”.
Is your disagreement specifically with the word “IQ” or with the broader point, that AI progress is continuing to make progress at a steady rate that implies things are going to happen soon-ish (2-4 years)?
If specifically with IQ, feel free to replace the word with “abstract units of machine intelligence” wherever appropriate.
If with “big things soon”, care to make a prediction?