Centrally, I’m thinking about big important things, like taking over the world, or making R&D go FOOM resulting in some other AI system which can take over the world. But I’m happy to have a broader conception of relevance on a case-by-case basis. Insofar as people have a broader conception of relevance than me, then that means AGI-by-their-definition might come hours or even several months later than AGI-by-my-definition. (The latter would happen in cases where R&D ability is significantly harder than world-takeover-ability. I guess in principle I could see this even resulting in a several-year gap, though I think that’s pretty unlikely.)
Relevant to what?
Centrally, I’m thinking about big important things, like taking over the world, or making R&D go FOOM resulting in some other AI system which can take over the world. But I’m happy to have a broader conception of relevance on a case-by-case basis. Insofar as people have a broader conception of relevance than me, then that means AGI-by-their-definition might come hours or even several months later than AGI-by-my-definition. (The latter would happen in cases where R&D ability is significantly harder than world-takeover-ability. I guess in principle I could see this even resulting in a several-year gap, though I think that’s pretty unlikely.)