I don’t know if you have already, but this might be the time to take a long and hard look at the probblem and consider whether deep learning is the key to solving it.
What is the problem?
reckless unilateralism? → go work for policy or chip manufacturing
inabillity to specify human values? → that problem looks not DL at all to me
powerful hackers stealing all the proto-AGIs in the next 4 years? → go cybersec
deception? → (why focus there? why make an AI that might deceive you in the first place?) but that’s pretty ML, though I’m not sure interp is the way to go there
corrigibility? → might be ML, though I’m not sure all theoretical squiggles are ironed out yet
OOD behavior? → probably ML
multi-agent dynamics? → probably ML
At the very least you ought to have a clear output channel if you’re going to work with hazardous technology. Do you have the safety-mindset that prevents you from having you dual-use tech on the streets? You’re probably familiar with the abysmal safety / capabilities ratio of people working in the field, any tech that helps safety as much as capability, will therefore in practice help capability more, if you don’t distribute it carefully.
I personally would want some organisation to step up to become the keeper of secrets. I’d want them to just go all-out on cybersec, have a web of trust and basically be the solution to the unilateralists curse. That’s not ML though.
I think this problem has a large ML-part to it, but the problem is being tackled nearly-solely by ML people. I think whatever part of the problem can be tackled with ML, won’t necessarily benefit by having more ML people on it.
I don’t know if you have already, but this might be the time to take a long and hard look at the probblem and consider whether deep learning is the key to solving it.
What is the problem?
reckless unilateralism? → go work for policy or chip manufacturing
inabillity to specify human values? → that problem looks not DL at all to me
powerful hackers stealing all the proto-AGIs in the next 4 years? → go cybersec
deception? → (why focus there? why make an AI that might deceive you in the first place?) but that’s pretty ML, though I’m not sure interp is the way to go there
corrigibility? → might be ML, though I’m not sure all theoretical squiggles are ironed out yet
OOD behavior? → probably ML
multi-agent dynamics? → probably ML
At the very least you ought to have a clear output channel if you’re going to work with hazardous technology. Do you have the safety-mindset that prevents you from having you dual-use tech on the streets? You’re probably familiar with the abysmal safety / capabilities ratio of people working in the field, any tech that helps safety as much as capability, will therefore in practice help capability more, if you don’t distribute it carefully.
I personally would want some organisation to step up to become the keeper of secrets. I’d want them to just go all-out on cybersec, have a web of trust and basically be the solution to the unilateralists curse. That’s not ML though.
I think this problem has a large ML-part to it, but the problem is being tackled nearly-solely by ML people. I think whatever part of the problem can be tackled with ML, won’t necessarily benefit by having more ML people on it.