Recently I’ve had success introducing people to AI risk by telling them about ELK, and specifically how human simulation is likely to be favored. Providing this or a similarly general argument, e.g. that power-seeking is convergent, seems both more intuitive (humans can also do simulation and power-seeking) and faithful to actual AI risk drivers to me than the video speed angle? ELK and power-seeking are also useful complements to specific AI risk scenarios.
The video speed framing seems to make the undesirable suggestion that AI ontology will be human-but-faster. I would prefer an example which highlighted the likely differences in ontology. Examples highlighting ontology mismatch have the benefit of neatly motivating the problems of value learning and misaligned proxy objectives.
Recently I’ve had success introducing people to AI risk by telling them about ELK, and specifically how human simulation is likely to be favored. Providing this or a similarly general argument, e.g. that power-seeking is convergent, seems both more intuitive (humans can also do simulation and power-seeking) and faithful to actual AI risk drivers to me than the video speed angle? ELK and power-seeking are also useful complements to specific AI risk scenarios.
The video speed framing seems to make the undesirable suggestion that AI ontology will be human-but-faster. I would prefer an example which highlighted the likely differences in ontology. Examples highlighting ontology mismatch have the benefit of neatly motivating the problems of value learning and misaligned proxy objectives.