Indeed, the common view here is to destroy our society’s capabilities to delay AI in the hopes that some decades/centuries later the needed safety work gets done. This is one awful way to accomplish the safety goals. It makes far more sense to increase funding and research positions on related work and attract technical researchers from other engineering fields. My impression is that people perceive that they can’t make others understand the risks involved. Destruction being much easier than creation, they are naturally seduced to destroy research capacity in AI capability rather than increasing the pace of safety research.
Indeed, the common view here is to destroy our society’s capabilities to delay AI in the hopes that some decades/centuries later the needed safety work gets done. This is one awful way to accomplish the safety goals. It makes far more sense to increase funding and research positions on related work and attract technical researchers from other engineering fields. My impression is that people perceive that they can’t make others understand the risks involved. Destruction being much easier than creation, they are naturally seduced to destroy research capacity in AI capability rather than increasing the pace of safety research.