Case one: would-be-catastrophe-insurers don’t believe in x-risks, don’t care to investigate. (At stake: their lives)
Case two: catastrophe-insurers don’t believe in x-risks, and either don’t care to investigate, or do for some reason I’m not seeing. (At stake: their lives and insurance profits (correlated)).
They can believe in catastrophic but non-existential risks. (Like, AI causes something like crowdstrike periodically if your not trying to prevent that )
I think people mostly don’t believe in extinction risk, so the incentive isn’t nearly as real/immediate.
+1, and even for those who do buy extinction risk to some degree, financial/status incentives usually have more day-to-day influence on behavior.
I’m imagining this:
Case one: would-be-catastrophe-insurers don’t believe in x-risks, don’t care to investigate. (At stake: their lives)
Case two: catastrophe-insurers don’t believe in x-risks, and either don’t care to investigate, or do for some reason I’m not seeing. (At stake: their lives and insurance profits (correlated)).
They can believe in catastrophic but non-existential risks. (Like, AI causes something like crowdstrike periodically if your not trying to prevent that )