I’m a Postdoctoral Research Fellow at Oxford University’s Global Priorities Institute.
Previously, I was a Philosophy Fellow at the Center for AI Safety.
So far, my work has mostly been about the moral importance of future generations. Going forward, it will mostly be about AI.
You can email me at elliott.thornley@philosophy.ox.ac.uk.
Thanks! We think that advanced POST-agents won’t deliberately try to get shut down, for the reasons we give in footnote 5 (relevant part pasted below). In brief:
advanced agents will be choosing between lotteries
we have theoretical reasons to expect that agents that satisfy POST (when choosing between trajectories) will be ‘neutral’ (when choosing between lotteries): they won’t spend resources to shift probability mass between different-length trajectories.
So (we think) neutral agents won’t deliberately try to get shut down if doing so costs resources.