Philosophy, global priorities and animal welfare research. My current specific interests include: philosophy of mind, moral weights, person-affecting views, preference-based views and subjectivism, moral uncertainty, decision theory, deep uncertainty/cluelessness and backfire risks, s-risks, and indirect effects on wild animals.
I’ve also done economic modelling for some animal welfare issues.
Want to leave anonymous feedback for me, positive, constructive or negative? https://www.admonymous.co/michael-st-jules
I suspect we need to involve our criteria for defining and picking bracketings here.
In practice, I think it doesn’t make sense to just bracket in the bad long-term effects or just bracket in the good ones. You might be able to carve out bracketings that include only bad (or only good) long-term effects and effects outweighed by them, but not all bad (or all good) long-term effects. But that will depend on the particulars.
I think if we only do spatiotemporal bracketing, it tells us to ignore the far future and causally inaccessible spacetime locations, because each such location is made neither determinately better off in expectation nor determinately worse off in expectation. I’m not entirely sure where the time cutoff should start in practice, but it would be related to AGI’s arrival. That could make us neartermist.
But we may also want to bracket out possibilities, not just ST locations. Maybe we can bracket out AGI by date X, for various X (or the min probability of it across choices, in case we affect its probability), and focus on non-AGI outcomes we may be more clueful about. If we bracket out the right set of possibilities, maybe some longtermist interventions will look best.