As I understand, the following two positions are largely accepted in the EA community:
Temporal position should not impact ethics (hence longtermism)
But if we are time-agnostic, then neutrality against making happy lives seems to imply a preference for extinction over any future where even a tiny amount of suffering exists.
So am I missing something here? (Perhaps “neutrality against creating happy lives” can’t be expressed in a way that’s temporally agnostic?)