But the expected value of existential risk reduction is—if not infinite, which I think it clearly is in expectation—extremely massive.
I commented something similar on your blog, but as soon as you allow that one decision is infinite in expectation you have to allow that all outcomes are, since whatever possibility of infinite value you have given that action must still be present without it.
If you think the Bostrom number of 10^52 happy people has a .01% chance of being right, then you’ll get 10^48 expected future people if we don’t go extinct, meaning reducing odds of existential risks by 1/10^20 creates 10^28 extra lives.
Reasoning like this seems kind of scope insensitive to me. In the real world, it’s common to see expected payoffs declining as offered rewards get larger, and I don’t see any reason to think this pattern shouldn’t typically generalise to most such prospects, even when the offer is astronomically large.
The odds are not trivial that if we get very advanced AI, we’ll basically eliminate any possibility of human extinction for billions of years.
I think the stronger case is just security in numbers. Get a civilisation around multiple star systems and capable of proliferating, and the odds of its complete destruction rapidly get indistinguishable from 0.
Not necessarily. It depends on
a) your credence distribution of TAI after this decade,
b) your estimate of annual risk per year of other catastrophes, and
c) your estimate of the comparative longterm cost of other catastrophes.
I don’t think it’s unreasonable to think, for example, that
there’s a very long tail to when TAI might arrive, given that its prospects of arriving in 2-3 decades are substantially related to to its prospects of arriving this decade) it arriving this decade (e.g. if we scale current models substantially and they still show no signs of becoming TAI, that undermines the case for future scaling getting us there under same paradigm); or
the more pessimistic annual risk estimates I talked about in the previous essay of 1-2% per year are correct, and that future civilisations will have a sufficiently increased difficulty for a collapse to cost to have near 50% the expected cost of extinction
And either of these beliefs (and others) would suggest we’re relatively overspending on on AI.
This is grossly disingenuous. Yudkowsky frames his call for airstrikes as what we ‘need’ to do, and describes them in the context of the hypothetical ‘if I had infinite freedom to write laws’. Hendrycks is slightly less direct in actively calling for it, claiming that it’s the default, but the document clearly states the intent of supporting it ‘we outline measures to maintain the conditions for MAIM’.
These aren’t the words of people dispassionately observing a phenomenon—they are both clearly trying to bring about the scenarios they describe when the lines they’ve personally drawn are crossed.