Forum? I’m against ’em!
utilistrutil
Yay! Where in the Bay are you located?
THW double the size of EA.
THB that EA-minded college freshmen should study Computer Science over Biology
Thanks for the thorough response! I agree with a lot of what you wrote, especially the third section on Epistemic Learned Helplessness: “Bayesianism + EUM, but only when I feel like it” is not a justification in any meaningful sense.
On Priors
I agree that we can construct thought experiments (Pascal’s Mugging, acausal trade) with arbitrarily high stakes to swamp commonsense priors (even without religious scenarios or infinite value, which are so contested I think it would be difficult to extract a sociological lesson from them).
On Higher Order Evidence
I still think a lot of speculative conclusions we encounter in the wild suffer from undiscovered evidence and model uncertainty, and even barring this we might want to defer taking action until we’ve had a chance to learn more.
Your response jumps over these cases to those where we have “~all the evidence we’re ever going to have,” but I’m skeptical these cases exist. Even with religion, we might expect some future miracles or divine revelations to provide new evidence; we have some impossibility theorems in ethics, but new ideas might come to light that resolve paradoxes or avoid them completely. In fact, soteriological research and finding the worldview that best acausally benefits observers are proposals to find new evidence.
But ok, yes, I think we can probably come up with cases where we do have ~all the evidence and still refrain from acting on speculative + fanatical conclusions.
Problem 1: Nicheness
From here on, I’m abandoning the justification thing. I agree that we’ve found some instances where the Fourth Principle holds without Bayesian + EUM justification. Instead, I’m getting more into the semantics of what is a “norm.”
The problem is that the support for this behavior among EAs comes from niche pieces of philosophy like Pascal’s Mugging, noncausal decision theory, and infinite ethics, ideas that are niche not just relative to the general population, but also within EA. So I feel like the Fourth Principle amounts to “the minority of EAs who are aware of these edge cases behave this way when confronted with them,” which doesn’t really seem like a norm about EA.
Problem 2: Everyone’s Doing It
(This is also not a justification, it’s an observation about the Fourth Principle)
The first three principles capture ways that EA differs from other communities. The Fourth Principle, on the other hand, seems like the kind of thing that all people do? For example, a lot of people write off earning to give when they first learn about it because it looks speculative and fanatical. Now, maybe EAs differ from other people on which crazy train stop they deem “speculative,” and I think that would qualify as a norm, but relative to each person’s threshold for “speculative,” I think this is more of a human-norm than an EA-norm.
Would love your thoughts on this, and I’m looking forward to your April post :)
Thanks for the excellent post!
I think you are right that this might be a norm/heuristic in the community, but in the spirit of a “justificatory story of our epistemic practices,” I want to look a little more at
4. When arguments lead us to conclusions that are both speculative and fanatical, treat this as a sign that something has gone wrong.
First, I’m not sure that “speculative” is an independent reason that conclusions are discounted, in the sense of a filter that is applied ex-post. In your 15AI thought experiment, for example, I think that expected value calculations would get you most of the way toward explaining an increase in fanaticism; the probability that we can solve the problem might increase on net, despite the considerations you note about replication. The remaining intuition might be explained by availability/salience bias, to which EA is not immune.
Now, “speculative” scenarios might be discounted during the reasoning process if we are anchored to commonsense priors, but this would fall under typical bayesian reasoning. The priors we use and the weight we grant various pieces of evidence are still epistemic norms worthy of examination! But a different kind than suggested by the fourth principle.
Suppose “speculative” arguments are discounted ex-post in EA. I think this practice can still be redeemed on purely bayesian grounds as a correction to the following problems:
Undiscovered Evidence: An argument seems speculative not just insofar as it is divorced from empirical observations, but also insofar as we have not thought about it very much. It seems that AI risk has become less speculative as people spend more time thinking about it, holding constant actual progress in AI capabilities. We have some sense of the space of possible arguments that might be made and evidence that might be uncovered, given further research on a topic. And these undiscovered arguments/evidence might not enter neatly into our initial reasoning process. We want some way to say “I haven’t thought of it yet, but I bet there’s a good reason this is wrong,” as we might respond to some clever conspiracy theorist who presents a superficially bulletproof case for a crazy theory we haven’t encountered before. And discounting speculative conclusions is one way to achieve this.
This point is especially relevant for speculative conclusions because they often rely on chains of uncertain premises, making our credence in their conclusions all the more sensitive to new information that could update multiple steps of the argument.
Model Uncertainty: Even in a domain where we have excavated all the major arguments available to us, we may still suffer from “reasoning in the dark,” ie, in the absence of solid empirics. When reasoning about extremely unlikely events, the probability our model is wrong can swamp our credence in its conclusion. Discounting speculative conclusions allows us to say “we should be fanatical insofar as my reasoning is correct, but I am not confident in my reasoning.”
We can lump uncertainty in our axiology, epistemology, and decision theory under this section. That is, a speculative conclusion might look good only under total utilitarian axiology, bayesian epistemology, and causal decision theory, but a more conventional conclusion might be more robust to alternatives in these categories. (Note that this is a prior question to the evidential-hedging double bind set up in Appendix B.)
Chains of uncertain premises also make model uncertainty doubly important for speculative conclusions. As Anders Sandberg points out, “if you have a long argument, the probability of there being some slight error somewhere is almost 1.”
Even after accounting for these considerations, we might find that the EV of pursuing the speculative path warrants fanaticism. In this event, discounting the speculative conclusion might be a pragmatic move to deprioritize actions on this front in anticipation of new evidence that will come to light, including evidence that will bear on model uncertainty. (We might treat this as a motivation for imprecise credences, prioritizing views with sharper credences over speculative views with fuzzier ones.)
Copy that. I removed “smash,” but I’m leaving the language kind of ambiguous because my understanding of this strategy is that it’s not restricted to conventional regulations, but instead will draw on every available tool, including informal channels.
Some Things I Heard about AI Governance at EAG
Thanks for following up and thanks for the references! Definitely agree these statements are evidence; I should have been more precise and said that they’re weak evidence / not likely to move your credences in the existence/prevalence of human consciousness.
a very close connection between an entity’s capacity to model its own mental states, and consciousness itself.
The 80k episode with David Chalmers includes some discussion of meta-consciousness and the relationship between awareness and awareness of awareness (of awareness of awareness...). Would recommend to anyone interested in hearing more!
They make the interesting analogy that we might learn more about God by studying how people think about God than by investigating God itself. Similarly we might learn more about consciousness by investigating how people think about it...
We trust human self-reports about consciousness, which makes them an indispensable tool for understanding the basis of human consciousness (“I just saw a square flash on the screen”; “I felt that pinprick”).
I want to clarify that these are examples of self-reports about consciousness and not evidence of consciousness in humans. A p-zombie would be able to report these stimuli without subjective experience of them.
They are “indispensable tools for understanding” insofar as we already have a high credence in human consciousness.
Another statement that “people are equal” from GWWC.
EMs?
Thanks for this post! I’m wondering what social change efforts you find most promising?
Oh I see! Ya, crazy stuff. I liked the attention it paid to the role of foundation funding. I’ve seen this critique of foundations included in some intro fellowships, so I wonder if it would also especially resonate with leftists who are fed up with cancel culture in light of the Intercept piece.
I don’t think anything here attempts a representation of “the situation in leftist orgs” ? But yes lol same
I would endorse all of this based on experience leading EA fellowships for college students! These are good principles not just for public media discussions, but also for talking to peers.