I might have higher probability thresholds for what I consider Pascalian, but it’s also a matter of how much of my time and resources I have to give. This feels directly intuitive to me, and it can be cashed out in terms of normative uncertainty about decision theory/my risk appetite. I limit my budget for views that are more risk neutral.
Voting is low commitment, so not Pascalian this way. Devoting your career to nuclear policy seems Pascalian to me. Working on nuclear policy among many other things you work on doesn’t seem Pascalian. Fighting in a war only with deciding the main outcome of who wins/loses in mind seems Pascalian.
Some of these things may have other benefits regardless of whether you change the main binary-ish outcome you might have in mind. That can make them not Pascalian.
Also, people do these things without thinking much or at all about the probability that they’d affect the main outcome. Sometimes they’re “doing their part”, or it’s a matter of identity or signaling. Those aren’t necessarily bad reasons. But they’re not even bothering to check whether it would be Pascalian.
EDIT: I’d also guess the people self-selecting into doing this work, especially without thinking about the probabilities, would have high implied probabilities of affecting the main binary-ish outcome, if we interpreted them as primarily concerned with that.
I might have higher probability thresholds for what I consider Pascalian, but it’s also a matter of how much of my time and resources I have to give. This feels directly intuitive to me, and it can be cashed out in terms of normative uncertainty about decision theory/my risk appetite. I limit my budget for views that are more risk neutral.
Voting is low commitment, so not Pascalian this way. Devoting your career to nuclear policy seems Pascalian to me. Working on nuclear policy among many other things you work on doesn’t seem Pascalian. Fighting in a war only with deciding the main outcome of who wins/loses in mind seems Pascalian.
Some of these things may have other benefits regardless of whether you change the main binary-ish outcome you might have in mind. That can make them not Pascalian.
Also, people do these things without thinking much or at all about the probability that they’d affect the main outcome. Sometimes they’re “doing their part”, or it’s a matter of identity or signaling. Those aren’t necessarily bad reasons. But they’re not even bothering to check whether it would be Pascalian.
EDIT: I’d also guess the people self-selecting into doing this work, especially without thinking about the probabilities, would have high implied probabilities of affecting the main binary-ish outcome, if we interpreted them as primarily concerned with that.