Thanks for taking the time to respond! I find your point of view more plausible now that I understand it a little bit better (though I’m still not sure of how convincing I find it overall).
Sure, and thank you for being interested in what I have written here.
I didn’t offer an argument meant to convince, more a listing of perspectives on what is actually happening around EA “updating”. For example, to know that an EA is confusing judgments of feeling intensity and truth probability, I would have to have evidence that they are acting in good faith to” update” in the first place, rather than (unconsciously) pursuing some other agenda. As another example, to know that an EA has a betting problem, the psychological kind, I would have see pathological betting behavior on their part (for example, that ruined their finances or their relationships). Different EA’s are typically doing different things with their “updating” at different times. Some of them are sure to have a bit of risk-seeking that is unhealthy and distorts their perspective, others are doing their best to measure their feelings of certainty as a fractional number, and still others are being cynical with their offerings of numbers. If the superforecasters among the EA’s are doing so well with their probability estimates, I wish they would offer some commentary on free will, game theory, and how they model indeterminism. I would learn something.
If there were a tight feedback loop about process, if everyone understood not only the math, but also the evidence that moves your credence numbers, if there were widespread agreement that a new bit of evidence X should move probability of credence Y an amount Z , then I could believe that there was systematic care about epistemics involved in the community. I could believe that EA folks are always training to improve their credence probability assignments. I could believe that EA folks train their guts, hone their rationality, and sharpen their research skills, all at the same time.
But what actually goes on in EA seems very subjective, anyone can make up any number and claim their evidence justifies it and don’t really have to prove anything, and in fact, it’s hard to make a case that is not subjective, and there’s a spread of numbers on most credences, or there really isn’t but I’m looking at the numbers and seeing most people quoting somebody else then, while everyday credences are just treated as regular beliefs (binary or constrained with experience) plus there is plenty of secondary gain potential, so I don’t believe that EA folks are using their probability estimates well at all, no.
NOTE: By “secondary gain potential”, I mean gain from an action other than the action’s ostensible gains. A person doing updating for its secondary gains might be developing:
feelings of belonging in the EA community.
a sense of control over their feelings of doubt or certainty.
an addiction to betting or other financial risk-taking.
a (false) sense of efficacy in rational arguments.
or something else that they would have to tell you.
I suspect that most EA’s are being sloppy in their offerings of probabilities and credence levels. They offer something closer to epistemic status metrics for their own statements, which rely on poor evidence or no evidence, just a feeling, maybe of worry or disappointment at failing other’s expectations, or the converse, a high metric corresponding to a poor judgement of the quality of their own work, typically basing that on other work in EA, unsurprisingly. EA folks choose numbers to emphasize a message to their audience that reflects what they’re comfortable sharing, not what they actually believe. There’s some real controversy among EA’s, but it comes from outside the community, apparently, through people like me.
So as I said, this whole updating thing has gone wrong in EA. EA as a research community would do better to drop updating and the bayesian kick until they have a better handle on contexts suitable for it. If superforecasting offers those contexts, then great!
Lets see superforecasters address existential risk with any success. Lets just see them do the research first so they that have some clear preliminary answers to what the dangers can shape up to be. Presuming that they are not using foresight but forecasting techniques, that should be interesting.
NOTE: Once I learn more about superforecasting, I might constrain my beliefs about how EA’s update to exclude superforecasters working in their successful forecasting domains. I would do that right now, but I’m suspicious of the superforecaster reputation, so I’ll wait until I can learn more.
Thanks for taking the time to respond! I find your point of view more plausible now that I understand it a little bit better (though I’m still not sure of how convincing I find it overall).
Sure, and thank you for being interested in what I have written here.
I didn’t offer an argument meant to convince, more a listing of perspectives on what is actually happening around EA “updating”. For example, to know that an EA is confusing judgments of feeling intensity and truth probability, I would have to have evidence that they are acting in good faith to” update” in the first place, rather than (unconsciously) pursuing some other agenda. As another example, to know that an EA has a betting problem, the psychological kind, I would have see pathological betting behavior on their part (for example, that ruined their finances or their relationships). Different EA’s are typically doing different things with their “updating” at different times. Some of them are sure to have a bit of risk-seeking that is unhealthy and distorts their perspective, others are doing their best to measure their feelings of certainty as a fractional number, and still others are being cynical with their offerings of numbers. If the superforecasters among the EA’s are doing so well with their probability estimates, I wish they would offer some commentary on free will, game theory, and how they model indeterminism. I would learn something.
If there were a tight feedback loop about process, if everyone understood not only the math, but also the evidence that moves your credence numbers, if there were widespread agreement that a new bit of evidence X should move probability of credence Y an amount Z , then I could believe that there was systematic care about epistemics involved in the community. I could believe that EA folks are always training to improve their credence probability assignments. I could believe that EA folks train their guts, hone their rationality, and sharpen their research skills, all at the same time.
But what actually goes on in EA seems very subjective, anyone can make up any number and claim their evidence justifies it and don’t really have to prove anything, and in fact, it’s hard to make a case that is not subjective, and there’s a spread of numbers on most credences, or there really isn’t but I’m looking at the numbers and seeing most people quoting somebody else then, while everyday credences are just treated as regular beliefs (binary or constrained with experience) plus there is plenty of secondary gain potential, so I don’t believe that EA folks are using their probability estimates well at all, no.
NOTE: By “secondary gain potential”, I mean gain from an action other than the action’s ostensible gains. A person doing updating for its secondary gains might be developing:
feelings of belonging in the EA community.
a sense of control over their feelings of doubt or certainty.
an addiction to betting or other financial risk-taking.
a (false) sense of efficacy in rational arguments.
or something else that they would have to tell you.
I suspect that most EA’s are being sloppy in their offerings of probabilities and credence levels. They offer something closer to epistemic status metrics for their own statements, which rely on poor evidence or no evidence, just a feeling, maybe of worry or disappointment at failing other’s expectations, or the converse, a high metric corresponding to a poor judgement of the quality of their own work, typically basing that on other work in EA, unsurprisingly. EA folks choose numbers to emphasize a message to their audience that reflects what they’re comfortable sharing, not what they actually believe. There’s some real controversy among EA’s, but it comes from outside the community, apparently, through people like me.
So as I said, this whole updating thing has gone wrong in EA. EA as a research community would do better to drop updating and the bayesian kick until they have a better handle on contexts suitable for it. If superforecasting offers those contexts, then great!
Lets see superforecasters address existential risk with any success. Lets just see them do the research first so they that have some clear preliminary answers to what the dangers can shape up to be. Presuming that they are not using foresight but forecasting techniques, that should be interesting.
NOTE: Once I learn more about superforecasting, I might constrain my beliefs about how EA’s update to exclude superforecasters working in their successful forecasting domains. I would do that right now, but I’m suspicious of the superforecaster reputation, so I’ll wait until I can learn more.