1. I think we should be more willing than we currently are to ban or softban people.
2. I think we should not assume that CEA’s Community Health team “has everything covered”
3. I think more people should feel empowered to tell CEA CH about their concerns, even (especially?) if other people appear to not pay attention or do not think it’s a major concern.
4. I think the community is responsible for helping the CEA CH team with having a stronger mandate to deal with interpersonal harm, including some degree of acceptance of mistakes of overzealous moderation.
(all views my own) I want to publicly register what I’ve said privately for a while:
For people (usually but not always men) who we have considerable suspicion that they’ve been responsible for significant direct harm within the community, we should be significantly more willing than we currently are to take on more actions and the associated tradeoffs of limiting their ability to cause more harm in the community.
Some of these actions may look pretty informal/unofficial (gossip, explicitly warning newcomers against specific people, keep an unofficial eye out for some people during parties, etc). However, in the context of a highly distributed community with many people (including newcomers) that’s also embedded within a professional network, we should be willing to take more explicit and formal actions as well.
This means I broadly think we should increase our willingness to a) ban potentially harmful people from events, b) reduce grants we make to people in ways that increase harmful people’s power, c) warn organizational leaders about hiring people in positions of power/contact with potentially vulnerable people.
I expect taking this seriously to involve taking on nontrivial costs. However, I think this is probably worth it.
I’m not sure why my opinion here is different from others[1]’, however I will try to share some generators of my opinion, in case it’s helpful:
A. We should aim to be a community that’s empowered to do the most good. This likely entails appropriately navigating the tradeoff of both attempting to reducing the harms of a) contributors feeling or being unwelcome due to sexual harassment or other harms and b) contributors feeling or being unwelcome due to false accusations or overly zealous response.
B. I think some of this is fundamentally a sensitivity vs specificity tradeoff. If we have a detection system that’s too tuned to reduce the risk of false positives (wrong accusations being acted on), we will overlook too many false negatives (people being too slow to be banned/censured, or not at all), and vice versa.
Avoid false negatives: take action if there’s reason to think someone is causing problems
Avoid false positives: don’t unfairly harm someone’s reputation / ability to participate in EA
In the world we live in, I’ve yet to hear of a single incidence where, in full context, I strongly suspect CEA CH (or for that matter, other prominent EA organizations) was overzealous in recommending bans due to interpersonal harm. If our institutions are designed to only reduce first-order harm (both from direct interpersonal harm and from accusations), I’d expect to see people err in both directions.
Given the (apparent) lack of false positives, I broadly expect we accept too high a rate of false negatives. More precisely, I do not think CEA CH’s current work on interpersonal harm will lead to a conclusion like “We’ve evaluated all the evidence available for the accusations against X. We currently think there’s only a ~45% chance that X has actually committed such harms, but given the magnitude of the potential harm, and our inability to get further clarity with more investigation, we’ve pre-emptively decided to ban X from all EA Globals pending further evidence.”
Instead, I get the impression that substantially more certainty is deemed necessary to take action. This differentially advantages conservatism, and increases the probability and allowance of predatory behavior.
C. I expect an environment with more enforcement is more pleasant than an environment with less enforcement.
I expect an environment where there’s a default expectation of enforcement for interpersonal harm is more pleasant for both men and women. Most directly in reducing the first-order harm itself, but secondarily an environment where people are less “on edge” for potential violence is generally more pleasant. As a man, I at least will find it more pleasant to interact with women in a professional context if I’m not worried that they’re worried I’ll harm them. I expect this to be true for most men, and the loud worries online about men being worried about false accusations to be heavily exaggerated and selection-skewed[2].
Additionally, I note that I expect someone who exhibit traits like reduced empathy, willingness to push past boundaries, sociopathy, etc, to also exhibit similar traits in other domains. So someone who is harmful in (e.g.) sexual matters is likely to also be harmful in friendly and professional matters. For example, in the more prominent cases I’m aware of where people accused of sexual assault were eventually banned, they also appeared to have done other harmful activities like a systematic history of deliberate deception, being very nasty to men, cheating on rent, harassing people online, etc. So I expect more bans to broadly be better for our community.
D. I expect people who are involved in EA for longer to be systematically biased in both which harms we see, and which things are the relevant warning signals.
The negative framing here is “normalization of deviance”. The more neutral framing here is that people (including women) who have been around EA for longer a) may be systematically less likely to be targeted (as they have more institutional power and cachet) and b) are selection-biased to be less likely to be harmed within our community (since the people who have received the most harm are more likely to have bounced off).
E. I broadly trust the judgement of CEA CH in general, and Julia Wise in particular.
I think their judgement is broadly reasonable, and they act well within the constraints that they’ve been given. If I did not trust them (e.g. if I was worried that they’d pursue political vendettas in the guise of harm-reduction), I’d be significantly more worried about given them more leeway to make mistakes with banning people.[3]
F. Nonetheless, the CEA CH team is just one group of individuals, and does a lot of work that’s not just on interpersonal harm. We should expect them to a) only have a limited amount of information to act on, and b) for the rest of EA to need to pick up some of the slack where they’ve left off.
For a), I think an appropriate action is for people to be significantly more willing to report issues to them, as well as make sure new members know about the existence of the CEA CH team and Julia Wise’s work within it. For b), my understanding is that CEA CH sees themself as having what I call a “limited purview”: e.g. they only have the authority to ban people from official CEA and maybe CEA-sponsored events, and not e.g. events hosted by local groups. So I think EA community-builders in a group organizing capacity should probably make it one of their priorities to be aware of the potential broken stairs in their community, and be willing to take decisive actions to reduce interpersonal harms.
Remember: EA is not a legal system. Our objective is to do the most good, not to wait to be absolutely certain of harm before taking steps to further limit harm.
One thing my post does not cover is opportunity cost. I mostly framed things as changing the decision-boundary. However, in practice I can see how having more bans is more costly in time and maybe money than the status quo. I don’t have good calculations here, however my intuition is strongly in the direction that having a safer and more cohesive is worth the relevant opportunity costs.
fwiw my guess is that the average person in EA leadership wishes the CEA CH team does more (is currently insufficiently punitive), rather than wish that they did less (is currently overzealous). I expect there’s significant variance in this opinion however.
I can imagine this being a crux for people who oppose greater action. If so, I’d like to a) see this argument explicitly being presented and debated, and b) see people propose alternatives for reducing interpersonal harm that routes around CEA CH.
Thank you so much for laying out this view. I completely agree, including every single subpoint (except the ones about the male perspective which I don’t have much of an opinion on). CEA has a pretty high bar for banning people. I’m in favour of lowering this bar as well as communicating more clearly that the bar is really high and therefore someone being part of the community certainly isn’t evidence they are safe.
Thank you in particular for point D. I’ve never been quite sure how to express the same point and I haven’t seen it written up elsewhere.
It’s a bit unfortunate that we don’t seem to have agreevote on shortforms.
As an aside, I dislike calling out gender like this, even with the “not always” disclaimer. Compare: “For people (usually but not always black people)” would be considered inappropriate.
I would prefer not to bring up gender at all. If someone commits sexual harassment, it doesn’t particularly matter what their gender is. And it may be true that men do it more than women, but that’s not really relevant, any more than it would be relevant if black people committed sexual harassment more than average.
And it may be true that men do it more than women, but that’s not really relevant.
It’s not that it “may be” true—it is true. I think it’s totally relevant: if some class of people are consistently the perpetuators of harm against another group, then surely we should be trying to figure out why that it is the case so we can stop it? Not providing that information seems like it could seriously impede our efforts to understand and address the problem (in this case, sexism & patriarchy).
I’m also confused by your analogy to race—I think you’re implying that it would be discriminatory to mention race if talking about other bad things being done, but I also feel like this is relevant. In this case I think it’s a bit different, however, as there’s other confounders present (e.g. black people are much more highly incarcerated, earn less on average, generally much less privileged) which all might increase rates of doing said bad thing. So in this case, it’s not a result of their race, but rather a result of the unequal socioeconomic conditions faced when someone is a certain race.
tl;dr:
In the context of interpersonal harm:
1. I think we should be more willing than we currently are to ban or softban people.
2. I think we should not assume that CEA’s Community Health team “has everything covered”
3. I think more people should feel empowered to tell CEA CH about their concerns, even (especially?) if other people appear to not pay attention or do not think it’s a major concern.
4. I think the community is responsible for helping the CEA CH team with having a stronger mandate to deal with interpersonal harm, including some degree of acceptance of mistakes of overzealous moderation.
(all views my own) I want to publicly register what I’ve said privately for a while:
For people (usually but not always men) who we have considerable suspicion that they’ve been responsible for significant direct harm within the community, we should be significantly more willing than we currently are to take on more actions and the associated tradeoffs of limiting their ability to cause more harm in the community.
Some of these actions may look pretty informal/unofficial (gossip, explicitly warning newcomers against specific people, keep an unofficial eye out for some people during parties, etc). However, in the context of a highly distributed community with many people (including newcomers) that’s also embedded within a professional network, we should be willing to take more explicit and formal actions as well.
This means I broadly think we should increase our willingness to a) ban potentially harmful people from events, b) reduce grants we make to people in ways that increase harmful people’s power, c) warn organizational leaders about hiring people in positions of power/contact with potentially vulnerable people.
I expect taking this seriously to involve taking on nontrivial costs. However, I think this is probably worth it.
I’m not sure why my opinion here is different from others[1]’, however I will try to share some generators of my opinion, in case it’s helpful:
A. We should aim to be a community that’s empowered to do the most good. This likely entails appropriately navigating the tradeoff of both attempting to reducing the harms of a) contributors feeling or being unwelcome due to sexual harassment or other harms and b) contributors feeling or being unwelcome due to false accusations or overly zealous response.
B. I think some of this is fundamentally a sensitivity vs specificity tradeoff. If we have a detection system that’s too tuned to reduce the risk of false positives (wrong accusations being acted on), we will overlook too many false negatives (people being too slow to be banned/censured, or not at all), and vice versa.
Consider the first section of “Difficult Tradeoffs”
In the world we live in, I’ve yet to hear of a single incidence where, in full context, I strongly suspect CEA CH (or for that matter, other prominent EA organizations) was overzealous in recommending bans due to interpersonal harm. If our institutions are designed to only reduce first-order harm (both from direct interpersonal harm and from accusations), I’d expect to see people err in both directions.
Given the (apparent) lack of false positives, I broadly expect we accept too high a rate of false negatives. More precisely, I do not think CEA CH’s current work on interpersonal harm will lead to a conclusion like “We’ve evaluated all the evidence available for the accusations against X. We currently think there’s only a ~45% chance that X has actually committed such harms, but given the magnitude of the potential harm, and our inability to get further clarity with more investigation, we’ve pre-emptively decided to ban X from all EA Globals pending further evidence.”
Instead, I get the impression that substantially more certainty is deemed necessary to take action. This differentially advantages conservatism, and increases the probability and allowance of predatory behavior.
C. I expect an environment with more enforcement is more pleasant than an environment with less enforcement.
I expect an environment where there’s a default expectation of enforcement for interpersonal harm is more pleasant for both men and women. Most directly in reducing the first-order harm itself, but secondarily an environment where people are less “on edge” for potential violence is generally more pleasant. As a man, I at least will find it more pleasant to interact with women in a professional context if I’m not worried that they’re worried I’ll harm them. I expect this to be true for most men, and the loud worries online about men being worried about false accusations to be heavily exaggerated and selection-skewed[2].
Additionally, I note that I expect someone who exhibit traits like reduced empathy, willingness to push past boundaries, sociopathy, etc, to also exhibit similar traits in other domains. So someone who is harmful in (e.g.) sexual matters is likely to also be harmful in friendly and professional matters. For example, in the more prominent cases I’m aware of where people accused of sexual assault were eventually banned, they also appeared to have done other harmful activities like a systematic history of deliberate deception, being very nasty to men, cheating on rent, harassing people online, etc. So I expect more bans to broadly be better for our community.
D. I expect people who are involved in EA for longer to be systematically biased in both which harms we see, and which things are the relevant warning signals.
The negative framing here is “normalization of deviance”. The more neutral framing here is that people (including women) who have been around EA for longer a) may be systematically less likely to be targeted (as they have more institutional power and cachet) and b) are selection-biased to be less likely to be harmed within our community (since the people who have received the most harm are more likely to have bounced off).
E. I broadly trust the judgement of CEA CH in general, and Julia Wise in particular.
EDIT 2023/02: I tentatively withhold my endorsement until this allegation is cleared up.
I think their judgement is broadly reasonable, and they act well within the constraints that they’ve been given. If I did not trust them (e.g. if I was worried that they’d pursue political vendettas in the guise of harm-reduction), I’d be significantly more worried about given them more leeway to make mistakes with banning people.[3]
F. Nonetheless, the CEA CH team is just one group of individuals, and does a lot of work that’s not just on interpersonal harm. We should expect them to a) only have a limited amount of information to act on, and b) for the rest of EA to need to pick up some of the slack where they’ve left off.
For a), I think an appropriate action is for people to be significantly more willing to report issues to them, as well as make sure new members know about the existence of the CEA CH team and Julia Wise’s work within it. For b), my understanding is that CEA CH sees themself as having what I call a “limited purview”: e.g. they only have the authority to ban people from official CEA and maybe CEA-sponsored events, and not e.g. events hosted by local groups. So I think EA community-builders in a group organizing capacity should probably make it one of their priorities to be aware of the potential broken stairs in their community, and be willing to take decisive actions to reduce interpersonal harms.
Remember: EA is not a legal system. Our objective is to do the most good, not to wait to be absolutely certain of harm before taking steps to further limit harm.
One thing my post does not cover is opportunity cost. I mostly framed things as changing the decision-boundary. However, in practice I can see how having more bans is more costly in time and maybe money than the status quo. I don’t have good calculations here, however my intuition is strongly in the direction that having a safer and more cohesive is worth the relevant opportunity costs.
fwiw my guess is that the average person in EA leadership wishes the CEA CH team does more (is currently insufficiently punitive), rather than wish that they did less (is currently overzealous). I expect there’s significant variance in this opinion however.
This is a potential crux.
I can imagine this being a crux for people who oppose greater action. If so, I’d like to a) see this argument explicitly being presented and debated, and b) see people propose alternatives for reducing interpersonal harm that routes around CEA CH.
Thank you so much for laying out this view. I completely agree, including every single subpoint (except the ones about the male perspective which I don’t have much of an opinion on). CEA has a pretty high bar for banning people. I’m in favour of lowering this bar as well as communicating more clearly that the bar is really high and therefore someone being part of the community certainly isn’t evidence they are safe.
Thank you in particular for point D. I’ve never been quite sure how to express the same point and I haven’t seen it written up elsewhere.
It’s a bit unfortunate that we don’t seem to have agreevote on shortforms.
As an aside, I dislike calling out gender like this, even with the “not always” disclaimer. Compare: “For people (usually but not always black people)” would be considered inappropriate.
Would you prefer “mostly but not always?”
I think the archetypal examples of things I’m calling out is sexual harassment or abuse, so gender is unusually salient here.
I would prefer not to bring up gender at all. If someone commits sexual harassment, it doesn’t particularly matter what their gender is. And it may be true that men do it more than women, but that’s not really relevant, any more than it would be relevant if black people committed sexual harassment more than average.
It’s not that it “may be” true—it is true. I think it’s totally relevant: if some class of people are consistently the perpetuators of harm against another group, then surely we should be trying to figure out why that it is the case so we can stop it? Not providing that information seems like it could seriously impede our efforts to understand and address the problem (in this case, sexism & patriarchy).
I’m also confused by your analogy to race—I think you’re implying that it would be discriminatory to mention race if talking about other bad things being done, but I also feel like this is relevant. In this case I think it’s a bit different, however, as there’s other confounders present (e.g. black people are much more highly incarcerated, earn less on average, generally much less privileged) which all might increase rates of doing said bad thing. So in this case, it’s not a result of their race, but rather a result of the unequal socioeconomic conditions faced when someone is a certain race.