I mean, I want employees at EA orgs to try to make EA orgs succeed insofar as that does the most good, and try to make EA orgs fail insofar as that does the most good instead. Likewise, I want them to try to strengthen the EA community if their model says this is good, and to try to weaken it (or just ignore it) otherwise.
(Obviously, in each case I’d want them to be open and honest about what they’re trying to do; you can oppose an org you think is bad without doing anything unethical or deceptive.)
I’m not sure what I think CEA’s role should be in EA. I do feel more optimistic about EA succeeding if major EA orgs in general focus more on developing a model of the world and trying to do the most good under their idiosyncratic world-view, rather than trying to represent or reflect EA-at-large; and I feel more optimistic about EA if sending our best and brightest to work at EA orgs doesn’t mean that they have to do massively more self-censoring now.
Maybe CEA or CEA-comms is an exception, but I’m not sold yet. I do think it’s good to have high epistemic standards, but I see that as compatible with expressing personal feelings, criticizing other orgs, wanting specific EA orgs to fail, etc.
I mean, I want employees at EA orgs to try to make EA orgs succeed insofar as that does the most good, and try to make EA orgs fail insofar as that does the most good instead. Likewise, I want them to try to strengthen the EA community if their model says this is good, and to try to weaken it (or just ignore it) otherwise.
(Obviously, in each case I’d want them to be open and honest about what they’re trying to do; you can oppose an org you think is bad without doing anything unethical or deceptive.)
I’m not sure what I think CEA’s role should be in EA. I do feel more optimistic about EA succeeding if major EA orgs in general focus more on developing a model of the world and trying to do the most good under their idiosyncratic world-view, rather than trying to represent or reflect EA-at-large; and I feel more optimistic about EA if sending our best and brightest to work at EA orgs doesn’t mean that they have to do massively more self-censoring now.
Maybe CEA or CEA-comms is an exception, but I’m not sold yet. I do think it’s good to have high epistemic standards, but I see that as compatible with expressing personal feelings, criticizing other orgs, wanting specific EA orgs to fail, etc.