TLDR: We don’t have some easy to summarise methodology and being rational is pretty hard. Generally we try our best and hold ourselves and each other accountable and try to set up the community in a way that encourages rationality. If what you’re looking for is a list of techniques to be more rational yourself you could read this book of rationality advice or talk to people about why they prioritise what they do in a discussion group
Some meta stuff on why I think you got unsatisfactory answers to the other questions
I wouldn’t try to answer either of the previous questions because the answers seem big and definitely incomplete. I don’t have a quick summary for how I would resolve a disagreement with another EA because there are a bunch of overlapping techniques that can’t be described in a quick answer.
To put it into perspective I’d say the foundation to how I personally try to rationally approach EA is in the Rationality A-Z book but that probably doesn’t cover everything in my head and I definitely wouldn’t put it forward as a complete methodology for finding the truth. For a specific EA spin just talking to people about why they prioritise what they prioritise is what I’ve found most helpful and an easy way to do that is in EA discussion groups (in person is better than online).
It is pretty unfortunate that there isn’t some easy to summarise methodology or curriculum for applying rationality for charity current EA curricula are pretty focussed on just laying out our current best guess and using those examples along with discussion to demonstrate our methodology.
How is EA rational then?
I think the main thing happening in EA is that there is a strong personal, social, and financial incentive for people to approach their work “rationally”. E.g people in the community will expect you to have some reasoning which led you to do what you’re doing, and they’ll feedback on that reasoning if they think it’s missing an important consideration. From that spawns a bunch of people thinking about how to reason about this stuff more rationally, and we end up with a big set of techniques or concepts which seem to guide us better.
I was responding mainly to the format. I don’t expect you to get complete answers to your earlier two questions because there’s a lot more rationality methodology in EA than can be expressed in the amount of time I expect someone to spend on an answer
If I had to put my finger on why I don’t feel like the failure to answer those questions is as concerning to me as it seems to be for you I’d say because.
A) Just because it’s hard to answer doesn’t mean EAs aren’t holding themselves and each other to a high epistemic standard
B) Something about perfect not being the enemy of good and about urgency of other work. I want humanity to have some good universal epistemic tools but currently I don’t have them and I don’t really have the option to wait to do good until I have them. So I’ll just focus on the best thing my flawed brain sees to work on at the moment (using what fuzzy technical tools it has but still being subject to bias) because I don’t have any other machinery to use
I could be wrong but my read from your comments on other answers is that we disagree most on B). E.g you think current EA work would be better directed if we were able to have a lot more formally rational discussions. To the point that EA work or priorities should be put on hold (or slowed down) until we can do this.
I think I disagree with you on both A and B, as well as some other things. Would you like to have a serious, high-effort discussion about it and try to reach a conclusion?
TLDR: We don’t have some easy to summarise methodology and being rational is pretty hard. Generally we try our best and hold ourselves and each other accountable and try to set up the community in a way that encourages rationality. If what you’re looking for is a list of techniques to be more rational yourself you could read this book of rationality advice or talk to people about why they prioritise what they do in a discussion group
Some meta stuff on why I think you got unsatisfactory answers to the other questions
I wouldn’t try to answer either of the previous questions because the answers seem big and definitely incomplete. I don’t have a quick summary for how I would resolve a disagreement with another EA because there are a bunch of overlapping techniques that can’t be described in a quick answer.
To put it into perspective I’d say the foundation to how I personally try to rationally approach EA is in the Rationality A-Z book but that probably doesn’t cover everything in my head and I definitely wouldn’t put it forward as a complete methodology for finding the truth. For a specific EA spin just talking to people about why they prioritise what they prioritise is what I’ve found most helpful and an easy way to do that is in EA discussion groups (in person is better than online).
It is pretty unfortunate that there isn’t some easy to summarise methodology or curriculum for applying rationality for charity current EA curricula are pretty focussed on just laying out our current best guess and using those examples along with discussion to demonstrate our methodology.
How is EA rational then?
I think the main thing happening in EA is that there is a strong personal, social, and financial incentive for people to approach their work “rationally”. E.g people in the community will expect you to have some reasoning which led you to do what you’re doing, and they’ll feedback on that reasoning if they think it’s missing an important consideration. From that spawns a bunch of people thinking about how to reason about this stuff more rationally, and we end up with a big set of techniques or concepts which seem to guide us better.
Trying to address only one thing at a time:
I don’t think I asked for an “easy to summarise methodology” and I’m unclear on where that idea is coming from.
I was responding mainly to the format. I don’t expect you to get complete answers to your earlier two questions because there’s a lot more rationality methodology in EA than can be expressed in the amount of time I expect someone to spend on an answer
If I had to put my finger on why I don’t feel like the failure to answer those questions is as concerning to me as it seems to be for you I’d say because.
A) Just because it’s hard to answer doesn’t mean EAs aren’t holding themselves and each other to a high epistemic standard
B) Something about perfect not being the enemy of good and about urgency of other work. I want humanity to have some good universal epistemic tools but currently I don’t have them and I don’t really have the option to wait to do good until I have them. So I’ll just focus on the best thing my flawed brain sees to work on at the moment (using what fuzzy technical tools it has but still being subject to bias) because I don’t have any other machinery to use
I could be wrong but my read from your comments on other answers is that we disagree most on B). E.g you think current EA work would be better directed if we were able to have a lot more formally rational discussions. To the point that EA work or priorities should be put on hold (or slowed down) until we can do this.
I think I disagree with you on both A and B, as well as some other things. Would you like to have a serious, high-effort discussion about it and try to reach a conclusion?