I think when I see the diagrams, I think of these as “low overhead roles” vs “high overhead roles”; where “low overhead roles” have peak marginal value much earlier than high overhead roles. If one is interested in scaling work, and assuming that requires also scaling labor, then scalable strategies would be ones that would have many low overhead roles, similar to your second diagram of “CEA in the Future”
That said, my main point above wasn’t that CEA should definitely grow, but that if CEA is having trouble/hesitancy/it-isn’t-ideal growing, I would expect that the strategy of “encouraging a bunch of new external nonprofits” to be limited in potential.
If CEA thinks it could help police new nonprofits, that would also take Max’s time or similar; the management time is coming from the same place, it’s just being used in different ways and there would ideally be less of it.
In the back of my mind, I’m thinking that OpenPhil theoretically has access to +$10Bil, and hypothetically much of this could go towards promotion of EA or EA-related principles, but right now there’s a big bottleneck here. I could imagine that it’s possible it could make sense to be rather okay wasting a fair bit of money and doing things quite unusual in order to get expansion to work somehow.
Around CEA and related organizations in particular, I am a bit worried that not all of the value of taking in good people is transparent. For example, if an org takes in someone promising and trains them up for 2 years, and then they leave for another org, that could have been a huge positive externality, but I’d bet it would get overlooked by funders. I’ve seen this happen previously. Right now it seems like there are a bunch of rather young EAs who really could use some training, but there are relatively few job openings, in part because existing orgs are quite hesitant to expand.
I imagine that hypothetically this could be an incredibly long conversation, and you definitely have a lot more inside knowledge than I do. I’d like to personally do more investigation to better understand what the main EA growth constraints are, we’ll see about this.
One thing we could make tractable progress in is in forecasting movement growth or these other things. I don’t have things in mind at the moment, but if you ever have ideas, do let me know, and we could see about developing them into questions in Metaculus or similar. I imagine having a group understanding of total EA movement growth could help a fair bit and make conversations like this more straightforward.
Thanks for the diagrams and explanation!
I think when I see the diagrams, I think of these as “low overhead roles” vs “high overhead roles”; where “low overhead roles” have peak marginal value much earlier than high overhead roles. If one is interested in scaling work, and assuming that requires also scaling labor, then scalable strategies would be ones that would have many low overhead roles, similar to your second diagram of “CEA in the Future”
That said, my main point above wasn’t that CEA should definitely grow, but that if CEA is having trouble/hesitancy/it-isn’t-ideal growing, I would expect that the strategy of “encouraging a bunch of new external nonprofits” to be limited in potential.
If CEA thinks it could help police new nonprofits, that would also take Max’s time or similar; the management time is coming from the same place, it’s just being used in different ways and there would ideally be less of it.
In the back of my mind, I’m thinking that OpenPhil theoretically has access to +$10Bil, and hypothetically much of this could go towards promotion of EA or EA-related principles, but right now there’s a big bottleneck here. I could imagine that it’s possible it could make sense to be rather okay wasting a fair bit of money and doing things quite unusual in order to get expansion to work somehow.
Around CEA and related organizations in particular, I am a bit worried that not all of the value of taking in good people is transparent. For example, if an org takes in someone promising and trains them up for 2 years, and then they leave for another org, that could have been a huge positive externality, but I’d bet it would get overlooked by funders. I’ve seen this happen previously. Right now it seems like there are a bunch of rather young EAs who really could use some training, but there are relatively few job openings, in part because existing orgs are quite hesitant to expand.
I imagine that hypothetically this could be an incredibly long conversation, and you definitely have a lot more inside knowledge than I do. I’d like to personally do more investigation to better understand what the main EA growth constraints are, we’ll see about this.
One thing we could make tractable progress in is in forecasting movement growth or these other things. I don’t have things in mind at the moment, but if you ever have ideas, do let me know, and we could see about developing them into questions in Metaculus or similar. I imagine having a group understanding of total EA movement growth could help a fair bit and make conversations like this more straightforward.