I’ve been thinking a bunch about a fundamental difference between the EA community and the LessWrong community.
LessWrong is optimized for the enjoyment of its members. Any LessWrong event I go to in any city the focus is on “what will we find fun to do?” This is great. Notice how the community isn’t optimized for “making the world more rational.” It is a community that selects for people interested in rationality and then when you get these kinds of people in the same room the community tries to optimize for FUN for these kinds of people.
EA as a community is NOT optimized for the enjoyment of its members. It is optimized for making the world a better place. This is a feature, not a bug. And surely it should be net positive since its goal should by definition be net positive. When planning an EAG or EA event you measure it on impact and say professional connections made and how many new high quality AI Alignment researchers you might have created on the margin. You don’t measure it on how much people enjoyed themselves (or you do, but for instrumental reasons to get more people to come so that you can continue to have impact).
As a community organizer in both spaces, I do notice it is easier that I can leave EA events I organized feeling more burnt out and less fulfilled than compared to similar LW/ACX events. I think the fundamental difference mentioned before explains why.
Dunno if I am pointing at anything that resonates with anyone. I don’t see this discussed much among community organizers. Seems important to highlight.
Basically in LW/ACX spaces—specifically as an organizer—I more easily feel like a fellow traveller up for a good time. In EA spaces—specifically as an organizer—I more easily feel like an unpaid recruiter.
hmmmm I’d disagree based on my experience running GWWC events and EA Taskmaster but I’m probably not going in with the mindset of optimising for short term impact/urgency based on recruitment into AI safety.
I think getting the ratio of newbies / experienced EAs right can be hard so I try to think carefully about how to attract the “hard side” of the network (experienced EAs). My approach is more creating a fun environment where people who are currently doing direct work and/or donating significantly and effectively want to hang out or find useful.
My model is that significant behaviour change requires multiple positive interactions and time in between for reflection.
I certainly agree with your model on behaviour change. Likewise, my approach has over the years simplified from more convoluted ideas to one simple maxim: “Just make sure you feed them. The rest will often take care of itself.”
I’m concerned about animal welfare, human welfare and AI safety—without the urgency of AI dominating entirely.
I think what I highlight is similar to how many professional communities are optimized for matching prospective employers with employees rather than the happiness and enjoyment of their members. If there are 100 members but employers are only interested in one candidate you will have 99 less happy members. But this is not a bad thing as the goal of the community is to matching particular employers. It could easily be a mistake to find different employers and different events to make it more likely that you’ll have more happy members—risks include value drift and reducing your actual goal of maximizing impact. Still, 99% of your members are disgruntled as a tradeoff.
Professional-adjacent communities like say “computer tinkerers who just do it for fun” do not have this problem. If 99% in the community are not happy then you either change what you are doing to what the community of tinkerers are interested in or the community ceases to exist—or at least this is a much more likely outcome.
I think of there being a sort of spectrum for events, ranging from fairly focused and goal-oriented on one end to fairly enjoyable, for-its-own sake on the other end. Maybe the type of events that EAs tend to organize tend to be more on one side of that (simplified) spectrum, but there isn’t anything inherent that makes it so. I suspect that a lot of people tend to focus EA events on the recruiting, upskilling, professional networking, and knowledge-sharing that is likely to have a positive expected value. I for one would be very happy to see you organize some EA events with people interested in EA ideas, getting these people in the same room, and organizing something fun for these kinds of people.
A potluck dinner, a casual picnic, a movie night, a book club, or a hike could all involve people interested in EA topics, and without putting such a focus on making the world a better place. But maybe we shouldn’t even consider those EA events, because they are just regular events that happen to be attended by a bunch of EAs? If 10 EAs go for a hike, it isn’t necessarily an “EA hike.”
Some of the less enjoyable EA events I’ve gone to were let’s practice this EA-relevant skill in order to increase our competence and network for jobs and the more enjoyable were we are a bunch of EAs who want to hang out and play frisbee/talk about relationships/do some activity for fun & enjoyment. I do think that there is a place for structured, outcome-focused events for the EA community, but I don’t want 95% of events to be like that. Maybe 60%? Maybe 40%?
I’ve been thinking a bunch about a fundamental difference between the EA community and the LessWrong community.
LessWrong is optimized for the enjoyment of its members. Any LessWrong event I go to in any city the focus is on “what will we find fun to do?” This is great. Notice how the community isn’t optimized for “making the world more rational.” It is a community that selects for people interested in rationality and then when you get these kinds of people in the same room the community tries to optimize for FUN for these kinds of people.
EA as a community is NOT optimized for the enjoyment of its members. It is optimized for making the world a better place. This is a feature, not a bug. And surely it should be net positive since its goal should by definition be net positive. When planning an EAG or EA event you measure it on impact and say professional connections made and how many new high quality AI Alignment researchers you might have created on the margin. You don’t measure it on how much people enjoyed themselves (or you do, but for instrumental reasons to get more people to come so that you can continue to have impact).
As a community organizer in both spaces, I do notice it is easier that I can leave EA events I organized feeling more burnt out and less fulfilled than compared to similar LW/ACX events. I think the fundamental difference mentioned before explains why.
Dunno if I am pointing at anything that resonates with anyone. I don’t see this discussed much among community organizers. Seems important to highlight.
Basically in LW/ACX spaces—specifically as an organizer—I more easily feel like a fellow traveller up for a good time. In EA spaces—specifically as an organizer—I more easily feel like an unpaid recruiter.
hmmmm I’d disagree based on my experience running GWWC events and EA Taskmaster but I’m probably not going in with the mindset of optimising for short term impact/urgency based on recruitment into AI safety.
I think getting the ratio of newbies / experienced EAs right can be hard so I try to think carefully about how to attract the “hard side” of the network (experienced EAs). My approach is more creating a fun environment where people who are currently doing direct work and/or donating significantly and effectively want to hang out or find useful.
My model is that significant behaviour change requires multiple positive interactions and time in between for reflection.
Thank you for the perspective!
I certainly agree with your model on behaviour change. Likewise, my approach has over the years simplified from more convoluted ideas to one simple maxim: “Just make sure you feed them. The rest will often take care of itself.”
I’m concerned about animal welfare, human welfare and AI safety—without the urgency of AI dominating entirely.
I think what I highlight is similar to how many professional communities are optimized for matching prospective employers with employees rather than the happiness and enjoyment of their members. If there are 100 members but employers are only interested in one candidate you will have 99 less happy members. But this is not a bad thing as the goal of the community is to matching particular employers. It could easily be a mistake to find different employers and different events to make it more likely that you’ll have more happy members—risks include value drift and reducing your actual goal of maximizing impact. Still, 99% of your members are disgruntled as a tradeoff.
Professional-adjacent communities like say “computer tinkerers who just do it for fun” do not have this problem. If 99% in the community are not happy then you either change what you are doing to what the community of tinkerers are interested in or the community ceases to exist—or at least this is a much more likely outcome.
Informal and rambling thoughts:
I think of there being a sort of spectrum for events, ranging from fairly focused and goal-oriented on one end to fairly enjoyable, for-its-own sake on the other end. Maybe the type of events that EAs tend to organize tend to be more on one side of that (simplified) spectrum, but there isn’t anything inherent that makes it so. I suspect that a lot of people tend to focus EA events on the recruiting, upskilling, professional networking, and knowledge-sharing that is likely to have a positive expected value. I for one would be very happy to see you organize some EA events with people interested in EA ideas, getting these people in the same room, and organizing something fun for these kinds of people.
A potluck dinner, a casual picnic, a movie night, a book club, or a hike could all involve people interested in EA topics, and without putting such a focus on making the world a better place. But maybe we shouldn’t even consider those EA events, because they are just regular events that happen to be attended by a bunch of EAs? If 10 EAs go for a hike, it isn’t necessarily an “EA hike.”
Some of the less enjoyable EA events I’ve gone to were let’s practice this EA-relevant skill in order to increase our competence and network for jobs and the more enjoyable were we are a bunch of EAs who want to hang out and play frisbee/talk about relationships/do some activity for fun & enjoyment. I do think that there is a place for structured, outcome-focused events for the EA community, but I don’t want 95% of events to be like that. Maybe 60%? Maybe 40%?
I’d be curious about what kinds of events you’ve enjoyed organizing. Do you have a list, or could you type out a few sentences describing them?