Much EA value comes from being a Schelling point

Link post

TL;DR: A significant way in which the EA community creates value is by acting as a Schelling point where talented, ambitious, and altruistic people tend to gather and can meet each other (in addition to more direct sources of EA value like identifying the most important problems and directly pushing people to work on them). It might be useful to think about what optimising for being a Schelling point looks like, and I list some vague thoughts on that.


A Schelling point, also known as a focal point, is what people decide on in the absence of communication, especially when it’s important to coordinate by coming to the same answer.

The classic example is: you were arranging a meeting with a stranger in New York City by telephone, but you used the last minute of your phone credit and the line cut off after you had agreed on the date but not location or time—where do you meet? “Grand Central Station at noon” is an answer that other people may be especially likely to converge on.

(Schelling points can be thought of as a type of acausal negotiation.)

When the Schelling point is the selling point

Schelling points are often extremely powerful and valuable. A key function of top universities is to be Schelling points for talented people. (Personally, I’d call it the most important function.) There are other valuable things too: courses that go deeper, the signalling value to employers, and so on. However, talented people generally have a preference for hanging out with other talented people, both for social reasons and to find collaborators for ambitious projects and future colleagues. At the same time, talented people are also generally spread out and present only at low densities. Top universities select hard on (some measures of) talent, and through this create environments with high talent density. A big chunk of the reason why people apply to top universities is because other people do so too, and I’d guess that even if the academic standards of Stanford, MIT, or Cambridge eroded significantly, the fact that they’ve established themselves as congregating points for smart people will keep people applying and visiting for a long time.

(Note that this is related to, but not equal to, the prestige and status of these places. It is possible to imagine Schelling points that are not prestigious. For example, my impression is that this described MIT at one point—it became a congregating point for uniquely ambitious STEM students and defence research before it achieved high academic status. It is also possible to imagine prestigious places that are not Schelling points, though this is a bit harder since anything with prestige becomes a Schelling point for high social status (though prestige Schelling points and talent Schelling points need not co-occur). More generally, since prestige is a thing many people care a lot about, there is a high correlation between a place being prestigious or high status and being a Schelling point for at least some type of person. However, the mechanisms are distinct—a person selecting their university based on status is selecting based on what they get to write on their CV, while a person selecting their university based on it being a Schelling point for smart people is selecting based on the fact that many other smart people that they can’t coordinate with but would like to meet will also choose to go there.)

Another example is Silicon Valley. Sure, the area has many strengths—being rich and inside a large stable free market—but by far the greatest argument for living in Silicon Valley is that others also choose it. This leads to a (for now) unique combination of entrepreneurial people, great programmers, venture capitalists, and all the other types of people you need for a thriving tech business ecosystem, all there primarily because all the others are there too (how touching!). There’s a lot of value of having everything in one place, and it would be very hard for all the different people who make up the value of Silicon Valley to coordinate to move to another place. That’s why the Schelling point value of Silicon Valley is so enduring that people continue to tolerate large numbers of homeless drug addicts and sell kidneys to pay rent for years on end.

Note that a big part of the mechanism isn’t that specific people you want to find are there, but that the types of person you’d want to find are likely to also be there, because both those people and yourself are likely to converge on the strategy of going there.

Schelling EA

The Effective Altruism (EA) community provides a lot of value, for example:

  • research into figuring out what are the most important problems to solve to maximise human flourishing;

  • research and concrete efforts into how to solve the most important problems discovered by the above;

  • high epistemic standards and truth-seeking discussion norms;

  • a uniquely wide-ranging and well-reasoned set of resources to help people pursue high-impact careers;

  • tens of billions of dollars in funding.

However, in addition to these, a very critical part of the value that EA provides is being a Schelling point for talented, ambitious, and altruistically-motivated people.

Even without EA, there would be researchers studying existential risks, animal welfare, and global poverty; people trying to assess charities; communities with high epistemic norms; and billionaires trying to use their fortunes for effective good. However, thanks to EA, people in each of these categories can go to the same Effective Altruism Global conference or quickly find people in local groups, and meet collaborators, co-founders, funders, and so on. A lot of the reason why this can happen is that if you hang out with a certain group of people or on the right websites, EA looms large.

The biggest personal source of value I’ve gotten from EA has been having a shortcut to meeting people very high in all of talent, ambition, and altruistic motivation.

Much of this is obvious—breaking news: communities bring people together and foster connections, more at 11 - but I think taking seriously just how much of counterfactual EA community impact comes from being a Schelling point leads to some less-obvious points about possible implications.

Implications

The Schelling-point-based (and therefore necessarily incomplete) answer to “what is the EA community for?” might be something like “be an obvious Schelling point where relevant people gather, the chance of interactions that lead to useful work is maximised, and have a community and infrastructure that pushes work in the most useful direction possible”. (This is in contrast to answers that emphasise e.g. directly increasing the number of people working on the most pressing problems.) (I will not argue for this being the best possible answer; my point is just that it is one possible answer, and an interesting one to examine further.)

If I were a Big Tech marketing consultant, I might call this “EA-as-a-platform”.

What might maximising for such a Schelling point strategy look like?

Being obvious

A Schelling point is not a Schelling point unless it’s obvious enough. For EA to be an effective Schelling point for talented/​ambitious/​altruistic people, those people must hear about it. Silicon Valley is obvious enough that entrepreneurial people from South Africa to Russia hear about it and decide it’s where they want to be. To maximise its Schelling point value, EA should have world-spanning levels of recognition.

Note that recognition does not equal prestige or likeability. We don’t care (for Schelling point reasons at least) if most people hear about EA and go “eh, sounds weird and unappealing”; what matters is that the core target demographic is excited enough to put effort into pursuing EA. Consider how Silicon Valley was not particularly high-prestige in the public even when it was already attracting tech entrepreneurs, or how many people hear about the intensity of academics at top universities and (very reasonably) think “no thanks”.

Providing value

Though most of a Schelling point’s value typically comes from the other people who congregate at it, a Schelling point is easier to create if it is obviously valuable. Even though the smart people they meet might be most of the benefit of university, high schoolers are still more likely to go to top universities if they provide good education, good facilities, and unambiguous social status.

Some obvious ways in which EA provides value are through funding sufficiently promising projects, and by having a very high concentration of intellectually interesting ideas.

There are risks to communicating loudly about the value-add, since this brings in people who are in it purely for personal gain (“the vultures are circling”, as one Forum post put it). This works for Schelling points like Silicon Valley, but not altruism.

Optimising for matchmaking

A specific way that Schelling points provide value is by making it easy to meet other people in the specific ways that lead to productive teams forming. An existing example of this is that everyone says one-on-one meetings are the main point of conferences, and there is (of course) a lot of thinking about how to make these effective. On the more informal end of the scale, Reciprocity exists.

However, the scope and value of EA matchmaking could be expanded. I’m not aware of many ways to match together entrepreneurial teams (the Charity Entrepreneurship incubation program is the only one that comes to mind). I recently took part in an informally-organised co-founder matching process and found it extremely helpful to quickly get a lot of information on what it’s like to work together with several promising people.

I’d advise for someone to think more about how to make the EA environment even more effective at matching people who should know about each other. However, I expect someone is already designing a 53-parameter one-on-one matching system with Calendly, Slack, and Matplotlib integration for the next conference, and therefore I will hold off on adding any more fuel to this fire.

Being legit

One of the specific ways in which a Schelling point becomes one is if things associated with it seem uniquely competent, successful, or otherwise good, in a clearly unfakeable way. It is helpful for Cambridge’s Schelling point status that it can brag about having 121 Nobel laureates. That so many successful tech companies emerged from Silicon Valley specifically is an unfakeable signal. Any government or city can afford to throw some millions at putting up posters advertising its startup-friendliness; few can consistently produce multi-billion dollar tech companies.

No amount of community-building or image-crafting is likely to replicate the Schelling point power of obviously being the place where things happen. In some areas, I think EA already has such power: much of the research and work on existential risks happens within EA, and it might be hard to be a researcher on those topics without running into the large body of EA-originating work. However, EA goals require more than just research; note how being a project/​organisation founder or working in an operations role have been creeping up the 80 000 Hours list of recommended career paths.

It would be extremely powerful, not just for direct impact reasons but also for building up EA’s Schelling point status, if the EA community clearly spawned very obviously successful real-world projects. Alvea succeeding or working Nucleic Acid Observatories being built would be powerful examples. Likewise if Charity Entrepreneurship-incubated charities become clear stars of the non-profit world.

Meritocracy and impartial judgement

Right now, I think if a person somewhere in the world has a well-thought out idea for how to make the world a better place, likely their best bet to get a fair hearing, useful feedback, and—if it is competitive with the most valuable existing projects—funding and support is to post it on the EA Forum. I don’t think this is very obvious outside the EA community. However, this fact, and awareness of it, could make EA a more useful Schelling point, in the same way that the impression that Silicon Valley doesn’t frown on weird ideas as long as they’re important enough makes it a better Schelling point.

That EA endorses cause neutrality, has high and transparent epistemic standards, and a quantitative mindset are key parts of this. However, to use this to increase EA Schelling point power, these properties need to be clearly visible to outsiders.

The most likely way for this to be become more obvious might be if specific EA organisations achieved such a reputation widely within their field (and then there was some path by which knowing of these organisations points people towards knowing about EA).

GiveWell might be an example of a clearly-EA-linked organisation with visibly high epistemics and judgement quality, though I don’t know what their image or recognition level is outside the EA community. Another example is if someone created successful and famous organisations along the lines of FTX Future Fund’s proposed epistemic appeals process or widespread expert polling projects.

Openness and approachability

Good Schelling points are easy to enter, and don’t select on attributes that they don’t have to.

Every human sub-group, even if loose and purpose-driven, tends to develop a distinctive culture that is much more specific than strictly implied by its purpose. Sometimes this is useful, since it makes it easy for humans in even a loose group to bond with each other. However, a strong and distinct internal culture is also a barrier to entry. EA is already high-risk for having a strong barrier to entry, because

  • many arguments and concepts in EA require background knowledge to understand, and sometimes dense philosophical or technical background knowledge (and this is not the case just for more formal things like Forum posts; I’ve frequently heard “EV [expected value]”, “QALY [quality-adjusted life year”, and “Pascal’s mugging” assumed as obvious common terminology in casual conversation);

  • EA (quite obviously, given what it’s about) has a high concentration of non-obvious arguments that are obscure in public discussion but have huge implications; and

  • perhaps the main route into EA is caring very strongly about intellectual arguments about abstract moral principles, which tends not to be a natural way for humans to join communities.

These largely unavoidable factors already make EA somewhat unapproachable, and seem like a tightly-knit weird in-group/​subculture (anecdotally, this seems to be the most common complaint about EA among Cambridge students). Weird cultural norms or quirks are (among other things!) barriers to entry. Therefore, they should be minimised—to the extent that they can be without impinging on what EA is about—if the goal is to maximise Schelling point value.

(Mostly implicit) selectivity for the right things

Some selection is usually part of a Schelling point’s value. Top universities select for academic merit (though perhaps less so in the US). Silicon Valley selects for openness and interest/​talent in tech/​business. EA selects for openness, altruistic orientation (especially if consequentialist-leaning), good epistemics, and quantitative thinking.

I think it is counterproductive to view openness and selectivity as two ends of one scale that apply to everything. You want to select on important features and be open otherwise (note that, when creating a Schelling point, most of the selection is usually implicit—what types of people you attract—rather than explicit filtering). The key choice is not “open or selective overall?” but rather “for which X do we want to appeal only to people who have a value of X in some specific range?”

Here’s a heuristic for when selectivity for X is useful: when the way X provides value is through its concentration rather than its amount. If you’re at a party where you can only talk to a subset of the people during its course, you’re going to care a lot about what fraction of people there are interesting − 10 interesting people in a party of 20 is better than 50 in a party of 5000.

Some cases are ambiguous. For example, if there exists a way for the good and important research to bubble to the top regardless of how much other research exists, it seems like total amount of (infohazard-free) research is the thing to maximise. However, a research area where the average paper is very high quality might help newcomers to the field, or might help lift the prestige of the field, so concentration matters at least somewhat.

To take another example, there was a recent debate over whether EA Global should be open access. Many of the arguments against boil down to thinking the path to impact runs through a uniquely high concentration of EA engagement (or other variables) among the participants; arguments in favour are often either claiming that concentration matters less than sheer amount of interactions, or that the choice of selection variable(s) is wrong, or that CEA fails to select on their chosen selection variable(s) so even if the intention is right the selection variable selected for in practice is wrong.

Finally, a key point of a Schelling point is that it is a point somewhere. Here, EA is increasingly better. Berkeley, Cambridge, Oxford, London, and Berlin all have large groups, and offices that you can apply to in order to work on EA-relevant things in the company of other EAs.

In Schelling point terms, there’s also a risk that it might be better to have one really obvious and strong hub than many weaker ones (I’ve heard some Bay Area EAs in particular endorsing this view; invariably, their hub of choice is the Bay Area, though there is push back). In practice, it seems that many physical hubs but one virtual/​intellectual hub may be best. Both airplanes and people’s desires to not uproot their lives are real and relevant things.

The organisers at each EA hub might benefit from applying Schelling point thinking to the context of their local scene.

Being one thing

Finally, a Schelling point needs to be one thing, at least in some loose sense. If New York had two Grand Central Stations, the classic Schelling point game would become a lot harder to solve.

One way to increase the One Thingness of the EA Schelling point is to merge it with other things. In Schelling point land, “merging” does not mean making them the same cluster, but rather creating an obvious and visible path from one thing to another. My understanding is that increasing the obviousness of EA in somewhat-adjacent communities (tech, longevity, space, and Emergent Ventures grantees) was a large part of what Future Forum tried to achieve.


Thanks to Hugo Eberhard for feedback and discussion on drafts and the general topic.