SAN FRANCISCO, Sept 25 (Reuters) - ChatGPT-maker OpenAI is working on a plan to restructure its core business into a for-profit benefit corporation that will no longer be controlled by its non-profit board, people familiar with the matter told Reuters, in a move that will make the company more attractive to investors.
The case for a legal challenge seems hugely overdetermined to me:
Stop/delay/complicate the restructuring, and otherwise make life appropriately hard for Sam Altman
Settle for a large huge amount of money that can be used to do a huge amount of good
Signal that you can’t just blatantly take advantage of OpenPhil/EV/EA as you please without appropriate challenge
I know OpenPhil has a pretty hands-off ethos and vibe; this shouldn’t stop them from acting with integrity when hands-on legal action is clearly warranted
I understand that OpenAI’s financial situation is not very good [edit: this may not be a high-quality source], and if they aren’t able to convert to a for-profit, things will become even worse:
OpenAI has two years from the [current $6.6 billion funding round] deal’s close to convert to a for-profit company, or its funding will convert into debt at a 9% interest rate.
As an aside: how will OpenAI pay that interest in the event they can’t convert to a for-profit business? Will they raise money to pay the interest rate? Will they get a loan?
It’s conceivable that OpenPhil suing OpenAI could buy us 10+ years of AI timeline, if the following dominoes fall:
OpenPhil sues, and OpenAI fails to convert to a for-profit.
As a result, OpenAI struggles to raise additional capital from investors.
Losing $4-5 billion a year with little additional funding in sight, OpenAI is forced to make some tough financial decisions. They turn off the free version of ChatGPT, stop training new models, and cut salaries for employees. They’re able to eke out some profit, but not much profit, because their product is not highly differentiated from other AI offerings.
Silicon Valley herd mentality kicks in. OpenAI has been the hottest startup in the Valley. If it becomes known as the next WeWork, its fall will be earth-shaking. Game-theoretically, it doesn’t make as much sense to invest in an early AI startup round if there’s no capital willing to invest in subsequent rounds. OpenAI’s collapse could generate the belief that AI startups will struggle to raise capital—and if many investors believe that, it could therefore become true.
The AI bubble deflates and the Valley refocuses on other industries.
It would be extremely ironic if the net effect of all Sam Altman and Mark Zuckerberg’s efforts is to make AI companies uninvestable and buy us a bunch of timeline. Sam by generating a bunch of hype that fails to deliver, and Mark by commoditizing LLMs. (I say “ironic” because EAs are used to thinking of both Sam and Mark as irresponsible actors in the AI space.)
EDIT: there is some criticism of OpenPhil’s approach to its public image here which may be relevant to the decision of whether to sue or not. Also, there’s the obvious point that OpenAI appears to be one of the worst actors in the AI space.
EDIT 2: One also needs to consider how Sam might respond, e.g. by starting a new company and attempting to poach all OpenAI employees.
Media is often bought on a CPM basis (cost per thousand views). A display ad on LinkedIn for e.g. might cost $30 CPM. So yeah I think merch is probably underrated.
This is not surprising to me given the different historical funding situations in the relevant cause areas, the sense that animal-welfare and global-health are not talent-constrained as much as funding-constrained, and the clearer presence of strong orgs in those areas with funding gaps.
For instance:
there are 15 references to “upskill” (or variants) in the list of microgrants, and it’s often hard to justify an upskilling grant in animal welfare given the funding gaps in good, shovel-ready animal-welfare projects.
Likewise, 10 references to “study,” 12 to “development,′ 87 to “research” (although this can have many meanings), 17 for variants of “fellow,” etc.
There are 21 references to “part-time,” and relatively small, short blocks of time may align better with community building, small research projects than (e.g.) running a corporate campaign
Seems pretty unsurprising—the animal welfare fund is mostly giving to orgs, while the others give to small groups or individuals for upskilling/outreach frequently.
I think the differences between the LTFF and AWF are largely explained by differences in salary expectations/standards between the cause areas. There are small groups and individuals getting money from the AWF, and they tend to get much less for similar duration projects. Salaries in effective animal advocacy are pretty consistently substantially lower than in AI safety (and software/ML, which AI safety employers and grantmakers might try to compete with somewhat), with some exceptions. This is true even for work in high-income countries like the US and the UK. And, of course, salary expectations are even lower in low- and middle-income countries, which are an area of focus of the AWF (within neglected regions). Plus, many AI safety folks are in the Bay Area specifically, which is pretty expensive (although animal advocates in London also aren’t paid as much).
Yeah but my (implicit, should have made explicit lol) question is “why this is the case?”
Like at a high level it’s not obvious that animal welfare as a cause/field should make less use of smaller projects than the others. I can imagine structural explanations (eg older field → organizations are better developed) but they’d all be post hoc.
I think getting enough people interested in working on animal welfare has not usually been the bottleneck, relative to money to directly deploy on projects, which tend to be larger.
This doesn’t obviously point in the direction of relatively and absolutely fewer small grants, though. Like naively it would shrink and/or shift the distribution to the left—not reshape it.
I don’t understand why you think this is the case. If you think of the “distribution of grants given” as a sum of multiple different distributions (e.g. upskilling, events, and funding programmes) of significantly varying importance across cause areas, then more or less dropping the first two would give your overall distribution a very different shape.
How much do you trust other EA Forum users to be genuinely interested in making the world better using EA principles?
This is one thing I’ve updated down quite a bit over the last year.
It seems to me that relatively few self-identified EA donors mostly or entirely give to the organization/whatever that they would explicitly endorse as being the single best recipient of a marginal dollar (do others disagree?)
Of course the more important question is whether most EA-inspired dollars are given in such a way (rather than most donors). Unfortunately, I think the answer to this is “no” as well, seeing as OpenPhil continues to donate a majority of dollars to human global health and development[1] (I threw together a Claude artifact that lets you get a decent picture of how OpenPhil has funded cause areas over time and in aggregate)[2]
Edit: to clarify, it could be the case that others have object-level disagreements about what the best use of a marginal dollar is. Clearly this is sometimes the case, but it’s not what I am getting at here. I am trying to get at the phenomenon where people implicitly say/reason “yes, EA principles imply that the best thing to do would be to donate to X, but I am going to donate to Y instead.” I’m guessing this mostly takes the form of people failing to endorse that they’re donations are optimally directed rather than that they do a bunch of ground-up reasoning and then decide to ignore the conclusion it gives, though.
Data is a few days old and there’s a bit of judgement about how to bin various subcategories of grants, but I doubt the general picture would change much if others redid the analysis/binning
In your original post, you talk about explicit reasoning, in the your later edit, you switch to implicit reasoning. Feels like this criticism can’t be both. I also think the implicit reasoning critique just collapses into object-level disagreements, and the explicit critique just doesn’t have much evidence.
The phenomenon you’re looking at, for instance, is:
“I am trying to get at the phenomenon where people implicitly say/reason “yes, EA principles imply that the best thing to do would be to donate to X, but I am going to donate to Y instead.”
And I think this might just be an ~empty set, compared to people having different object-level beliefs about what EA principles are or imply they should do, and also disagree with you on what the best thing to do would be.[1] I really don’t think there’s many people saying “the bestthing to do is donate to X, but I will donate to Y”. (References please if so—clarification in footnotes[2]) Even on OpenPhil, I think Dustin just genuinely believes in worldview diversification is the best thing, so there’s no contradiction there where he implies the best thing would be to do X but in practice does do Y.
I think causing this to ‘update downwards’ on your views of the genuine interest of others—as opposed to, say, them being human and fallible despite trying to do the best they can—in the movement feels… well Jason used ‘harsh’, I might use a harsher word to describe this behavior.
I think maybe there might be a difference between the best thing (or best thing using simple calculations) and the right thing. I think people think in terms of the latter and not the former, and unless you buy into strong or even naïve consequentialism we shouldn’t always expect the two to go together
Thanks and I think your second footnote makes an excellent distinction that I failed to get across well in my post.
I do think it’s at least directionally an “EA principle” that “best” and “right” should go together, although of course there’s plenty of room for naive first-order calculation critiques, heuristics/intuitions/norms that might push against some less nuanced understanding of “best”.
I still think there’s a useful conceptual distinction to be made between these terms, but maybe those ancillary (for lack of a better word) considerations relevant to what one thinks is the “best” use of money blur the line enough to make it too difficult to distinguish these in practice.
Re: your last paragraph, I want to emphasize that my dispute is with the terms “using EA principles”. I have no doubt whatsoever about the first part, “genuinely interested in making the world better”
Thanks Aaron, I think you’re responses to me and Jason do clear things up. I still think the framing of it is a bit off though:
I accept that you didn’t intend your framing to be insulting to others, but using “updating down” about the “genuine interest” of others read as hurtful on my first read. As a (relative to EA) high contextualiser it’s the thing that stood out for me, so I’m glad you endorse that the ‘genuine interest’ part isn’t what you’re focusing on, and you could probably reframe your critique without it.
My current understanding of your position is that it is actually: “I’ve come to realise over the last year that many people in EA aren’t directing their marginal dollars/resources to the efforts that I see as most cost-effective, since I also think those are the efforts that EA principles imply are the most effective.”[1] To me, this claim is about the object-level disagreement on what EA principles imply.
However, in your response to Jason you say “it’s possible I’m mistaken over the degree to which direct resources to the place you think needs them most” is a consensus-EA principle which switches back to people not being EA? Or not endorsing this view? But you’ve yet to provide any evidence that people aren’t doing this, as opposed to just disagreeing about what those places are.[2]
Secondary interpretation is: “EA principles imply one should make a quantitative point estimate of the good of all your relevant moral actions, and then act on the leading option in a ‘shut-up-and-calculate’ way. I now believe many fewer actors in the EA space actually do this than I did last year”
For example, in Ariel’s piece, Emily from OpenPhil implies that they have much lower moral weights on animal life than Rethink does, not that they don’t endorse doing ‘the most good’ (I think this is separable from OP’s commitment to worldview diversification).
This is one thing I’ve updated down quite a bit over the last year.
It seems a bit harsh to treat other user-donors’ disagreement with your views on concentrating funding on their top-choice org (or even cause area) as significant evidence against the proposition that they are “genuinely interested in making the world better using EA principles.”
It seems to me that relatively few self-identified EA donors mostly or entirely give to the organization/whatever that they would explicitly endorse as being the single best recipient of a marginal dollar (do others disagree?)
I think a world in which everyone did this would have some significant drawbacks. While I understand how that approach would make sense through an individual lens, and am open to the idea that people should concentrate their giving more, I’d submit that we are trying to do the most good collectively. For instance: org funding is already too concentrated on a too-small number of donors. If (say) each EA is donating to an average of 5 orgs, then a norm of giving 100% to a single org would decrease the number of donors by 80%. That would impose significant risks on orgs even if their total funding level was not changed.
It’s also plausible that the number of first-place votes an org (or even a cause area) would get isn’t a super-strong reflection of overall community sentiment. If a wide range of people identified Org X as in their top 10%, then that likely points to some collective wisdom about Org X’s cost-effectiveness even if no one has them at number 1. Moreover, spreading the wealth can be seen as deferring to broader community views to some extent—which could be beneficial insofar as one found little reason to believe that wealthier community members are better at deciding where donation dollars should go than the community’s collective wisdom. Thus, there are reasons—other than a lack of genuine interest in EA principles by donors—that donors might reasonably choose to act in accordance with a practice of donation spreading.
Thanks, it’s possible I’m mistaken over the degree to which “direct resources to the place you think needs them most” is a consensus-EA principle.
Also, I recognize that “genuinely interested in making the world better using EA principles” is implicitly value-laden, and to be clear I do wish it was more the case, but I also genuinely intend my claim to be an observation that might have pessimistic implications depending on other beliefs people may have rather than an insult or anything like it, if that makes any sense.
In Q1 we will launch a GPT builder revenue program. As a first step, US builders will be paid based on user engagement with their GPTs. We’ll provide details on the criteria for payments as we get closer.
Interesting lawsuit; thanks for sharing! A few hot (unresearched, and very tentative) takes, mostly on the Musk contract/fraud type claims rather than the unfair-competition type claims related to x.ai:
One of the overarching questions to consider when reading any lawsuit is that of remedy. For instance, the classic remedy for breach of contract is money damages . . . and the potential money damages here don’t look that extensive relative to OpenAI’s money burn.
Broader “equitable” remedies are sometimes available, but they are more discretionary and there may be some significant barriers to them here. Specifically, a court would need to consider the effects of any equitable relief on third parties who haven’t done anything wrongful (like the bulk of OpenAI employees, or investors who weren’t part of an alleged conspiracy, etc.), and consider whether Musk unreasonably delayed bringing this lawsuit (especially in light of those third-party interests). On hot take, I am inclined to think these factors would weigh powerfully against certain types of equitable remedies.
Stated more colloquially, the adverse effects on third parties and the delay (“laches”) would favor a conclusion that Musk will have to be content with money damages, even if they fall short of giving him full relief.
Third-party interests and delay may be less of a barrier to equitable relief against Altman himself.
Musk is an extremely sophisticated party capable of bargaining for what he wanted out of his grants (e.g., a board seat), and he’s unlikely to get the same sort of solicitude on an implied contract theory that an ordinary individual might. For example, I think it was likely foreseeable in 2015 to January 2017 -- when he gave the bulk of the funds in question—that pursuing AGI could be crazy expensive and might require more commercial relationships than your average non-profit would ever consider. So I’d be hesitant to infer much in the way of implied-contractual constraints on OpenAI’s conduct than section 501(c)(3) of the Internal Revenue Code and California non-profit law require.
The fraud theories are tricky because the temporal correspondence between accepting the bulk of the funds and the alleged deceit feels shaky here. By way of rough analogy, running up a bunch of credit card bills you never intended to pay back is fraud. Running up bills and then later deciding that you aren’t going to pay them back is generally only a contractual violation. I’m not deep into OpenAI drama, but a version of the story in which the heel turn happened later in the game than most/all of Musk’s donations and assistance seems plausible to me.
I’m pretty happy with how this “Where should I donate, under my values?” Manifold market has been turning out. Of course all the usual caveats pertaining to basically-fake “prediction” markets apply, but given the selection effects of who spends manna on an esoteric market like this I put a non-trivial weight into the (live) outcomes.
I guess I’d encourage people with a bit more money to donate to do something similar (or I guess defer, if you think I’m right about ethics!), if just as one addition to your portfolio of donation-informing considerations.
Thanks! Let me write them as a loss function in python (ha)
For real though:
Some flavor of hedonic utilitarianism
I guess I should say I have moral uncertainty (which I endorse as a thing) but eh I’m pretty convinced
Longtermism as explicitly defined is true
Don’t necessarily endorse the cluster of beliefs that tend to come along for the ride though
“Suffering focused total utilitarian” is the annoying phrase I made up for myself
I think many (most?) self-described total utilitarians give too little consideration/weight to suffering, and I don’t think it really matters (if there’s a fact of the matter) whether this is because of empirical or moral beliefs
Maybe my most substantive deviation from the default TU package is the following (defended here):
“Under a form of utilitarianism that places happiness and suffering on the same moral axis and allows that the former can be traded off against the latter, one might nevertheless conclude that some instantiations of suffering cannot be offset or justified by even an arbitrarily large amount of wellbeing.”
Moral realism for basically all the reasons described by Rawlette on 80k but I don’t think this really matters after conditioning on normative ethical beliefs
Nothing besides valenced qualia/hedonic tone has intrinsic value
I think that might literally be it—everything else is contingent!
According to Kevin Esvelt on the recent 80,000k podcast (excellent btw, mostly on biosecurity), eliminating the New World New World screwwormcould be an important farmed animal welfare (infects livestock), global health (infects humans), development (hurts economies), science/innovation intervention, and most notably quasi-longtermist wild animal suffering intervention.
More, if you think there’s a non-trivial chance of human disempowerment, societal collapse, or human extinction in the next 10 years, this would be important to do ASAP because we may not be able to later.
From the episode:
Kevin Esvelt:
...
But from an animal wellbeing perspective, in addition to the human development, the typical lifetime of an insect species is several million years. So 106 years times 109 hosts per year means an expected 1015 mammals and birds devoured alive by flesh-eating maggots. For comparison, if we continue factory farming for another 100 years, that would be 1013 broiler hens and pigs. So unless it’s 100 times worse to be a factory-farmed broiler hen than it is to be devoured alive by flesh-eating maggots, then when you integrate over the future, it is more important for animal wellbeing that we eradicate the New World screwworm from the wild than it is that we end factory farming tomorrow.
Dropping a longer quote with more context in this footnote.[1] A quick Google Images search makes this all the more visceral, but be warned that it’s kinda graphic.
I would really love to see someone (maybe me) do a deeper dive into this and write up a proper Forum post.
Kevin Esvelt: So the fourth one might actually be the easiest to get going: the New World screwworm, which has the amazing scientific name of Cochliomyia hominivorax: “the man devourer.” But it doesn’t primarily eat humans; it feeds indiscriminately on warm-blooded things, so mammals and birds. It’s a botfly that lays its eggs in open wounds, anything as small as a tick bite. And it’s called the screwworm because the larvae are screw-shaped and they drill their way into living flesh, devouring it. And as they do, they cultivate bacteria that attract new gravid females that lay more eggs and continue the cycle.
So you have this macabre dance of parasitisation that results in the animal being devoured alive by flesh-eating maggots. And we know that it’s horrendously painful, because people get affected by this, and the standard of treatment is you give them morphine immediately so that surgeons can cut the things out — because it’s just that painful; it’s unbelievably agonising. And by my back-of-the-envelope calculations, there’s about a billion hosts of this every year — so a billion animals are devoured alive by flesh-eating maggots every single year.
We even know that we can eradicate this species from at least many ecosystems and not see any effects, because it used to be present in North America too, and we wiped it out using nuclear technology, oddly enough. Some clever folks noticed if you irradiate the larvae, then they grow up sterile. And if you release enough of them, then the wild ones will mate with a sterile one, and they only mate once, so you can suppress the population to the point of not being there anymore.
So we did this first up through Florida and then across the West, and then down through Texas to the Mexican border. The US Department of Agriculture then inked a deal with the Mexican government to eradicate them from Mexico because the southern border was shorter and therefore cheaper. And then they just went country by country down Central America to Panama. The southern border of Panama is the shortest, so American taxpayer dollars today contribute to the creation and maintenance of a living wall of sterile screwworm flies released in southern Panama that prevents the South American screwworm from reinvading North America — 10 million released every week.
Luisa Rodriguez: Wow.
Kevin Esvelt: But there’s too many of them in South America to wipe out by that means. And so the way forward is obviously gene drive. If the Mercosur countries agree that they want to get rid of the New World screwworm, they can start with something like a daisy drive locally — and Uruguay is working on this — then they can wipe it out from their country. Uruguay loses about 0.1% of their total country’s GDP to the screwworm because they’re so dependent on animal exports. I mean, Uruguay and beef is… To those listeners who eat beef, I’m going to start fights here, but it’s better than beef from Argentina, even. But anyway, they’re all very concerned about their beef, and screwworm is horrific.
It also, of course, preferentially hurts poor farmers who struggle to afford the veterinary treatments for their animals. And of course, they hate to see it, because here you’re watching these animals that you’re caring for literally get devoured by flesh-eating maggots, and it’s agonisingly painful.
But from an animal wellbeing perspective, in addition to the human development, the typical lifetime of an insect species is several million years. So 106 years times 109 hosts per year means an expected 1015 mammals and birds devoured alive by flesh-eating maggots. For comparison, if we continue factory farming for another 100 years, that would be 1013 broiler hens and pigs. So unless it’s 100 times worse to be a factory-farmed broiler hen than it is to be devoured alive by flesh-eating maggots, then when you integrate over the future, it is more important for animal wellbeing that we eradicate the New World screwworm from the wild than it is that we end factory farming tomorrow.
1) How often (in absolute and relative terms) a given forum topic appears with another given topic
2) Visualizing the popularity of various tags
An updated Forum scrape including the full text and attributes of 10k-ish posts as of Christmas, ’22
See the data without full text in Google Sheets here
Post explaining version 1.0 from a few months back
From the data in no. 2, a few effortposts that never garnered an accordant amount of attention (qualitatively filtered from posts with (1) long read times (2) modest positive karma (3) not a ton of comments.
I think the proxy question is “after what period of time is it reasonable to assume that any work building or expanding on the post would have been published?” and my intuition here is about 1 year but would be interested in hearing others thoughts
Eh I’m not actually sure how bad this would be. Of course it could be overdone, but a post’s author is its obvious best advocate, and a simple “I think this deserves more attention” vote doesn’t seem necessarily illegitimate to me
This post is half object level, half experiment with “semicoherent audio monologue ramble → prose” AI (presumably GPT-3.5/4 based) program audiopen.ai.
In the interest of the latter objective, I’m including 3 mostly-redundant subsections:
A ’final’ mostly-AI written text, edited and slightly expanded just enough so that I endorse it in full (though recognize it’s not amazing or close to optimal)
The raw AI output
The raw transcript
1) Dubious asymmetry argument in WWOTF
In Chapter 9 of his book, What We Are the Future, Will MacAskill argues that the future holds positive moral value under a total utilitarian perspective. He posits that people generally use resources to achieve what they want—either for themselves or for others—and thus good outcomes are easily explained as the natural consequence of agents deploying resources for their goals. Conversely, bad outcomes tend to be side effects of pursuing other goals. While malevolence and sociopathy do exist, they are empirically rare.
MacAskill argues that in a future with continued economic growth and no existential risk, we will likely direct more resources towards doing good things due to self-interest and increased impartial altruism. He contrasts this eutopian scenario with an anti-eutopia: the worst possible world, which he argues (compellingly, I think) less probable because it requires convoluted explanations as opposed to simple desires like enjoying ice cream. He concludes that the probability of achieving a eutopia outweighs the low likelihood but extreme negative consequences of an anti-eutopia.
However, I believe McCaskill’s analysis neglects an important aspect: considering not only these two extremes but also the middle distribution where neither significant amounts of resources nor agentic intervention occur.
When physics operates without agency-driven resource allocation, we have good reason to expect evolution to create conscious beings whose suffering we can attribute to the ease with which animals (or animal-like beings) can lose all future expected genetic reproduction—as MacAskill himself argues elsewhere in the book.
Importantly, though, this non-agentic suffering, seems more likely to complement agentic resource deployment—not substitute for it as one might intuit. That’s because human or post-human expansion necessarily entails the expansion of concentrated physical energy, and seems likely to entail the expansion of other scarce, pro-biotic resources such as DNA, water, and computation.
Although McCaskill does not explicitly claim his binary model comparing eutopia and anti-eutopia is sufficient for understanding this complex problem, it seems to me to be implied;
Only upon attempting to draft a blog post revisiting his work did I noticed the line “We can make some progress by focusing on just two extreme scenarios: the best or worst possible futures, eutopia and anti-eutopia,” acknowledging status of this model as a bit of evidence in the larger question of the future’s value.
In sum, I think MacAskill’s analysis would benefit from addressing the morally relevant middle distribution to provide a more accurate representation of the future under total utilitarianism.
2) Utilitarian Utopia, Anti-Utopia and the Neglected Middle
In Chapter 9 of his book, What We Are the Future, Will McCaskill argues that the future holds positive moral value under a total utilitarian perspective. He posits that people generally use resources to achieve what they want—either for themselves or others—and thus good outcomes are often intentional. Conversely, bad outcomes tend to be side effects of pursuing other goals. While malevolence and sociopathy do exist, they are empirically rare.
McCaskill then extrapolates this argument to suggest that in a future with continued economic growth (assuming no existential risk), we will likely direct more resources towards doing good things due to self-interest and increased impartial altruism. He contrasts this utopian scenario with an anti-utopia: the worst possible world which is less probable because it requires convoluted explanations as opposed to simple desires like enjoying ice cream.
He concludes that the probability of achieving a utopia outweighs the low likelihood but extreme negative consequences of an anti-utopia. However, I believe McCaskill’s analysis neglects an important aspect: considering not only these two extremes but also the middle distribution where neither significant amounts of resources nor agentic intervention occur.
In such cases where physics operates without agency-driven resource allocation, evolution can create conscious beings like plants and animals who experience suffering without any intentionality behind it. This middle distribution may actually skew negatively since evolution favors reproductive fitness at any cost; as a result, sentient beings could suffer more than they experience happiness during their lives.
I argue that wild animal welfare is net negative overall; if given a choice between having them continue living or ceasing their existence altogether, I would choose nonexistence on moral grounds. Although McCaskill does not explicitly claim his heuristic comparison between utopia and anti-utopia is sufficient for understanding this complex problem, he strongly implies it throughout most of his chapter.
Upon revisiting his work while drafting my response blog post, I noticed a single line acknowledging the limitations of his approach. However, this caveat does not adequately emphasize that his argument should be considered only as a first pass and not definitive proof. In my opinion, McCaskill’s analysis would benefit from addressing the morally relevant middle distribution to provide a more accurate representation of the future under total utilitarianism.
-------
3) Original Transcript
Okay, so I’m going to describe where I think I disagree with Will McCaskill in Chapter 9 of his book, What We Are the Future, where he basically makes an argument that the future is positive in expectation, positive moral value under a total utilitarian perspective. And so his argument is basically that people, it’s very easy to see that people deploy the resources in order to get what they want, which is either to help themselves and sometimes to help other people, whether it’s just their family or more impartial altruism. Basically you can always explain why somebody does something good just because it’s good and they want it, which is kind of, I think that’s correct and compelling. Whereas when something bad happens, it’s generally the side effect of something else. At least, yeah. So while there is malevolence and true sociopathy, those things are in fact empirically quite rare, but if you undergo a painful procedure, like a medical procedure, it’s because there’s something affirmative that you want and that’s a necessary side effect. It’s not because you actually sought that out in particular. And all this I find true and correct and compelling. And so then he uses this to basically say that in the future, presumably conditional on continued economic growth, which basically just means no existential risk and humans being around, we’ll be employing a lot of resources in the direction of doing things well or doing good. Largely just because people just want good things for themselves and hopefully to some extent because there will be more impartial altruists willing to both trade and to put their own resources in order to help others. And once again, all true, correct, compelling in my opinion. So on the other side, so basically utopia in this sense, utopia basically meaning employing a lot of, the vast majority of resources in the direction of doing good is very likely and very good. On the other side, it’s how likely and how bad is what he calls anti-utopia, which is basically the worst possible world. And he basically using… I don’t need to get into the particulars, but basically I think he presents a compelling argument that in fact it would be worse than the best world is good, at least to the best of our knowledge right now. But it’s very unlikely because it’s hard to see how that comes about. You actually can invent stories, but they get kind of convoluted. And it’s not nearly as simple as, okay, people like ice cream and so they buy ice cream. It’s like, you have to explain why so many resources are being deployed in the direction of doing good things and you still end up with a terrible world. Then he basically says, okay, all things considered, the probability of good utopia wins out relative to the badness, but very low probability of anti-utopia. Again, a world full of misery. And where I think he goes wrong is that he neglects the middle of the distribution where the distribution is ranging from… I don’t know how to formalize this, but something like percentage or amount of… Yeah, one of those two, percentage or amount of resources being deployed in the direction of on one side of the spectrum causing misery and then the other side of the spectrum causing good things to come about. And so he basically considers the two extreme cases. But I claim that, in fact, the middle of the distribution is super important. And actually when you include that, things look significantly worse because the middle of the distribution is basically like, what does the world look like when you don’t have agents essentially deploying resources in the direction of anything? You just have the universe doing its thing. We can set aside the metaphysics or physics technicalities of where that becomes problematic. Anyway, so basically the middle of the distribution is just universe doing its thing, physics operating. I think there’s the one phenomenon that results from this that we know of to be morally important or we have good reason to believe is morally important is basically evolution creating conscious beings that are not agentic in the sense that I care about now, but basically like plants and animals. And presumably I think you have good reason to believe animals are sentient. And evolution, I claim, creates a lot of suffering. And so you look at the middle of the distribution and it’s not merely asymmetrical, but it’s asymmetrical in the opposite direction. So I claim that if you don’t have anything, if you don’t have lots of resources being deployed in any direction, this is a bad world because you can expect evolution to create a lot of suffering. The reason for that is, as he gets into, something like either suffering is intrinsically more important, which I put some weight on that. It’s not exactly clear how to distinguish that from the empirical case. And the empirical case is basically it’s very easy to lose all your reproductive fitness in the evolutionary world very quickly. It’s relatively hard to massively gain a ton. Reproduction is like, even having sex, for example, only increases your relative reproductive success a little bit, whereas you can be killed in an instant. And so this creates an asymmetry where if you buy a functional view of qualia, then it results in there being an asymmetry where animals are just probably going to experience more pain over their lives, by and large, than happiness. And I think this is definitely true. I think wild animal welfare is just net negative. I wish if I could just… If these are the only two options, have there not be any wild animals or have them continue living as they are, I think it would be overwhelmingly morally important to not have them exist anymore. And so tying things back. Yeah, so McCaskill doesn’t actually… I don’t think he makes a formally incorrect statement. He just strongly implies that this case, that his heuristic of comparing the two tails is a pretty good proxy for the best we can do. And that’s where I disagree. I think there’s actually one line in the chapter where he basically says, we can get a grip on this very hard problem by doing the following. But I only noticed that when I went back to start writing a blog post. And the vast majority of the chapter is basically just the object level argument or evidence presentation. There’s no repetition emphasizing that this is a really, I guess, sketchy, for lack of a better word, dubious case. Or first pass, I guess, is a better way of putting it. This is just a first pass, don’t put too much weight on this. That’s not how it comes across, at least in my opinion, to the typical reader. And yeah, I think that’s everything.
I think there’s a case to be made for exploring the wide range of mediocre outcomes the world could become.
Recent history would indicate that things are getting better faster though. I think MacAskill’s bias towards a range of positive future outcomes is justified, but I think you agree too.
Maybe you could turn this into a call for more research into the causes of mediocre value lock-in. Like why have we had periods of growth and collapse, why do some regions regress, what tools can society use to protect against sinusoidal growth rates.
I didn’t have a good response to @DanielFilan, and I’m pretty inclined to defer to orgs like CEA to make decisions about how to use their own scarce resources.
At least for EA Global Boston 2024 (which ended yesterday), there was the option to pay a “cost covering” ticket fee (of what I’m told is $1000).[1]
All this is to say that I am now more confident (although still <80%) that marginal rejected applicants who are willing to pay their cost-covering fee would be good to admit.[2]
In part this stems from an only semi-legible background stance that, on the whole, less impressive-seeming people have more ~potential~ and more to offer than I think “elite EA” (which would those running EAG admissions) tend to think. And this, in turn, has a lot to do with the endogeneity/path dependence of I’d hastily summarize as “EA involvement.”
That is, many (most?) people need a break-in point to move from something like “basically convinced that EA is good, interested in the ideas and consuming content, maybe donating 10%” to anything more ambitious.
For some, that comes in the form of going to an elite college with a vibrant EA group/community. Attending EAG is another—or at least could be. But if admission is dependent on doing the kind of things and/or having the kinds of connections that a person might only pursue after getting on such an on-ramp, you have a vicious cycle of endogenous rejection.
The impetus for writing this is seeing a person who was rejected with some characteristics that seem plausibly pretty representative of a typical marginal EAG rejectee:
College educated but not via an elite university
Donates 10%, mostly to global health
Normal-looking middle or upper-middle class career
Interested in EA ideas but not a huge amount to show for it
Never attended EAG
Of course n=1, this isn’t a tremendous amount of evidence, I don’t have strictly more information than the admissions folks, the optimal number of false-negatives is not zero, etc., etc. But if a person with those above characteristics who is willing to write a reasonably thoughtful application and spend their personal time and money traveling to and taking part in EA Global (and, again, covering their cost)[3] is indeed likely to get rejected, I just straightforwardly think that admission has too high a bar; does CEA really think such a person is actively harmful to the event on net?
I don’t want to say that there is literally zero potential downside from admitting more people and “diluting the attendee pool” for lack of a more thoughtful term, but it’s not immediately obvious to me what that downside would be especially at the current margin (say for example a 25% increase in the number of attendees, not a 2000% increase). And, needless to say, there is a lot of potential upside via both the impact of this marginal attendee themselves and via the information/experience/etc. that they bring to the whole group.
If this were an econ paper, I’d probably want to discuss the fiscal relevance of marginal vs average cost-covering tickets. I suspect that the “cost covering ticket” actually advertised is based on the average cost, but I’m not sure.
If this is true, and marginal cost < average cost as seems intuitive, then admitting a marginal attendee who then pays the average cost would be financially net-positive for CEA.
That is, many (most?) people need a break-in point to move from something like “basically convinced that EA is good, interested in the ideas and consuming content, maybe donating 10%” to anything more ambitious.
I am under the impression that EAGx can be such a break-in point, and has lower admission standards than EAG. In particular, there is EAGxVirtual (Applications are open!).
Has the rejected person you are thinking of applied to any EAGx conference?
I agree. One minor issue with your “low bar” is the giving 10 percent. Giving this much is extremely uncommon to any cause, so for me might be more of a “medium bar” ;)
Would this (generally) be a one-time deal? The idea that some people would benefit from a bolus of EA as a “break-in point” or “on-ramp” seems plausible, and willingness to pay a hefty admission fee / other expenses would certainly have a signaling value.[1] However, the argument probably gets weaker after the first marginal admission (unless the marginal applicant is a lot closer to the line on the second time around).
Maybe allowing only one marginal admission per person absent special circumstances would mitigate concerns about “diluting” the event.
I recognize the downsides of a pay-extra-to-attend approach as far as perceived fairness, equity, accessibility to people from diverse backgrounds, and so on. That would be a tradeoff to consider.
Random sorta gimmicky AI safety community building idea: tabling at universities but with a couple laptop signed into Claude Pro with different accounts. Encourage students (and profs) to try giving it some hard question from eg a problem set and see how it performs. Ideally have a big monitor for onlookers to easily see.
Most college students are probably still using ChatGPT-3.5, if they use LLMs at all. There’s a big delta now between that and the frontier.
A (potential) issue with MacAskill’s presentation of moral uncertainty
Not able to write a real post about this atm, though I think it deserves one.
MacAskill makes a similar point in WWOTF, but IMO the best and most decision-relevant quote comes from his second appearance on the 80k podcast:
There are possible views in which you should give more weight to suffering...I think we should take that into account too, but then what happens? You end up with kind of a mix between the two, supposing you were 50⁄50 between classical utilitarian view and just strict negative utilitarian view. Then I think on the natural way of making the comparison between the two views, you give suffering twice as much weight as you otherwise would.
I don’t think the second bolded sentence follows in any objective or natural manner from the first. Rather, this reasoning takes a distinctly total utilitarian meta-level perspective, summing the various signs of utility and then implicitly considering them under total utilitarianism.
Even granting that the mora arithmetic is appropriate and correct, it’s not at all clear what to do once the 2:1 accounting is complete. MacAskill’s suffering-focused twin might have reasoned instead that
Negative and total utilitarianism are both 50% likely to be true, so we must give twice the normal amount of weight to happiness. However, since any sufficiently severe suffering morally outweighs any amount of happiness, the moral outlook on a world with twice as much wellbeing is the same as before
A better proxy for genuine neutrality (and the best one I can think of) might be to simulate bargaining over real-world outcomes from each perspective, which would probably result in at least some proportion of one’s resources being deployed as though negative utilitarianism were true (perhaps exactly 50%, though I haven’t given this enough thought to make the claim outright).
The recent 80k podcast on the contingency of abolition got me wondering what, if anything, the fact of slavery’s abolition says about the ex ante probability of abolition—or more generally, what one observation of a binary random variable X says about p as in
Turns out there is an answer (!), and it’s found starting in paragraph 3 of subsection 1 of section 3 of the Binomial distribution Wikipedia page:
Don’t worry, I had no idea what Beta(α,β) was until 20 minutes ago. In the Shortform spirit, I’m gonna skip any actual explanation and just link Wikipedia and paste this image (I added the uniform distribution dotted line because why would they leave that out?)
So...
Cool, so for the n=1 case, we get that if you have a prior over the ex ante probability space[0,1] described by one of those curves in the image, you...
0) Start from ‘zero empirical information guesstimate’ E[Beta(α,β)]=αα+β
1a) observe that the thing happens (x=1), moving you, Ideal Bayesian Agent, to updated probability ^pb=1+α1+α+β>αα+β OR
1b) observe that the thing doesn’t happen (x=0), moving you to updated probability ^pb=α1+α+β<αα+β
In the uniform case (which actually seems kind of reasonable for abolition), you...
0) Start from prior E[p]=1/2
1a) observe that the thing happens, moving you to updated probability ^pb=2/3
1a) observe that the thing doesn’thappen, moving you to updated probability ^pb=13
In terms of result, yeah it does, but I sorta half-intentionally left that out because I don’t actually think LLS is true as it seems to often be stated.
Since we have the prior knowledge that we are looking at an experiment for which both success and failure are possible, our estimate is as if we had observed one success and one failure for sure before we even started the experiments.
seems both unconvincing as stated and, if assumed to be true, doesn’t depend on that crucial assumption
Idea/suggestion: an “Evergreen” tag, for old (6 months month? 1 year? 3 years?) posts (comments?), to indicate that they’re still worth reading (to me, ideally for their intended value/arguments rather than as instructive historical/cultural artifacts)
yeah some posts are sufficiently 1. good/useful, and 2. generic/not overly invested in one particular author’s voice or particularities that they make more sense as a wiki entry than a “blog”-adjacent post.
Hypothesis: from the perspective of currently living humans and those who will be born in the currrent <4% growth regime only (i.e. pre-AGI takeoff or I guess stagnation) donations currently earmarked for large scale GHW, Givewell-type interventions should be invested (maybe in tech/AI correlated securities) instead with the intent of being deployed for the same general category of beneficiaries in <25 (maybe even <1) years.
The arguments are similar to those for old school “patient philanthropy” except now in particular seems exceptionally uncertain wrt how to help humans because of AI.
For example, it seems plausible that the most important market the global poor don’t have access to is literally the NYSE (ie rather than for malaria nets), because ~any growth associated with (AGI + no ‘doom’) will leave the global poor no better off by default (i.e. absent redistribution or immigration reform)unlike e.g., middle class westerners who might own a bit of the S&P500. A solution could be for e.g. OpenPhil to invest on their behalf.
(More meta: I worry that segmenting off AI as fundamentally longtermist is leaving a lot of good on the table; e.g. insofar as this isn’t currently the case, I think OP’s GHW side should look into what kind of AI-associated projects could do a lot of good for humans and animals in the next few decades.)
I’m skeptical of this take. If you think sufficiently transformative + aligned AI is likely in the next <25 years, then from the perspective of currently living humans and those who will be born in the current <4% growth regime, surviving until transformative AI arrives would be a huge priority. Under that view, you should aim to deploy resources as fast as possible to lifesaving interventions rather than sitting on them.
I tried making a shortform → Twitter bot (ie tweet each new top level ~quick take~) and long story short it stopped working and wasn’t great to begin with.
I feel like this is the kind of thing someone else might be able to do relatively easily. If so, I and I think much of EA Twitter would appreciate it very much! In case it’s helpful for this, a quick takes RSS feed is at https://ea.greaterwrong.com/shortform?format=rss
Note: this sounds like it was written by chatGPT because it basically was (from a recorded ramble)🤷
I believe the Forum could benefit from a Shorterform page, as the current Shortform forum, intended to be a more casual and relaxed alternative to main posts, still seems to maintain high standards. This is likely due to the impressive competence of contributors who often submit detailed and well-thought-out content. While some entries are just a few well-written sentences, others resemble blog posts in length and depth.
As such, I find myself hesitant to adhere to the default filler text in the submission editor when visiting this page. However, if it were more informal and less intimidating in nature, I’d be inclined to post about various topics that might otherwise seem out of place. To clarify, I’m not suggesting we resort to jokes or low-quality “shitposts,” but rather encourage genuine sharing of thoughts without excessive analysis.
Perhaps adopting an amusing name like “EA Shorterform” would help create a more laid-back atmosphere for users seeking lighter discussions. By doing so, we may initiate a preference falsification cascade where everyone feels comfortable enough admitting their desire for occasional brevity within conversations. Who knows? Maybe I’ll start with posting just one sentence soon!
WWOTF: what did the publisher cut? [answer: nothing]
Contextual note: this post is essentially a null result. It seemed inappropriate both as a top-level post and as an abandoned Google Doc, so I’ve decided to put out the key bits (i.e., everything below) as Shortform. Feel free to comment/message me if you think that was the wrong call!
Actual post
On his recent appearance on the 80,000 Hours Podcast, Will MacAskill noted that Doing Good Better was significantly influenced by the book’s publisher:[1]
Rob Wiblin: …But in 2014 you wrote Doing Good Better, and that somewhat soft pedals longtermism when you’re introducing effective altruism. So it seems like it was quite a long time before you got fully bought in.
Will MacAskill: Yeah. I should say for 2014, writing Doing Good Better, in some sense, the most accurate book that was fully representing my and colleagues’ EA thought would’ve been broader than the particular focus. And especially for my first book, there was a lot of equivalent of trade — like agreement with the publishers about what gets included. I also wanted to include a lot more on animal issues, but the publishers really didn’t like that, actually. Their thought was you just don’t want to make it too weird.
Rob Wiblin: I see, OK. They want to sell books and they were like, “Keep it fairly mainstream.”
Will MacAskill: Exactly...
I thought it was important to know whether the same was true with respect to What We Owe the Future, so I reached out to Will’s team and received the following response from one of his colleagues [emphasis mine]:
Hi Aaron, thanks for sending these questions and considering to make this info publicly available.
However, in contrast to what one might perhaps reasonably expect given what Will said about Doing Good Better, I think there is actually very little of interest that can be said on this topic regarding WWOTF. In particular:
I’m not aware of any material that was cut, or any other significant changes to the content of the book that were made significantly because of the publisher’s input. (At least since I joined Forethought in mid-2021; it’s possible there was some of this at earlier stages of the project, though I doubt it.) To be clear: The UK publisher’s editor read multiple drafts of the book and provided helpful comments, but Will generally changed things in response to these comments if and only if he was actually convinced by them.
(There are things other than the book’s content where the publisher exerted more influence – for instance, the publishers asked us for input on the book’s cover but made clear that the cover is ultimately their decision. Similarly, the publisher set the price of the book, and this is not something we were involved in at all.)
As Will talks about in more detail here, the book’s content would have been different in some ways if it had been written for a different audience – e.g., people already engaged in the EA community as opposed to the general public. But this was done by Will’s own choice/design rather than because of publisher intervention. And to be clear, I think this influenced the content in mundane and standard ways that are present in ~all communication efforts – understanding what your audience is, aiming to meet them where they are and delivering your messages in way that is accessible to them (rather than e.g. using overly technical language the audience might not be familiar with).
That^ is too big for Google Sheets, so here’s the same thing just without a breakdown by country that you should be able to open easily if you want to take a look.
Basically the UN data generally used for tracking/analyzing the amount of fish and other marine life captured/farmed and killed only tracks the total weight captured for a given country-year-species (or group of species).
I had chatGPT-4 provide estimated lower and upper bounds on the average weight of individual members of given species/groups, thereby allowing me to guesstimate individual numbers of fish+ from the UN live weight data. I only did this for species/groups of which >5000 metric tons have been harvested since 1961, which covers about 99.9% of the total mass.
Happy to describe anything in more detail if it would actually be of use!
There’s a ton there, but one anecdote from yesterday: referred me to this $5 IOS desktop app which (among other more reasonable uses) made me this full quality, fully intra-linked >3600 page PDF of (almost) every file/site linked to by every file/site linked to from Tomasik’s homepage (works best with old-timey simpler sites like that)
It wasn’t too hard to put together a text doc with (at least some of each of) all 1470ish shortform posts, which you can view or download here.
Pros: (practically) infinite scroll of insight porn
Cons:
longer posts get cut off at about 300 words
Each post is an ugly block of text
No links to the original post [see doc for more]
Various other disclaimers/notes at the top of the document
I was starting to feel like the If You Give a Mouse a Cookie’s eternally-doomed protagonist (it’ll look presentable if I just do this one more thing), so I’m cutting myself off here to see whether it might be worth me (or someone else) making it better.
Newer Thing (?)
I do think this could be an MVP (minimal viable product) for a much nicer-looking and readable document, such as:
“this but without the posts cut off and with spacing figured out”
“nice-looking searchable pdf with original media and formatting”
“WWOTF-level-production book and audiobook”
Any of those ^ three options but only for the top 10/100/n posts
So by all means, copy and paste and turn it into something better!
Oh yeah and, if you haven’t done so already, I highly recommend going through the top Shortform posts for each of the last four years here
Note: inspired by the FTX+Bostrom fiascos and associated discourse. May (hopefully) develop into longform by explicitly connecting this taxonomy to those recent events (but my base rate of completing actual posts cautions humility)
Event as evidence
The default: normal old Bayesian evidence
The realm of “updates,” “priors,” and “credences”
Pseudo-definition: Induces [1] a change to or within a model (of whatever the model’s user is trying to understand)
Corresponds to models that are (as is often assumed):
Well-defined (i.e. specific, complete, and without latent or hidden information)
Stable except in response to ‘surprising’ new information
Event as spotlight
Pseudo-definition: Alters the how a person views, understands, or interacts with a model, just as a spotlight changes how an audience views what’s on stage
In particular, spotlights change the salience of some part of a model
This can take place both/either:
At an individual level (think spotlight before an audience of one); and/or
To a community’s shared model (think spotlight before an audience of many)
They can also which information latent in a model is functionally available to a person or community, just as restricting one’s field of vision increases the resolution of whichever part of the image shines through
Example
You’re hiking a bit of the Appalachian Trail with two friends, going north, using the following of a map (the “external model”)
An hour in, your mental/internal model probably looks like this:
Event: the collapse of a financial institutionyou hear traffic
As evidence, thiscauses you to change where you think you are—namely, a bit south of the first road you were expecting to cross
As spotlight, this causes the three of you to stare at the same map as before model but in such a way that your internal models are all very similar, each looking something like this
Are you factoring in that CEA pays a few hundred bucks per attendee? I’d have a high-ish bar to pay that much for someone to go to a conference myself. Altho I don’t have a good sense of what the marginal attendee/rejectee looks like.
Ok so things that get posted in the Shortform tab also appear in your (my) shortform post , which can be edited to not have the title “___′s shortform” and also has a real post body that is empty by default but you can just put stuff in.
There’s also the usual “frontpage” checkbox, so I assume an individual’s own shortform page can appear alongside normal posts(?).
From Reuters:
I sincerely hope OpenPhil (or Effective Ventures, or both—I don’t know the minutia here) sues over this. Read the reasoning for and details of the $30M grant here.
The case for a legal challenge seems hugely overdetermined to me:
Stop/delay/complicate the restructuring, and otherwise make life appropriately hard for Sam Altman
Settle for a large huge amount of money that can be used to do a huge amount of good
Signal that you can’t just blatantly take advantage of OpenPhil/EV/EA as you please without appropriate challenge
I know OpenPhil has a pretty hands-off ethos and vibe; this shouldn’t stop them from acting with integrity when hands-on legal action is clearly warranted
Good Ventures rather than Effective Ventures, no?
What’s the legal case for a lawsuit?
There’s a broader point here about the takeover of a non-profit organization by financial interests that I’d really like to see fought back against.
You could also add:
”Negotiate safety conditions as part of a settlement”
I understand that OpenAI’s financial situation is not very good [edit: this may not be a high-quality source], and if they aren’t able to convert to a for-profit, things will become even worse:
https://www.wheresyoured.at/oai-business/
It’s conceivable that OpenPhil suing OpenAI could buy us 10+ years of AI timeline, if the following dominoes fall:
OpenPhil sues, and OpenAI fails to convert to a for-profit.
As a result, OpenAI struggles to raise additional capital from investors.
Losing $4-5 billion a year with little additional funding in sight, OpenAI is forced to make some tough financial decisions. They turn off the free version of ChatGPT, stop training new models, and cut salaries for employees. They’re able to eke out some profit, but not much profit, because their product is not highly differentiated from other AI offerings.
Silicon Valley herd mentality kicks in. OpenAI has been the hottest startup in the Valley. If it becomes known as the next WeWork, its fall will be earth-shaking. Game-theoretically, it doesn’t make as much sense to invest in an early AI startup round if there’s no capital willing to invest in subsequent rounds. OpenAI’s collapse could generate the belief that AI startups will struggle to raise capital—and if many investors believe that, it could therefore become true.
The AI bubble deflates and the Valley refocuses on other industries.
It would be extremely ironic if the net effect of all Sam Altman and Mark Zuckerberg’s efforts is to make AI companies uninvestable and buy us a bunch of timeline. Sam by generating a bunch of hype that fails to deliver, and Mark by commoditizing LLMs. (I say “ironic” because EAs are used to thinking of both Sam and Mark as irresponsible actors in the AI space.)
EDIT: there is some criticism of OpenPhil’s approach to its public image here which may be relevant to the decision of whether to sue or not. Also, there’s the obvious point that OpenAI appears to be one of the worst actors in the AI space.
EDIT 2: One also needs to consider how Sam might respond, e.g. by starting a new company and attempting to poach all OpenAI employees.
A couple takes from Twitter on the value of merch and signaling that I think are worth sharing here:
1)
2)
Media is often bought on a CPM basis (cost per thousand views). A display ad on LinkedIn for e.g. might cost $30 CPM. So yeah I think merch is probably underrated.
Interesting that the Animal Welfare Fund gives out so few small grants relative to the Infrastructure and Long Term Future funds (Global Health and Development has only given out 20 grants, all very large, so seems to be a more fundamentally different type of thing(?)). Data here.
A few stats:
The 25th percentile AWF grant was $24,250, compared to $5,802 for Infrastructure and $7,700 for LTFF (and median looks basically the same).
AWF has only made just nine grants of less than $10k, compared to 163 (Infrastructure) and 132 (LTFF).
Proportions under $threshold
Grants under $threshold
Summary stats (rounded)
This is not surprising to me given the different historical funding situations in the relevant cause areas, the sense that animal-welfare and global-health are not talent-constrained as much as funding-constrained, and the clearer presence of strong orgs in those areas with funding gaps.
For instance:
there are 15 references to “upskill” (or variants) in the list of microgrants, and it’s often hard to justify an upskilling grant in animal welfare given the funding gaps in good, shovel-ready animal-welfare projects.
Likewise, 10 references to “study,” 12 to “development,′ 87 to “research” (although this can have many meanings), 17 for variants of “fellow,” etc.
There are 21 references to “part-time,” and relatively small, short blocks of time may align better with community building, small research projects than (e.g.) running a corporate campaign
Seems pretty unsurprising—the animal welfare fund is mostly giving to orgs, while the others give to small groups or individuals for upskilling/outreach frequently.
I think the differences between the LTFF and AWF are largely explained by differences in salary expectations/standards between the cause areas. There are small groups and individuals getting money from the AWF, and they tend to get much less for similar duration projects. Salaries in effective animal advocacy are pretty consistently substantially lower than in AI safety (and software/ML, which AI safety employers and grantmakers might try to compete with somewhat), with some exceptions. This is true even for work in high-income countries like the US and the UK. And, of course, salary expectations are even lower in low- and middle-income countries, which are an area of focus of the AWF (within neglected regions). Plus, many AI safety folks are in the Bay Area specifically, which is pretty expensive (although animal advocates in London also aren’t paid as much).
Yeah but my (implicit, should have made explicit lol) question is “why this is the case?”
Like at a high level it’s not obvious that animal welfare as a cause/field should make less use of smaller projects than the others. I can imagine structural explanations (eg older field → organizations are better developed) but they’d all be post hoc.
I think getting enough people interested in working on animal welfare has not usually been the bottleneck, relative to money to directly deploy on projects, which tend to be larger.
This doesn’t obviously point in the direction of relatively and absolutely fewer small grants, though. Like naively it would shrink and/or shift the distribution to the left—not reshape it.
I don’t understand why you think this is the case. If you think of the “distribution of grants given” as a sum of multiple different distributions (e.g. upskilling, events, and funding programmes) of significantly varying importance across cause areas, then more or less dropping the first two would give your overall distribution a very different shape.
Yeah you’re right, not sure what I missed on the first read
Very interesting, thanks for pulling this data!
There’s a question on the forum user survey:
This is one thing I’ve updated down quite a bit over the last year.
It seems to me that relatively few self-identified EA donors mostly or entirely give to the organization/whatever that they would explicitly endorse as being the single best recipient of a marginal dollar (do others disagree?)
Of course the more important question is whether most EA-inspired dollars are given in such a way (rather than most donors). Unfortunately, I think the answer to this is “no” as well, seeing as OpenPhil continues to donate a majority of dollars to human global health and development[1] (I threw together a Claude artifact that lets you get a decent picture of how OpenPhil has funded cause areas over time and in aggregate)[2]
Edit: to clarify, it could be the case that others have object-level disagreements about what the best use of a marginal dollar is. Clearly this is sometimes the case, but it’s not what I am getting at here. I am trying to get at the phenomenon where people implicitly say/reason “yes, EA principles imply that the best thing to do would be to donate to X, but I am going to donate to Y instead.” I’m guessing this mostly takes the form of people failing to endorse that they’re donations are optimally directed rather than that they do a bunch of ground-up reasoning and then decide to ignore the conclusion it gives, though.
See Open Phil Should Allocate Most Neartermist Funding to Animal Welfare for a sufficient but not necessary case against this.
Data is a few days old and there’s a bit of judgement about how to bin various subcategories of grants, but I doubt the general picture would change much if others redid the analysis/binning
In your original post, you talk about explicit reasoning, in the your later edit, you switch to implicit reasoning. Feels like this criticism can’t be both. I also think the implicit reasoning critique just collapses into object-level disagreements, and the explicit critique just doesn’t have much evidence.
The phenomenon you’re looking at, for instance, is:
And I think this might just be an ~empty set, compared to people having different object-level beliefs about what EA principles are or imply they should do, and also disagree with you on what the best thing to do would be.[1] I really don’t think there’s many people saying “the best thing to do is donate to X, but I will donate to Y”. (References please if so—clarification in footnotes[2]) Even on OpenPhil, I think Dustin just genuinely believes in worldview diversification is the best thing, so there’s no contradiction there where he implies the best thing would be to do X but in practice does do Y.
I think causing this to ‘update downwards’ on your views of the genuine interest of others—as opposed to, say, them being human and fallible despite trying to do the best they can—in the movement feels… well Jason used ‘harsh’, I might use a harsher word to describe this behavior.
For context, I think Aaron thinks that GiveWell deserves ~0 EA funding afaict
I think maybe there might be a difference between the best thing (or best thing using simple calculations) and the right thing. I think people think in terms of the latter and not the former, and unless you buy into strong or even naïve consequentialism we shouldn’t always expect the two to go together
Thanks and I think your second footnote makes an excellent distinction that I failed to get across well in my post.
I do think it’s at least directionally an “EA principle” that “best” and “right” should go together, although of course there’s plenty of room for naive first-order calculation critiques, heuristics/intuitions/norms that might push against some less nuanced understanding of “best”.
I still think there’s a useful conceptual distinction to be made between these terms, but maybe those ancillary (for lack of a better word) considerations relevant to what one thinks is the “best” use of money blur the line enough to make it too difficult to distinguish these in practice.
Re: your last paragraph, I want to emphasize that my dispute is with the terms “using EA principles”. I have no doubt whatsoever about the first part, “genuinely interested in making the world better”
Thanks Aaron, I think you’re responses to me and Jason do clear things up. I still think the framing of it is a bit off though:
I accept that you didn’t intend your framing to be insulting to others, but using “updating down” about the “genuine interest” of others read as hurtful on my first read. As a (relative to EA) high contextualiser it’s the thing that stood out for me, so I’m glad you endorse that the ‘genuine interest’ part isn’t what you’re focusing on, and you could probably reframe your critique without it.
My current understanding of your position is that it is actually: “I’ve come to realise over the last year that many people in EA aren’t directing their marginal dollars/resources to the efforts that I see as most cost-effective, since I also think those are the efforts that EA principles imply are the most effective.”[1] To me, this claim is about the object-level disagreement on what EA principles imply.
However, in your response to Jason you say “it’s possible I’m mistaken over the degree to which direct resources to the place you think needs them most” is a consensus-EA principle which switches back to people not being EA? Or not endorsing this view? But you’ve yet to provide any evidence that people aren’t doing this, as opposed to just disagreeing about what those places are.[2]
Secondary interpretation is: “EA principles imply one should make a quantitative point estimate of the good of all your relevant moral actions, and then act on the leading option in a ‘shut-up-and-calculate’ way. I now believe many fewer actors in the EA space actually do this than I did last year”
For example, in Ariel’s piece, Emily from OpenPhil implies that they have much lower moral weights on animal life than Rethink does, not that they don’t endorse doing ‘the most good’ (I think this is separable from OP’s commitment to worldview diversification).
It seems a bit harsh to treat other user-donors’ disagreement with your views on concentrating funding on their top-choice org (or even cause area) as significant evidence against the proposition that they are “genuinely interested in making the world better using EA principles.”
I think a world in which everyone did this would have some significant drawbacks. While I understand how that approach would make sense through an individual lens, and am open to the idea that people should concentrate their giving more, I’d submit that we are trying to do the most good collectively. For instance: org funding is already too concentrated on a too-small number of donors. If (say) each EA is donating to an average of 5 orgs, then a norm of giving 100% to a single org would decrease the number of donors by 80%. That would impose significant risks on orgs even if their total funding level was not changed.
It’s also plausible that the number of first-place votes an org (or even a cause area) would get isn’t a super-strong reflection of overall community sentiment. If a wide range of people identified Org X as in their top 10%, then that likely points to some collective wisdom about Org X’s cost-effectiveness even if no one has them at number 1. Moreover, spreading the wealth can be seen as deferring to broader community views to some extent—which could be beneficial insofar as one found little reason to believe that wealthier community members are better at deciding where donation dollars should go than the community’s collective wisdom. Thus, there are reasons—other than a lack of genuine interest in EA principles by donors—that donors might reasonably choose to act in accordance with a practice of donation spreading.
Thanks, it’s possible I’m mistaken over the degree to which “direct resources to the place you think needs them most” is a consensus-EA principle.
Also, I recognize that “genuinely interested in making the world better using EA principles” is implicitly value-laden, and to be clear I do wish it was more the case, but I also genuinely intend my claim to be an observation that might have pessimistic implications depending on other beliefs people may have rather than an insult or anything like it, if that makes any sense.
Banger from @Linch
Love this quick take, and would appreciate more similar short fun/funny takes to lift the mood :D.
Boy do I have a website for you (twitter.com)!
(I unironically like twitter for the lower stakes and less insanely high implicit standards)
On mobile now so can’t add image but https://x.com/aaronbergman18/status/1782164275731001368?s=46
I made a custom GPT that is just normal, fully functional ChatGPT-4, but I will donate any revenue this generates[1] to effective charities.
Presenting: Donation Printer
OpenAI is rolling out monetization for custom GPTs:
Re: a recent quick take in which I called on OpenPhil to sue OpenAI: a new document in Musk’s lawsuit mentions this explicitly (page 91)
Interesting lawsuit; thanks for sharing! A few hot (unresearched, and very tentative) takes, mostly on the Musk contract/fraud type claims rather than the unfair-competition type claims related to x.ai:
One of the overarching questions to consider when reading any lawsuit is that of remedy. For instance, the classic remedy for breach of contract is money damages . . . and the potential money damages here don’t look that extensive relative to OpenAI’s money burn.
Broader “equitable” remedies are sometimes available, but they are more discretionary and there may be some significant barriers to them here. Specifically, a court would need to consider the effects of any equitable relief on third parties who haven’t done anything wrongful (like the bulk of OpenAI employees, or investors who weren’t part of an alleged conspiracy, etc.), and consider whether Musk unreasonably delayed bringing this lawsuit (especially in light of those third-party interests). On hot take, I am inclined to think these factors would weigh powerfully against certain types of equitable remedies.
Stated more colloquially, the adverse effects on third parties and the delay (“laches”) would favor a conclusion that Musk will have to be content with money damages, even if they fall short of giving him full relief.
Third-party interests and delay may be less of a barrier to equitable relief against Altman himself.
Musk is an extremely sophisticated party capable of bargaining for what he wanted out of his grants (e.g., a board seat), and he’s unlikely to get the same sort of solicitude on an implied contract theory that an ordinary individual might. For example, I think it was likely foreseeable in 2015 to January 2017 -- when he gave the bulk of the funds in question—that pursuing AGI could be crazy expensive and might require more commercial relationships than your average non-profit would ever consider. So I’d be hesitant to infer much in the way of implied-contractual constraints on OpenAI’s conduct than section 501(c)(3) of the Internal Revenue Code and California non-profit law require.
The fraud theories are tricky because the temporal correspondence between accepting the bulk of the funds and the alleged deceit feels shaky here. By way of rough analogy, running up a bunch of credit card bills you never intended to pay back is fraud. Running up bills and then later deciding that you aren’t going to pay them back is generally only a contractual violation. I’m not deep into OpenAI drama, but a version of the story in which the heel turn happened later in the game than most/all of Musk’s donations and assistance seems plausible to me.
I’m pretty happy with how this “Where should I donate, under my values?” Manifold market has been turning out. Of course all the usual caveats pertaining to basically-fake “prediction” markets apply, but given the selection effects of who spends manna on an esoteric market like this I put a non-trivial weight into the (live) outcomes.
I guess I’d encourage people with a bit more money to donate to do something similar (or I guess defer, if you think I’m right about ethics!), if just as one addition to your portfolio of donation-informing considerations.
This is a really interesting idea! What are your values, so I can make an informed decision?
Thanks! Let me write them as a loss function in python (ha)
For real though:
Some flavor of hedonic utilitarianism
I guess I should say I have moral uncertainty (which I endorse as a thing) but eh I’m pretty convinced
Longtermism as explicitly defined is true
Don’t necessarily endorse the cluster of beliefs that tend to come along for the ride though
“Suffering focused total utilitarian” is the annoying phrase I made up for myself
I think many (most?) self-described total utilitarians give too little consideration/weight to suffering, and I don’t think it really matters (if there’s a fact of the matter) whether this is because of empirical or moral beliefs
Maybe my most substantive deviation from the default TU package is the following (defended here):
“Under a form of utilitarianism that places happiness and suffering on the same moral axis and allows that the former can be traded off against the latter, one might nevertheless conclude that some instantiations of suffering cannot be offset or justified by even an arbitrarily large amount of wellbeing.”
Moral realism for basically all the reasons described by Rawlette on 80k but I don’t think this really matters after conditioning on normative ethical beliefs
Nothing besides valenced qualia/hedonic tone has intrinsic value
I think that might literally be it—everything else is contingent!
I was inspired to create this market! I would appreciate it if you weighed in. :)
Some shrinsight (shrimpsight?) from the comments:
According to Kevin Esvelt on the recent 80,000k podcast (excellent btw, mostly on biosecurity), eliminating the New World New World screwworm could be an important farmed animal welfare (infects livestock), global health (infects humans), development (hurts economies), science/innovation intervention, and most notably quasi-longtermist wild animal suffering intervention.
More, if you think there’s a non-trivial chance of human disempowerment, societal collapse, or human extinction in the next 10 years, this would be important to do ASAP because we may not be able to later.
From the episode:
Dropping a longer quote with more context in this footnote.[1] A quick Google Images search makes this all the more visceral, but be warned that it’s kinda graphic.
I would really love to see someone (maybe me) do a deeper dive into this and write up a proper Forum post.
A few Forum meta things you might find useful or interesting:
Two super basic interactive data viz apps
1) How often (in absolute and relative terms) a given forum topic appears with another given topic
2) Visualizing the popularity of various tags
An updated Forum scrape including the full text and attributes of 10k-ish posts as of Christmas, ’22
See the data without full text in Google Sheets here
Post explaining version 1.0 from a few months back
From the data in no. 2, a few effortposts that never garnered an accordant amount of attention (qualitatively filtered from posts with (1) long read times (2) modest positive karma (3) not a ton of comments.
Columns labels should be (left to right):
Title/link
Author(s)
Date posted
Karma (as of a week ago)
Comments (as of a week ago)
Open Philanthropy: Our Approach to Recruiting a Strong Team
Histories of Value Lock-in and Ideology Critique
Why I think strong general AI is coming soon
Anthropics and the Universal Distribution
Range and Forecasting Accuracy
A Pin and a Balloon: Anthropic Fragility Increases Chances of Runaway Global Warming
Strategic considerations for effective wild animal suffering work
Red teaming a model for estimating the value of longtermist interventions - A critique of Tarsney's "The Epistemic Challenge to Longtermism"
Welfare stories: How history should be written, with an example (early history of Guam)
Summary of Evidence, Decision, and Causality
Some AI research areas and their relevance to existential safety
Maximizing impact during consulting: building career capital, direct work and more.
Independent Office of Animal Protection
Investigating how technology-focused academic fields become self-sustaining
Using artificial intelligence (machine vision) to increase the effectiveness of human-wildlife conflict mitigations could benefit WAW
Crucial questions about optimal timing of work and donations
Will we eventually be able to colonize other stars? Notes from a preliminary review
Philanthropists Probably Shouldn't Mission-Hedge AI Progress
I went ahead and made an “Evergreen” tag as proposed in my quick take from a while back:
cool, but I don’t think a year is right. I would have said 3 years.
I think the proxy question is “after what period of time is it reasonable to assume that any work building or expanding on the post would have been published?” and my intuition here is about 1 year but would be interested in hearing others thoughts
I think that the number of years depends on how fast EA is growing.
Hopefully people will be sparing in applying it to their own recent posts!
Eh I’m not actually sure how bad this would be. Of course it could be overdone, but a post’s author is its obvious best advocate, and a simple “I think this deserves more attention” vote doesn’t seem necessarily illegitimate to me
This post is half object level, half experiment with “semicoherent audio monologue ramble → prose” AI (presumably GPT-3.5/4 based) program audiopen.ai.
In the interest of the latter objective, I’m including 3 mostly-redundant subsections:
A ’final’ mostly-AI written text, edited and slightly expanded just enough so that I endorse it in full (though recognize it’s not amazing or close to optimal)
The raw AI output
The raw transcript
1) Dubious asymmetry argument in WWOTF
In Chapter 9 of his book, What We Are the Future, Will MacAskill argues that the future holds positive moral value under a total utilitarian perspective. He posits that people generally use resources to achieve what they want—either for themselves or for others—and thus good outcomes are easily explained as the natural consequence of agents deploying resources for their goals. Conversely, bad outcomes tend to be side effects of pursuing other goals. While malevolence and sociopathy do exist, they are empirically rare.
MacAskill argues that in a future with continued economic growth and no existential risk, we will likely direct more resources towards doing good things due to self-interest and increased impartial altruism. He contrasts this eutopian scenario with an anti-eutopia: the worst possible world, which he argues (compellingly, I think) less probable because it requires convoluted explanations as opposed to simple desires like enjoying ice cream. He concludes that the probability of achieving a eutopia outweighs the low likelihood but extreme negative consequences of an anti-eutopia.
However, I believe McCaskill’s analysis neglects an important aspect: considering not only these two extremes but also the middle distribution where neither significant amounts of resources nor agentic intervention occur.
When physics operates without agency-driven resource allocation, we have good reason to expect evolution to create conscious beings whose suffering we can attribute to the ease with which animals (or animal-like beings) can lose all future expected genetic reproduction—as MacAskill himself argues elsewhere in the book.
Importantly, though, this non-agentic suffering, seems more likely to complement agentic resource deployment—not substitute for it as one might intuit. That’s because human or post-human expansion necessarily entails the expansion of concentrated physical energy, and seems likely to entail the expansion of other scarce, pro-biotic resources such as DNA, water, and computation.
Although McCaskill does not explicitly claim his binary model comparing eutopia and anti-eutopia is sufficient for understanding this complex problem, it seems to me to be implied;
Only upon attempting to draft a blog post revisiting his work did I noticed the line “We can make some progress by focusing on just two extreme scenarios: the best or worst possible futures, eutopia and anti-eutopia,” acknowledging status of this model as a bit of evidence in the larger question of the future’s value.
In sum, I think MacAskill’s analysis would benefit from addressing the morally relevant middle distribution to provide a more accurate representation of the future under total utilitarianism.
2) Utilitarian Utopia, Anti-Utopia and the Neglected Middle
In Chapter 9 of his book, What We Are the Future, Will McCaskill argues that the future holds positive moral value under a total utilitarian perspective. He posits that people generally use resources to achieve what they want—either for themselves or others—and thus good outcomes are often intentional. Conversely, bad outcomes tend to be side effects of pursuing other goals. While malevolence and sociopathy do exist, they are empirically rare.
McCaskill then extrapolates this argument to suggest that in a future with continued economic growth (assuming no existential risk), we will likely direct more resources towards doing good things due to self-interest and increased impartial altruism. He contrasts this utopian scenario with an anti-utopia: the worst possible world which is less probable because it requires convoluted explanations as opposed to simple desires like enjoying ice cream.
He concludes that the probability of achieving a utopia outweighs the low likelihood but extreme negative consequences of an anti-utopia. However, I believe McCaskill’s analysis neglects an important aspect: considering not only these two extremes but also the middle distribution where neither significant amounts of resources nor agentic intervention occur.
In such cases where physics operates without agency-driven resource allocation, evolution can create conscious beings like plants and animals who experience suffering without any intentionality behind it. This middle distribution may actually skew negatively since evolution favors reproductive fitness at any cost; as a result, sentient beings could suffer more than they experience happiness during their lives.
I argue that wild animal welfare is net negative overall; if given a choice between having them continue living or ceasing their existence altogether, I would choose nonexistence on moral grounds. Although McCaskill does not explicitly claim his heuristic comparison between utopia and anti-utopia is sufficient for understanding this complex problem, he strongly implies it throughout most of his chapter.
Upon revisiting his work while drafting my response blog post, I noticed a single line acknowledging the limitations of his approach. However, this caveat does not adequately emphasize that his argument should be considered only as a first pass and not definitive proof. In my opinion, McCaskill’s analysis would benefit from addressing the morally relevant middle distribution to provide a more accurate representation of the future under total utilitarianism.
-------
3) Original Transcript
Okay, so I’m going to describe where I think I disagree with Will McCaskill in Chapter 9 of his book, What We Are the Future, where he basically makes an argument that the future is positive in expectation, positive moral value under a total utilitarian perspective. And so his argument is basically that people, it’s very easy to see that people deploy the resources in order to get what they want, which is either to help themselves and sometimes to help other people, whether it’s just their family or more impartial altruism. Basically you can always explain why somebody does something good just because it’s good and they want it, which is kind of, I think that’s correct and compelling. Whereas when something bad happens, it’s generally the side effect of something else. At least, yeah. So while there is malevolence and true sociopathy, those things are in fact empirically quite rare, but if you undergo a painful procedure, like a medical procedure, it’s because there’s something affirmative that you want and that’s a necessary side effect. It’s not because you actually sought that out in particular. And all this I find true and correct and compelling. And so then he uses this to basically say that in the future, presumably conditional on continued economic growth, which basically just means no existential risk and humans being around, we’ll be employing a lot of resources in the direction of doing things well or doing good. Largely just because people just want good things for themselves and hopefully to some extent because there will be more impartial altruists willing to both trade and to put their own resources in order to help others. And once again, all true, correct, compelling in my opinion. So on the other side, so basically utopia in this sense, utopia basically meaning employing a lot of, the vast majority of resources in the direction of doing good is very likely and very good. On the other side, it’s how likely and how bad is what he calls anti-utopia, which is basically the worst possible world. And he basically using… I don’t need to get into the particulars, but basically I think he presents a compelling argument that in fact it would be worse than the best world is good, at least to the best of our knowledge right now. But it’s very unlikely because it’s hard to see how that comes about. You actually can invent stories, but they get kind of convoluted. And it’s not nearly as simple as, okay, people like ice cream and so they buy ice cream. It’s like, you have to explain why so many resources are being deployed in the direction of doing good things and you still end up with a terrible world. Then he basically says, okay, all things considered, the probability of good utopia wins out relative to the badness, but very low probability of anti-utopia. Again, a world full of misery. And where I think he goes wrong is that he neglects the middle of the distribution where the distribution is ranging from… I don’t know how to formalize this, but something like percentage or amount of… Yeah, one of those two, percentage or amount of resources being deployed in the direction of on one side of the spectrum causing misery and then the other side of the spectrum causing good things to come about. And so he basically considers the two extreme cases. But I claim that, in fact, the middle of the distribution is super important. And actually when you include that, things look significantly worse because the middle of the distribution is basically like, what does the world look like when you don’t have agents essentially deploying resources in the direction of anything? You just have the universe doing its thing. We can set aside the metaphysics or physics technicalities of where that becomes problematic. Anyway, so basically the middle of the distribution is just universe doing its thing, physics operating. I think there’s the one phenomenon that results from this that we know of to be morally important or we have good reason to believe is morally important is basically evolution creating conscious beings that are not agentic in the sense that I care about now, but basically like plants and animals. And presumably I think you have good reason to believe animals are sentient. And evolution, I claim, creates a lot of suffering. And so you look at the middle of the distribution and it’s not merely asymmetrical, but it’s asymmetrical in the opposite direction. So I claim that if you don’t have anything, if you don’t have lots of resources being deployed in any direction, this is a bad world because you can expect evolution to create a lot of suffering. The reason for that is, as he gets into, something like either suffering is intrinsically more important, which I put some weight on that. It’s not exactly clear how to distinguish that from the empirical case. And the empirical case is basically it’s very easy to lose all your reproductive fitness in the evolutionary world very quickly. It’s relatively hard to massively gain a ton. Reproduction is like, even having sex, for example, only increases your relative reproductive success a little bit, whereas you can be killed in an instant. And so this creates an asymmetry where if you buy a functional view of qualia, then it results in there being an asymmetry where animals are just probably going to experience more pain over their lives, by and large, than happiness. And I think this is definitely true. I think wild animal welfare is just net negative. I wish if I could just… If these are the only two options, have there not be any wild animals or have them continue living as they are, I think it would be overwhelmingly morally important to not have them exist anymore. And so tying things back. Yeah, so McCaskill doesn’t actually… I don’t think he makes a formally incorrect statement. He just strongly implies that this case, that his heuristic of comparing the two tails is a pretty good proxy for the best we can do. And that’s where I disagree. I think there’s actually one line in the chapter where he basically says, we can get a grip on this very hard problem by doing the following. But I only noticed that when I went back to start writing a blog post. And the vast majority of the chapter is basically just the object level argument or evidence presentation. There’s no repetition emphasizing that this is a really, I guess, sketchy, for lack of a better word, dubious case. Or first pass, I guess, is a better way of putting it. This is just a first pass, don’t put too much weight on this. That’s not how it comes across, at least in my opinion, to the typical reader. And yeah, I think that’s everything.
I think there’s a case to be made for exploring the wide range of mediocre outcomes the world could become.
Recent history would indicate that things are getting better faster though. I think MacAskill’s bias towards a range of positive future outcomes is justified, but I think you agree too.
Maybe you could turn this into a call for more research into the causes of mediocre value lock-in. Like why have we had periods of growth and collapse, why do some regions regress, what tools can society use to protect against sinusoidal growth rates.
So the EA Forum has, like, an ancestor? Is this common knowledge? Lol
Felicifia: not functional anymore but still available to view. Learned about thanks to a Tweet from Jacy
From Felicifia Is No Longer Accepting New Users:
Update: threw together
some data with authors, post title names, date, and number of replies (and messed one section up so some rows are missing links)
A rather long PDF with the posts and replies together (for quick keyword searching), with decent but not great formatting
Wow, blast from the past!
A little while ago I posted this quick take:
I didn’t have a good response to @DanielFilan, and I’m pretty inclined to defer to orgs like CEA to make decisions about how to use their own scarce resources.
At least for EA Global Boston 2024 (which ended yesterday), there was the option to pay a “cost covering” ticket fee (of what I’m told is $1000).[1]
All this is to say that I am now more confident (although still <80%) that marginal rejected applicants who are willing to pay their cost-covering fee would be good to admit.[2]
In part this stems from an only semi-legible background stance that, on the whole, less impressive-seeming people have more ~potential~ and more to offer than I think “elite EA” (which would those running EAG admissions) tend to think. And this, in turn, has a lot to do with the endogeneity/path dependence of I’d hastily summarize as “EA involvement.”
That is, many (most?) people need a break-in point to move from something like “basically convinced that EA is good, interested in the ideas and consuming content, maybe donating 10%” to anything more ambitious.
For some, that comes in the form of going to an elite college with a vibrant EA group/community. Attending EAG is another—or at least could be. But if admission is dependent on doing the kind of things and/or having the kinds of connections that a person might only pursue after getting on such an on-ramp, you have a vicious cycle of endogenous rejection.
The impetus for writing this is seeing a person who was rejected with some characteristics that seem plausibly pretty representative of a typical marginal EAG rejectee:
College educated but not via an elite university
Donates 10%, mostly to global health
Normal-looking middle or upper-middle class career
Interested in EA ideas but not a huge amount to show for it
Never attended EAG
Of course n=1, this isn’t a tremendous amount of evidence, I don’t have strictly more information than the admissions folks, the optimal number of false-negatives is not zero, etc., etc. But if a person with those above characteristics who is willing to write a reasonably thoughtful application and spend their personal time and money traveling to and taking part in EA Global (and, again, covering their cost)[3] is indeed likely to get rejected, I just straightforwardly think that admission has too high a bar; does CEA really think such a person is actively harmful to the event on net?
I don’t want to say that there is literally zero potential downside from admitting more people and “diluting the attendee pool” for lack of a more thoughtful term, but it’s not immediately obvious to me what that downside would be especially at the current margin (say for example a 25% increase in the number of attendees, not a 2000% increase). And, needless to say, there is a lot of potential upside via both the impact of this marginal attendee themselves and via the information/experience/etc. that they bring to the whole group.
I’m not sure if this was an option when I wrote my previous take, and if so whether I just didn’t know about it or what.
To be clear none of this is because I’ve ever been rejected (I haven’t). Kinda cringe to say but worth the clarification I think.
If this were an econ paper, I’d probably want to discuss the fiscal relevance of marginal vs average cost-covering tickets. I suspect that the “cost covering ticket” actually advertised is based on the average cost, but I’m not sure.
If this is true, and marginal cost < average cost as seems intuitive, then admitting a marginal attendee who then pays the average cost would be financially net-positive for CEA.
I am under the impression that EAGx can be such a break-in point, and has lower admission standards than EAG. In particular, there is EAGxVirtual (Applications are open!).
Has the rejected person you are thinking of applied to any EAGx conference?
I agree. One minor issue with your “low bar” is the giving 10 percent. Giving this much is extremely uncommon to any cause, so for me might be more of a “medium bar” ;)
Would this (generally) be a one-time deal? The idea that some people would benefit from a bolus of EA as a “break-in point” or “on-ramp” seems plausible, and willingness to pay a hefty admission fee / other expenses would certainly have a signaling value.[1] However, the argument probably gets weaker after the first marginal admission (unless the marginal applicant is a lot closer to the line on the second time around).
Maybe allowing only one marginal admission per person absent special circumstances would mitigate concerns about “diluting” the event.
I recognize the downsides of a pay-extra-to-attend approach as far as perceived fairness, equity, accessibility to people from diverse backgrounds, and so on. That would be a tradeoff to consider.
Random sorta gimmicky AI safety community building idea: tabling at universities but with a couple laptop signed into Claude Pro with different accounts. Encourage students (and profs) to try giving it some hard question from eg a problem set and see how it performs. Ideally have a big monitor for onlookers to easily see.
Most college students are probably still using ChatGPT-3.5, if they use LLMs at all. There’s a big delta now between that and the frontier.
I have a vague fear that this doesn’t do well on the ‘try not to have the main net effect be AI hypebuilding’ heuristic.
A (potential) issue with MacAskill’s presentation of moral uncertainty
Not able to write a real post about this atm, though I think it deserves one.
MacAskill makes a similar point in WWOTF, but IMO the best and most decision-relevant quote comes from his second appearance on the 80k podcast:
I don’t think the second bolded sentence follows in any objective or natural manner from the first. Rather, this reasoning takes a distinctly total utilitarian meta-level perspective, summing the various signs of utility and then implicitly considering them under total utilitarianism.
Even granting that the mora arithmetic is appropriate and correct, it’s not at all clear what to do once the 2:1 accounting is complete. MacAskill’s suffering-focused twin might have reasoned instead that
A better proxy for genuine neutrality (and the best one I can think of) might be to simulate bargaining over real-world outcomes from each perspective, which would probably result in at least some proportion of one’s resources being deployed as though negative utilitarianism were true (perhaps exactly 50%, though I haven’t given this enough thought to make the claim outright).
Made a podcast feed with EAG talks. Now has both the recent Bay Area and London ones:
Spotify
RSS (can plug into any other platform)
Full vids on the CEA Youtube page
The recent 80k podcast on the contingency of abolition got me wondering what, if anything, the fact of slavery’s abolition says about the ex ante probability of abolition—or more generally, what one observation of a binary random variable X says about p as in
Turns out there is an answer (!), and it’s found starting in paragraph 3 of subsection 1 of section 3 of the Binomial distribution Wikipedia page:
Don’t worry, I had no idea what Beta(α,β) was until 20 minutes ago. In the Shortform spirit, I’m gonna skip any actual explanation and just link Wikipedia and paste this image (I added the uniform distribution dotted line because why would they leave that out?)
So...
Cool, so for the n=1 case, we get that if you have a prior over the ex ante probability space[0,1] described by one of those curves in the image, you...
0) Start from ‘zero empirical information guesstimate’ E[Beta(α,β)]=αα+β
1a) observe that the thing happens (x=1), moving you, Ideal Bayesian Agent, to updated probability ^pb=1+α1+α+β>αα+β OR
1b) observe that the thing doesn’t happen (x=0), moving you to updated probability ^pb=α1+α+β<αα+β
In the uniform case (which actually seems kind of reasonable for abolition), you...
0) Start from prior E[p]=1/2
1a) observe that the thing happens, moving you to updated probability ^pb=2/3
1a) observe that the thing doesn’t happen, moving you to updated probability ^pb=13
The uniform prior case just generalizes to Laplace’s Law of Succession, right?
In terms of result,yeah it does, but I sorta half-intentionally left that out because I don’t actually think LLS is true as it seems to often be stated.Why the strikethrough: after writing the shortform, I get that e.g., “if we know nothing more about them” and “in the absence of additional information” mean “conditional on a uniform prior,” but I didn’t get that before. And Wikipedia’s explanation of the rule,
seems both unconvincing as stated and, if assumed to be true, doesn’t depend on that crucial assumption
The last line contains a typo, right?
Fixed, thanks!
Idea/suggestion: an “Evergreen” tag, for old (6 months month? 1 year? 3 years?) posts (comments?), to indicate that they’re still worth reading (to me, ideally for their intended value/arguments rather than as instructive historical/cultural artifacts)
As an example, I’d highlight Log Scales of Pleasure and Pain, which is just about 4 years old now.
I know I could just create a tag, and maybe I will, but want to hear reactions and maybe generate common knowledge.
I think we want someone to push them back into the discussion.
Or you know, have editable wiki versions of them.
yeah some posts are sufficiently 1. good/useful, and 2. generic/not overly invested in one particular author’s voice or particularities that they make more sense as a wiki entry than a “blog”-adjacent post.
Hypothesis: from the perspective of currently living humans and those who will be born in the currrent <4% growth regime only (i.e. pre-AGI takeoff or I guess stagnation) donations currently earmarked for large scale GHW, Givewell-type interventions should be invested (maybe in tech/AI correlated securities) instead with the intent of being deployed for the same general category of beneficiaries in <25 (maybe even <1) years.
The arguments are similar to those for old school “patient philanthropy” except now in particular seems exceptionally uncertain wrt how to help humans because of AI.
For example, it seems plausible that the most important market the global poor don’t have access to is literally the NYSE (ie rather than for malaria nets), because ~any growth associated with (AGI + no ‘doom’) will leave the global poor no better off by default (i.e. absent redistribution or immigration reform) unlike e.g., middle class westerners who might own a bit of the S&P500. A solution could be for e.g. OpenPhil to invest on their behalf.
(More meta: I worry that segmenting off AI as fundamentally longtermist is leaving a lot of good on the table; e.g. insofar as this isn’t currently the case, I think OP’s GHW side should look into what kind of AI-associated projects could do a lot of good for humans and animals in the next few decades.)
I’m skeptical of this take. If you think sufficiently transformative + aligned AI is likely in the next <25 years, then from the perspective of currently living humans and those who will be born in the current <4% growth regime, surviving until transformative AI arrives would be a huge priority. Under that view, you should aim to deploy resources as fast as possible to lifesaving interventions rather than sitting on them.
LessWrong has a new feature/type of post called “Dialogues”. I’m pretty excited to use it, and hope that if it seems usable, reader friendly, and generally good the EA Forum will eventually adopt it as well.
I tried making a shortform → Twitter bot (ie tweet each new top level ~quick take~) and long story short it stopped working and wasn’t great to begin with.
I feel like this is the kind of thing someone else might be able to do relatively easily. If so, I and I think much of EA Twitter would appreciate it very much! In case it’s helpful for this, a quick takes RSS feed is at https://ea.greaterwrong.com/shortform?format=rss
I would be interested in following this bot if it were made. Thanks for trying!
Note: this sounds like it was written by chatGPT because it basically was (from a recorded ramble)🤷
I believe the Forum could benefit from a Shorterform page, as the current Shortform forum, intended to be a more casual and relaxed alternative to main posts, still seems to maintain high standards. This is likely due to the impressive competence of contributors who often submit detailed and well-thought-out content. While some entries are just a few well-written sentences, others resemble blog posts in length and depth.
As such, I find myself hesitant to adhere to the default filler text in the submission editor when visiting this page. However, if it were more informal and less intimidating in nature, I’d be inclined to post about various topics that might otherwise seem out of place. To clarify, I’m not suggesting we resort to jokes or low-quality “shitposts,” but rather encourage genuine sharing of thoughts without excessive analysis.
Perhaps adopting an amusing name like “EA Shorterform” would help create a more laid-back atmosphere for users seeking lighter discussions. By doing so, we may initiate a preference falsification cascade where everyone feels comfortable enough admitting their desire for occasional brevity within conversations. Who knows? Maybe I’ll start with posting just one sentence soon!
WWOTF: what did the publisher cut? [answer: nothing]
Contextual note: this post is essentially a null result. It seemed inappropriate both as a top-level post and as an abandoned Google Doc, so I’ve decided to put out the key bits (i.e., everything below) as Shortform. Feel free to comment/message me if you think that was the wrong call!
Actual post
On his recent appearance on the 80,000 Hours Podcast, Will MacAskill noted that Doing Good Better was significantly influenced by the book’s publisher:[1]
I thought it was important to know whether the same was true with respect to What We Owe the Future, so I reached out to Will’s team and received the following response from one of his colleagues [emphasis mine]:
Quote starts at 39:47
New fish data with estimated individuals killed per country/year/species (super unreliable, read below if you’re gonna use!)
That^ is too big for Google Sheets, so here’s the same thing just without a breakdown by country that you should be able to open easily if you want to take a look.
Basically the UN data generally used for tracking/analyzing the amount of fish and other marine life captured/farmed and killed only tracks the total weight captured for a given country-year-species (or group of species).
I had chatGPT-4 provide estimated lower and upper bounds on the average weight of individual members of given species/groups, thereby allowing me to guesstimate individual numbers of fish+ from the UN live weight data. I only did this for species/groups of which >5000 metric tons have been harvested since 1961, which covers about 99.9% of the total mass.
Happy to describe anything in more detail if it would actually be of use!
A resource that might be useful: https://tinyapps.org/
There’s a ton there, but one anecdote from yesterday: referred me to this $5 IOS desktop app which (among other more reasonable uses) made me this full quality, fully intra-linked >3600 page PDF of (almost) every file/site linked to by every file/site linked to from Tomasik’s homepage (works best with old-timey simpler sites like that)
New Thing
Last week I complained about not being able to see all the top shortform posts in one list. Thanks to Lorenzo for pointing me to the next best option:
It wasn’t too hard to put together a text doc with (at least some of each of) all 1470ish shortform posts, which you can view or download here.
Pros: (practically) infinite scroll of insight porn
Cons:
longer posts get cut off at about 300 words
Each post is an ugly block of text
No links to the original post [see doc for more]
Various other disclaimers/notes at the top of the document
I was starting to feel like the If You Give a Mouse a Cookie’s eternally-doomed protagonist (it’ll look presentable if I just do this one more thing), so I’m cutting myself off here to see whether it might be worth me (or someone else) making it better.
Newer Thing (?)
I do think this could be an MVP (minimal viable product) for a much nicer-looking and readable document, such as:
“this but without the posts cut off and with spacing figured out”
“nice-looking searchable pdf with original media and formatting”
“WWOTF-level-production book and audiobook”
Any of those ^ three options but only for the top 10/100/n posts
So by all means, copy and paste and turn it into something better!
Oh yeah and, if you haven’t done so already, I highly recommend going through the top Shortform posts for each of the last four years here
Infinitely easier said than done, of course, but some Shortform feedback/requests
The link to get here from the main page is awfully small and inconspicuous (1 of 145 individual links on the page according to a Chrome extension)
I can imagine it being near/stylistically like:
“All Posts” (top of sidebar)
“Recommendations” in the center
“Frontpage Posts”, but to the main section’s side or maybe as a replacement for it you can easily toggle back and forth from
Would be cool to be able to sort and aggregate like with the main posts (nothing to filter by afaik)
I’d really appreciate being able to see the highest-scoring Shortform posts ever, but afaik can’t easily do that atm.
For 2.a the closest I found is https://forum.effectivealtruism.org/allPosts?sortedBy=topAdjusted&timeframe=yearly&filter=all, you can see the inflation-adjusted top posts and shortforms by year.
For 1 it’s probably best to post in the EA Forum feature suggestion thread
Late but thanks on both, and commented there!
Events as evidence vs. spotlights
Note: inspired by the FTX+Bostrom fiascos and associated discourse. May (hopefully) develop into longform by explicitly connecting this taxonomy to those recent events (but my base rate of completing actual posts cautions humility)
Event as evidence
The default: normal old Bayesian evidence
The realm of “updates,” “priors,” and “credences”
Pseudo-definition: Induces [1] a change to or within a model (of whatever the model’s user is trying to understand)
Corresponds to models that are (as is often assumed):
Well-defined (i.e. specific, complete, and without latent or hidden information)
Stable except in response to ‘surprising’ new information
Event as spotlight
Pseudo-definition: Alters the how a person views, understands, or interacts with a model, just as a spotlight changes how an audience views what’s on stage
In particular, spotlights change the salience of some part of a model
This can take place both/either:
At an individual level (think spotlight before an audience of one); and/or
To a community’s shared model (think spotlight before an audience of many)
They can also which information latent in a model is functionally available to a person or community, just as restricting one’s field of vision increases the resolution of whichever part of the image shines through
Example
You’re hiking a bit of the Appalachian Trail with two friends, going north, using the following of a map (the “external model”)
An hour in, your mental/internal model probably looks like this:
Event:
the collapse of a financial institutionyou hear trafficAs evidence, this causes you to change where you think you are—namely, a bit south of the first road you were expecting to cross
As spotlight, this causes the three of you to stare at the same map as before model but in such a way that your internal models are all very similar, each looking something like this
Or fails to induce
EAG(x)s should have a lower acceptance bar. I find it very hard to believe that accepting the marginal rejectee would be bad on net.
Are you factoring in that CEA pays a few hundred bucks per attendee? I’d have a high-ish bar to pay that much for someone to go to a conference myself. Altho I don’t have a good sense of what the marginal attendee/rejectee looks like.
What is the acceptance bar?
Ok so things that get posted in the Shortform tab also appear in your (my) shortform post , which can be edited to not have the title “___′s shortform” and also has a real post body that is empty by default but you can just put stuff in.
There’s also the usual “frontpage” checkbox, so I assume an individual’s own shortform page can appear alongside normal posts(?).
The link is: [Draft] Used to be called “Aaron Bergman’s shortform” (or smth)
I assume only I can see this but gonna log out and check
Effective Altruism Georgetown will be interviewing Rob Wiblin for our inaugural podcast episode this Friday! What should we ask him?