Are you an earn-to-giver or (aspiring) philanthropist who has short AGI timelines and/or high p(doom|AGI)? Do you want to discuss donation opportunities with others who share your goal of slowing down / pausing / stopping AI development[1]? If so, I want to hear from you!
For some context, I’ve been extremely concerned about short-term AI x-risk since March 2023 (post-GPT-4), and have, since then, thought that more AI Safety research will not be enough to save us (or AI Governance that isn’t focused[2] on slowing down AI or a global moratorium on further capabilities advances). Thus I think that on the margin far more resources need to be going into slowing down AI (there are already many dedicated funds for the wider space of AI Safety).
I posted this to an EA investing group in late April:
And this—AGI rising: why we are in a new era of acute risk and increasing public awareness, and what to do now—to the EA Forum in early May. My p(doom|AGI) is ~90% as things stand (Doom is default outcome of AGI). But my p(doom) overall is ~50% by 2030, because I think there’s a decent chance we can actually get a Stop[3]. My timelines are ~0-5 years:I have donated >$150k[4] to people and projects focused on slowing down AI since (mostly as kind of seed funding—to individuals, and projects so new they don’t have official orgs yet[5]), but I want to do a lot more. Having people with me would be great for multiplying impact and also for my motivation!
I’m thinking 4-6 people, each committing ~$100k(+) over 2024, would be good. The idea would be to discuss donation opportunities in the “slowing down AI” space during a monthly call (e.g. Google Meet), and have an informal text chat for the group (e.g. Whatsapp or Messenger). Fostering a sense of unity of purpose[6], but nothing too demanding or official. Active, but low friction and low total time commitment. Donations would be made independently rather than from a pooled fund, but we can have some coordination to get “win-wins” based on any shared preferences of what to fund. Meta-charity Funders is a useful model.
We could maybe do something like an S-process for coordination, like what Jaan Tallinn’s Survival and Flourishing Fund does[7]; it helps avoid “donor chicken” situations. Or we could do something simpler like rank the value of donating successive marginal $10k amounts to each project. Or just stick to more qualitative discussion. This is all still to be determined by the group.
Please join me if you can[8], or share with others you think may be interested. Feel free to DM me here or on X, book a call with me, or fill in this form.
If you oppose AI for other reasons (e.g. ethics, job loss, copyright), as long as you are looking to fund strategies that aim to show results in the short term (say within a year), then I’d be interested in you joining the circle.
It’s a little complicated, but here’s a short description: “Everyone individually decides how much value each project creates at various funding levels. We find an allocation of funds that’s fair and maximises the funders’ expressed preferences (using a number of somewhat dubious but probably not too terrible assumptions). Funders can adjust how much money they want to distribute after seeing everyone’s evaluations, including fully pulling out.” (paraphrased from Lightspeed Grants [funders and evaluators would probably be the same people in our case, but we could outsource evaluation]). See link (the footnote) for more detail.
Funding circle aimed at slowing down AI—looking for participants
Are you an earn-to-giver or (aspiring) philanthropist who has short AGI timelines and/or high p(doom|AGI)? Do you want to discuss donation opportunities with others who share your goal of slowing down / pausing / stopping AI development[1]? If so, I want to hear from you!
For some context, I’ve been extremely concerned about short-term AI x-risk since March 2023 (post-GPT-4), and have, since then, thought that more AI Safety research will not be enough to save us (or AI Governance that isn’t focused[2] on slowing down AI or a global moratorium on further capabilities advances). Thus I think that on the margin far more resources need to be going into slowing down AI (there are already many dedicated funds for the wider space of AI Safety).
I posted this to an EA investing group in late April:
And this—AGI rising: why we are in a new era of acute risk and increasing public awareness, and what to do now—to the EA Forum in early May. My p(doom|AGI) is ~90% as things stand (Doom is default outcome of AGI). But my p(doom) overall is ~50% by 2030, because I think there’s a decent chance we can actually get a Stop[3]. My timelines are ~0-5 years:I have donated >$150k[4] to people and projects focused on slowing down AI since (mostly as kind of seed funding—to individuals, and projects so new they don’t have official orgs yet[5]), but I want to do a lot more. Having people with me would be great for multiplying impact and also for my motivation!I’m thinking 4-6 people, each committing ~$100k(+) over 2024, would be good. The idea would be to discuss donation opportunities in the “slowing down AI” space during a monthly call (e.g. Google Meet), and have an informal text chat for the group (e.g. Whatsapp or Messenger). Fostering a sense of unity of purpose[6], but nothing too demanding or official. Active, but low friction and low total time commitment. Donations would be made independently rather than from a pooled fund, but we can have some coordination to get “win-wins” based on any shared preferences of what to fund. Meta-charity Funders is a useful model.
We could maybe do something like an S-process for coordination, like what Jaan Tallinn’s Survival and Flourishing Fund does[7]; it helps avoid “donor chicken” situations. Or we could do something simpler like rank the value of donating successive marginal $10k amounts to each project. Or just stick to more qualitative discussion. This is all still to be determined by the group.
Please join me if you can[8], or share with others you think may be interested. Feel free to DM me here or on X, book a call with me, or fill in this form.
If you oppose AI for other reasons (e.g. ethics, job loss, copyright), as long as you are looking to fund strategies that aim to show results in the short term (say within a year), then I’d be interested in you joining the circle.
I think Jaan Tallinn’s new top priorities are great!
After 2030, if we have a Stop and are still here, we can keep kicking the can down the road..
I’ve made a few more donations since that tweet.
Public examples include Holly Elmore, giving away copies of Uncontrollable, and AI-Plans.com.
Right now I feel quite isolated making donations in this space.
It’s a little complicated, but here’s a short description: “Everyone individually decides how much value each project creates at various funding levels. We find an allocation of funds that’s fair and maximises the funders’ expressed preferences (using a number of somewhat dubious but probably not too terrible assumptions). Funders can adjust how much money they want to distribute after seeing everyone’s evaluations, including fully pulling out.” (paraphrased from Lightspeed Grants [funders and evaluators would probably be the same people in our case, but we could outsource evaluation]). See link (the footnote) for more detail.
If there is enough interest from people who want to contribute but can’t stretch to $100k, we can look at other potential structures.