Thanks, Martijn. Yes, that’s another point of confusion; someone would need to pick up the tab for the added costs associated with better welfare.
SiobhanBall
Thanks for engaging, Constance, this does answer some of my questions.
I take your point that the window for the EU AI Act was short and that missing it would have meant missing it entirely. Are there any other win-win situations have been found and packaged so far, beyond the EU AI Act?
The BOTEC is a careful piece of reasoning, but it’s a model that compares two categories of work without specifying what either involves in practice. It doesn’t answer my questions about what concretely gets done, and how animals stand to benefit.
Your homepage describes Sentient Futures as existing to ‘identify the leverage points for weaving welfare into the core of future systems and cultivate the foundational community needed to activate them.’ A casual reader can’t translate that into what work is happening in real terms.
You acknowledged the communication gap yourself; I wonder if that’s where some effort could go, given you’re trying to build a field that hinges on persuading people of the gravity and impact of changes that need to happen now, with time-sensitive urgency.
AI and animal welfare: am I missing something?
Oh, no. I had no idea they could be this bad. And I’m speaking from experience… is anybody working on this?
‘When EA becomes cool, people who don’t care about EA show up to secure money and status. When EA becomes cool, that reputation needs defending, which is often corrosive to truth.’
I think that’s very well said.
P.S I’m also unsure of how hiring from near-misses = fewer charities. There’s probably a large stockpile of finalists, just from the final 1% of previous rounds.
Hi Joey, thanks for replying. Zooming in on this point: ‘If we did closed rounds including people from prior cohorts, we would likely lose ~50% of our talent pool but save ~90% of team time on vetting and comms.’
What’s a ballpark figure for the amount of money that 90% of team time represents? This is the amount that could potentially be saved per round.
Having that figure is important; it would help one to assess whether it’s a potentially good trade-off. If the amount would be substantial, then the savings could fund the fewer charities that are incubated, for example.
They may be, although absolute candidate quality doesn’t seem to be the only consideration; for context, the feedback I received as a finalist suggested a different mechanism. I was told that a main consideration was that I was relatively locked into a single idea, and that this idea was among the most popular in the cohort. ‘This does mean we have to make the difficult decision of turning down talented potential founders like yourself who are better suited to some popular ideas.’
The implication was that they expected to be able to find founders for that idea regardless, so they prioritised more flexible candidates or those interested in less popular ideas in order to maximise the total number of charities launched.So I don’t think there’s a pure ‘fixed-bar’. Some candidates who clear the bar might still be turned down due to idea-level constraints.
I’m not sure to what extent both of these are operating simultaneously (e.g. a minimum bar + then matching), but if the latter is a meaningful factor, it seems to strengthen the case for tracking near-miss candidates across rounds rather than resetting the pool each time.
P.S I originally posted this as a quick take and have been thinking about this question in more detail since.
Why doesn’t AIM maintain a ‘near-miss’ founder pool between CEIP rounds?
Lol you had me for a moment there!
This looks great! When will the date be determined?
This is great! I’m excited to see some potentially high-impact, real-world interventions. It also helps to make the area of WAW more tangible for readers.
Since you explicitly asked for criticism:
- Pie chart labels are too small.
- I don’t like the thick partial border around the work segment—makes it look like a clock, which is perhaps deliberate as it’s gesturing at time—but I look at it and think ‘this is a book about 5 o’ clock’.
- I don’t like The Cornerstone Press label going vertically, and I’m unsure of how it relates to the Penguin logo.
- I think ‘that does good’ is a bit of a flat subtitle. Too vague. I much prefer the wording on the site: ridiculously in-depth is ballsy, and I like it! That should be on the cover.
- I would’ve tried for an endorsement from a more mainstream figure in society, like a celebrity or something… I don’t know, Stephen Fry? Lex Friedman? Rutger Bregman is very known in EA, I’m not sure how known outside of EA, and since he’s written a book on a similar topic, one might see that and think ‘why should I read both?’
Also the link to the alternative design doesn’t work.
Learning of this book launch also has me wondering: how is this book different to Moral Ambition or the HIP playbook?
Anyway, there’s my tuppence worth. Are you asking because the cover is still editable? Congratulations on the launch!
Question for AIM folks: what’s the thinking behind running a very involved process twice per year, as opposed to recruiting from near-misses from previous rounds?
Are there savings to be made here? Asking as someone deeply concerned with cost effectiveness as a vital principle of EA… and a former finalist!
That seems like a very high rate of remission to me, congratulations! And on the EAIF funding as well.
It seems that this problem is best solved by regulation + import controls, with industry-wide agreements being an ‘in the meantime’ solution. Doesn’t this make a case for redirecting more resources toward legislative strategy, and less toward corporate campaigns?
Hahaha, alright. I appreciate the self-awareness. :) I, too, have words to live by, even Spanish ones.
And they call us a cult. A 𝑐𝑢𝑙𝑡!
I skimmed your examples and think neither should meet the bar for a ‘little telling off.’
In light of the Netflix doccos coming out later this year, and whatever other scandal might be waiting around the corner, I think we ought to ease up on this level of public language policing. The critics are coming; let’s not make their job easy for them (nor convert from within our own community).
Ok, my updated understanding is that Sentient Futures is primarily focused on field-building, with a view to supporting interventions as they emerge over time.
One thing I’m still trying to get a better grip on is how this translates into impact on animals, and ideally, on what timescale. I’ve had similar questions when thinking about wild animal welfare more broadly: when does investment in building a field start to produce concrete outcomes that benefit animals?
In the AI x animals case, it seems slightly more pressing because of the time-sensitivity point. I’m trying to reconcile the idea that ‘this is urgent’ with an approach that is upstream and preparatory.
I’m also conscious that most of what I’m seeing is the public-facing layer, and you mentioned that a lot of the communication is happening in more private or high-context settings; so it may be that the picture looks more abstract from the outside than it does from within.
Thanks for the invite. I’ll join the showcase. Looking forward to seeing what interventions are being worked on.