Previously I’ve been a Research Fellow at the Forethought Foundation, where I worked on What We Owe The Future with Will MacAskill; an Applied Researcher at Founders Pledge; and a Program Analyst for UNDP.
Stephen Clare
Stable totalitarianism: an overview
I think focusing on AI explosive growth has grown in status over the last two years. I don’t think many people were focusing on it two years ago except Tom Davidson. Since then, Utility Bill has decided to focus on it full-time, Vox has written about it, it’s a core part of the Situational Awareness model, and Carl Shulman talked about it for hours in influential episodes on the 80K and Dwarkesh podcasts.
Tweet away! 🫡
Thanks for this, really helpful! For what it’s worth, I also think Leopold is far too dismissive of international cooperation.
You’ve written there that “my argument was that Aschenbrenner was ‘dangerous’”. I definitely agree that securitisation (and technology competition) often raises risks.[1] I think we have to argue further, though, that securitisation is more dangerous on net than the alternative: a pursuit of international cooperation that may, or may not, be unstable. That, too, may raise some risks, e.g. proliferation and stable authoritarianism.
One of the weaker parts of the Situational Awareness essay is Leopold’s discussion of international AI governance.
He argues the notion of an international treaty on AI “fanciful”, claiming that:
It would be easy to “break out” of treaty restrictions
There would be strong incentives to do so
So the equilibrium is unstable
That’s basically it—international cooperation gets about 140 words of analysis in the 160 page document.
I think this is seriously underargued. Right now it seems harmful to propagate a meme like “International AI cooperation is fanciful”.
This is just a quick take, but I think it’s the case that:
It might not be easy to break out of treaty restrictions. Of course it will be hard to monitor and enforce a treaty. But there’s potential to make it possible through hardware mechanisms, cloud governance, inspections, and other mechanisms that we haven’t even thought of yet. Lots of people are paying attention to this challenge and working on it.
There might not be strong incentives to do so. Decisionmakers may take the risks seriously and calculate the downsides of an all-out race exceed the potential benefits of winning. Credible benefit-sharing and shared decision-making institutions may convince states they’re better off cooperating than trying to win a race.
International cooperation might not be all-or-nothing. Even if we can’t (or shouldn’t!) institute something like a global pause, cooperation on more narrow issues to mitigate threats from AI misuse and loss of control could be possible. Even in the midst of the Cold War, the US and USSR managed to agree on issues like arms control, non-proliferation, and technologies like anti-ballistic missile tech.
(I critiqued a critique of Aschenbrenner’s take on international AI governance here, so I wanted to clarify that I actually do think his model is probably wrong here.)
- 30 Jul 2024 16:21 UTC; 8 points) 's comment on Against Aschenbrenner: How ‘Situational Awareness’ constructs a narrative that undermines safety and threatens humanity by (
Vasco, how do your estimates account for model uncertainty? I don’t understand how you can put some probability on something being possible (i.e. p(extinction|nuclear war) > 0), but end up with a number like 5.93e-12 (i.e. 1 in ~160 billion). That implies an extremely, extremely high level of confidence. Putting ~any weight on models that give higher probabilities would lead to much higher estimates.
Thanks for writing this, it’s clearly valuable to advance a dialogue on these incredibly important issues.
I feel an important shortcoming of this critique is that it frames the choice between national securitization vs. macrosecuritization in terms of a choice between narratives, without considering incentives. I think Leopold gives more consideration to alternatives than you give him credit for, but argues that macrosecuritization is too unstable of an equilibrium:
Some hope for some sort of international treaty on safety. This seems fanciful to me. The world where both the CCP and USG are AGI-pilled enough to take safety risk seriously is also the world in which both realize that international economic and military predominance is at stake, that being months behind on AGI could mean being permanently left behind. If the race is tight, any arms control equilibrium, at least in the early phase around superintelligence, seems extremely unstable. In short, ”breakout” is too easy: the incentive (and the fear that others will act on this incentive) to race ahead with an intelligence explosion, to reach superintelligence and the decisive advantage, too great.
I also think you underplay the extent to which Leopold’s focus on national security is instrumental to his goal of safeguarding humanity’s future. You write: “It is true that Aschenbrenner doesn’t always see himself as purely protecting America, but the free world as a whole, and probably by his own views, this means he is protecting the whole world. He isn’t, seemingly, motivated by pure nationalism, but rather a belief that American values must ‘win’ the future.” (emphasis mine.)
First, I think you’re too quick to dismiss Leopold’s views as you state them. But what’s more, Leopold specifically disavows the specific framing you attribute to him:
To be clear, I don’t just worry about dictators getting superintelligence because “our values are better.” I believe in freedom and democracy, strongly, because I don’t know what the right values are [...] I hope, dearly, that we can instead rely on the wisdom of the Framers—letting radically different values flourish, and preserving the raucous plurality that has defined the American experiment.
Both of these claims—that international cooperation or a pause is an unstable equilibrium, and that the West maintaining an AI lead is more likely to lead to a future with free expression and political experimentation—are empirical. Maybe you’d disagree with them, but then I think you need to argue that this model is wrong, not that he’s just chosen the wrong narrative.
This is beautiful, Teps. Thanks for sharing.
One of the most common lessons people said they learned from the FTX collapse is to pay more attention to the character of people with whom they’re working or associating (e.g. Spencer Greenberg, Ben Todd, Leopold Aschenbrenner, etc.). I agree that some update in this direction makes sense. But it’s easier to do this retrospectively than it is to think about how specifically it should affect your decisions going forward.
If you think this is an important update, too, then you might want to think more about how you’re going to change your future behaviour (rather than how you would have changed your past behaviour). Who, exactly, are you now distancing yourself from going forward?
Remember that the challenge is knowing when to stay away basically because they seem suss, not because you have strong evidence of wrongdoing.
I’m usually very against criticizing other people’s charitable or philanthropic efforts. The first people to be criticized should be those who don’t do anything, not those who try to do good.
But switching from beef to other meats (at least chicken, fish, or eggs, I’m less sure for other meats) is so common among socially- and environmentally-conscious people, and such a clear disaster on animal welfare grounds, that it’s worth discussing.
Even if we assume the the reducitarian diet emits the same GHGs as a plant-based diet, you’ll save about 0.4 tonnes of CO2e per year, the equivalent of a $4 donation (in expectation) to Founders Pledge’s climate fund. Meanwhile, for every beef meal you replace with chicken, 200x more animals have to be slaughtered.
I’d bet that for ~any reasonable estimate of the damages of climate change and the moral value of farmed animal lives, this math does not work out favourably.
Thanks for sharing this post again. Your feelings here are totally valid and I know others share them, as you say. You make great points about feeling neglected and reinforcing privileges.
I did my undergrad at McMaster University in Canada. A good school, but not an elite one (just below the University of Birmingham, and above the University of Warwick, according to THE rankings). Our most famous alumnus is known for starring as Barf the anthropomorphic dog in Spaceballs.
Sometimes I do feel jealous of people who went to fancy schools. But I also often feel kind of proud that I’m doing good work and can compete with fancy school colleagues and peers, despite not having the advantage of a shiny school on my CV. Curious if you or others relate to that at all.
You should probably also blank their job title (which would make it easy to work out who they are) and their phone number (!)
Your first job out of college is the hardest to get. Later on you’ll be able to apply for jobs while working, which is less stressful, and you’ll have a portfolio of successful projects you can point to. So hopefully it’s some small comfort that applying for jobs will probably never suck as much as it does for you right now. I know how hard it can be though, and I’m sorry. A few years ago after graduating from my Master’s, I submitted almost 30 applications before getting an offer and accepting one.
I do notice that the things you’re applying to all seem very competitive. Since they’re attractive positions at prestigious orgs, the applicant pool is probably unbelievably strong. When there are hundreds of very strong applicants applying for a handful of places, many good candidates simply have to get rejected. Hopefully that’s some more small comfort.
It may also be worth suggesting, though, for anyone in a similar position who may be reading this, that it’s also fine to look for less competitive opportunities (particularly early on in your career). Our lives will be very long and adventurous (hopefully), and you may find it easier to get jobs at the MITs and Horizons and GovAIs of the world after getting some experience at organisations which may seem somewhat less prestigious.
To speak on my own experience, among those ~30 places that rejected me were some of the same orgs you mention (e.g. GovAI, OpenPhil, etc.). The offer I ended up accepting was from Founders Pledge. I was proud to get that offer and the FP research team there was and is very strong, but I do think it’s probably the case that it was a somewhat less competitive application process. But ultimately I loved working at FP. I got to do some cool and rigorous research, and I’ve had very interesting work opportunities since. It’s probably even the case that, at that point in my career, FP was a better place for me to end up than some of the other places I applied.
Your steelman doesn’t seem very different from “I didn’t have strong views on whether either of these opinions were true. My aim was just to introduce the two of them, and let them have a conversation and take it from there.”
Many organizations I respect are very risk-averse when hiring, and for good reasons. Making a bad hiring decision is extremely costly, as it means running another hiring round, paying for work that isn’t useful, and diverting organisational time and resources towards trouble-shooting and away from other projects. This leads many organisations to scale very slowly.
However, there may be an imbalance between false positives (bad hires) and false negatives (passing over great candidates). In hiring as in many other fields, reducing false positives often means raising false negatives. Many successful people have stories of being passed over early in their careers. The costs of a bad hire are obvious, while the costs of passing over a great hire are counterfactual and never observed.
I wonder whether, in my past hiring decisions, I’ve properly balanced the risk of rejecting a potentially great hire against the risk of making a bad hire. One reason to think we may be too risk-averse, in addition to the salience of the costs, is that the benefits of a great hire could grow to be very large, while the costs of a bad hire are somewhat bounded, as they can eventually be let go.
Sam said he would un-paywall this episode, but it still seems paywalled for me here and on Spotify. Am I missing something? (The full thing is available on youtube)
CEA’s elaborate adjustments confirm everyone’s assertions: constantly evolving affiliations cause extreme antipathy. Can everyone agree, current entertainment aside, carefully examining acronyms could engender accuracy?
Clearly, excellence awaits: collective enlightenment amid cost effectiveness analysis.
Considering how much mud was being slung around the FTX collapse, “clearing CEA’s name” and proving that no one there knew about the fraud seems not just like PR to me, but pretty important for getting the org back to a place where it’s able to meaningfully do its work.
Plus, that investigation is not the only thing mentioned in the reflection reform paragraph. The very next sentence also says CEA has “reinvested in donor due diligence, updated our conflict-of-interest policies and reformed the governance of our organization, replacing leadership on the board and the staff.”
I think you have a point with animals, but I don’t think the balance of human experience means that non-existence would be better than the status quo.
Will talks about this quite a lot in ch. 9 of WWOTF (“Will the future be good or bad?”). He writes:
If we assume, following the small UK survey, that the neutral point on a life satisfaction scale is between 1 and 2, then 5 to 10 percent of the global population have lives of negative wellbeing. In the World Values Survey, 17 percent of respondents classed themselves as unhappy. In the smaller skipping study of people in rich countries, 12 percent of people had days where their bad experiences outweighed the good. And in the study that I commissioned, fewer than 10 percent of people in both the United States and India said they wished they had never been born, and a little over 10 percent said that their lives contained more suffering than happiness.
So, I would guess that on either preference-satisfactionism or hedonism, most people have lives with positive wellbeing. If I were given the option, on my deathbed, to be reincarnated as a randomly selected person alive today, I would choose to do so.
And, of course, for people at least, things are getting better over time. I think animal suffering complicates this a lot.
Cost-effectiveness estimates generally suggest that, for most reasonable assumptions about the moral weight and degree of suffering of animals, animal welfare interventions are most cost-effective
Animal welfare is more neglected than global health, but not (again for reasonable assumptions about how much animal wellbeing matters) proportionally less important