Error
Unrecognized LW server error:
Field "fmCrosspost" of type "CrosspostOutput" must have a selection of subfields. Did you mean "fmCrosspost { ... }"?
Unrecognized LW server error:
Field "fmCrosspost" of type "CrosspostOutput" must have a selection of subfields. Did you mean "fmCrosspost { ... }"?
Thank you for writing this post. An evergreen difficulty that applies to discussing topics of such a broad scope is the large number of matters that are relevant, difficult to judge, and where oneâs judgement (whatever it may be) can be reasonably challenged. I hope to offer a crisper summary of why I am not persuaded.
I understand from this the primary motivation of MCE is avoiding AI-based dystopias, with the implied causal chain being along the lines of, âIf we ensure the humans generating the AI have a broader circle of moral concern, the resulting post-human civilization is less likely to include dystopic scenarios involving great multitudes of suffering sentiences.â
There are two considerations that speak against this being a greater priority than AI alignment research: 1) Back-chaining from AI dystopias leaves relatively few occasions where MCE would make a crucial difference. 2) The current portfolio of âEA-basedâ MCE is poorly addressed to averting AI-based dystopias.
Re. 1): MCE may prove neither necessary nor sufficient for ensuring AI goes well. On one hand, AI designers, even if speciesist themselves, might nonetheless provide the right apparatus for value learning such that resulting AI will not propagate the moral mistakes of its creators. On the other, even if the AI-designers have the desired broad moral circle, they may have other crucial moral faults (maybe parochial in other respects, maybe selfish, maybe insufficiently reflective, maybe some mistaken particular moral judgements, maybe naive approaches to cooperation or population ethics, and so on) - even if they do not, there are manifold ways in the wider environment (e.g. arms races), or in terms of technical implementation, that may incur disaster.
It seems clear to me that, pro tanto, the less speciesist the AI-designer, the better the AI. Yet for this issue to be of such fundamental importance to be comparable to AI safety research generally, the implication is of an implausible doctrine of âAI immaculate conceptionâ: only by ensuring we ourselves are free from sin can we conceive an AI which will not err in a morally important way.
Re 2): As Plant notes, MCE does not arise from animal causes alone: global poverty, climate change also act to extend moral circles, as well as propagating other valuable moral norms. Looking at things the other way, one should expect the animal causes found most valuable from the perspective of avoiding AI-based dystopia to diverge considerably from those picked on face-value animal welfare. Companion animal causes are far inferior from the latter perspective, but unclear on the former if this a good way of fostering concern for animals; if the crucial thing is for AI-creators not to be speciest over the general population, targeted interventions like âStart a petting zoo at Deepmindâ look better than broader ones, like the abolition of factory farming.
The upshot is that, even if there are some particularly high yield interventions in animal welfare from the far future perspective, this should be fairly far removed from typical EAA activity directed towards having the greatest near-term impact on animals. If this post heralds a pivot of Sentience Institute to directions pretty orthogonal to the principal component of effective animal advocacy, this would be welcome indeed.
Notwithstanding the above, the approach outlined above has a role to play in some ideal âfar future portfolioâ, and it may be reasonable for some people to prioritise work on this area, if only for reasons of comparative advantage. Yet I aver it should remain a fairly junior member of this portfolio compared to AI-safety work.
Those considerations make sense. I donât have much more to add for/âagainst than what I said in the post.
On the comparison between different MCE strategies, Iâm pretty uncertain which are best. The main reasons I currently favor farmed animal advocacy over your examples (global poverty, environmentalism, and companion animals) are that (1) farmed animal advocacy is far more neglected, (2) farmed animal advocacy is far more similar to potential far future dystopias, mainly just because it involves vast numbers of sentient beings who are largely ignored by most of society. Iâm not relatively very worried about, for example, far future dystopias where dog-and-cat-like-beings (e.g. small, entertaining AIs kept around for companionship) are suffering in vast numbers. And environmentalism is typically advocating for non-sentient beings, which I think is quite different than MCE for sentient beings.
I think the better competitors to farmed animal advocacy are advocating broadly for antispeciesism/âfundamental rights (e.g. Nonhuman Rights Project) and advocating specifically for digital sentience (e.g. a larger, more sophisticated version of People for the Ethical Treatment of Reinforcement Learners). There are good arguments against these, however, such as that it would be quite difficult for an eager EA to get much traction with a new digital sentience nonprofit. (We considered founding Sentience Institute with a focus on digital sentience. This was a big reason we didnât.) Whereas given the current excitement in the farmed animal space (e.g. the coming release of âclean meat,â real meat grown without animal slaughter), the farmed animal space seems like a fantastic place for gaining traction.
Iâm currently not very excited about âStart a petting zoo at Deepmindâ (or similar direct outreach strategies) because it seems like it would produce a ton of backlash because it seems too adversarial and aggressive. There are additional considerations for/âagainst (e.g. I worry that itâd be difficult to push a niche demographic like AI researchers very far away from the rest of society, at least the rest of their social circles; I also have the same traction concern I have with advocating for digital sentience), but this one just seems quite damning.
I agree this is a valid argument, but given the other arguments (e.g. those above), I still think itâs usually right for EAAs to focus on farmed animal advocacy, including Sentience Institute at least for the next year or two.
(FYI for readers, Gregory and I also discussed these things before the post was published when he gave feedback on the draft. So our comments might seem a little rehearsed.)
Wild animal advocacy is far more neglected than farmed animal advocacy, and it involves even larger numbers of sentient beings ignored by most of society. If the superiority of farmed animal advocacy over global poverty along these two dimensions is a sufficient reason for not working on global poverty, why isnât the superiority of wild animal advocacy over farmed animal advocacy along those same dimensions not also a sufficient reason for not working on farmed animal advocacy?
I personally donât think WAS is as similar to the most plausible far future dystopias, so Iâve been prioritizing it less even over just the past couple of years. I donât expect far future dystopias to involve as much naturogenic (nature-caused) suffering, though of course itâs possible (e.g. if humans create large numbers of sentient beings in a simulation, but then let the simulation run on its own for a while, then the simulation could come to be viewed as naturogenic-ish and those attitudes could become more relevant).
I think if one wants something very neglected, digital sentience advocacy is basically across-the-board better than WAS advocacy.
That being said, Iâm highly uncertain here and these reasons arenât overwhelming (e.g. WAS advocacy pushes on more than just the âcare about naturogenic sufferingâ lever), so I think WAS advocacy is still, in Gregoryâs words, an important part of the âfar future portfolio.â And often one can work on it while working on other things, e.g. I think Animal Charity Evaluatorsâ WAS content (e.g. ]guest blog post by Oscar Horta](https://ââanimalcharityevaluators.org/ââblog/ââwhy-the-situation-of-animals-in-the-wild-should-concern-us/ââ)) has helped them be more well-rounded as an organization, and didnât directly trade off with their farmed animal content.
But humanity/âAI is likely to expand to other planets. Wonât those planets need to have complex ecosystems that could involve a lot of suffering? Or do you think it will all be done with some fancy tech thatâll be too different from todayâs wildlife for it to be relevant? Itâs true that those ecosystems would (mostly?) be non-naturogenic but Iâm not that sure that people would care about them, itâd still be animals/âdiseases/âhunger.etc. hurting animals. Maybe itâd be easier to engineer an ecosystem without predation and diseases but that is a non-trivial assumption and suffering could then arise in other ways.
Also, some humans want to spread life to other planets for its own sake and relatively few people need to want that to cause a lot of suffering if no one works on preventing it.
This could be less relevant if you think that most of the expected value comes from simulations that wonât involve ecosystems.
Yes, terraforming is a big way in which close-to-WAS scenarios could arise. I do think itâs smaller in expectation than digital environments that develop on their own and thus are close-to-WAS.
I donât think terraforming would be done very differently than todayâs wildlife, e.g. done without predation and diseases.
Ultimately I still think the digital, not-close-to-WAS scenarios seem much larger in expectation.
In Stuart Russellâs Human Compatible (2019), he advocates for AGI to follow preference utilitarianism, maximally satisfying the values of humans. As for animal interests, he seems to think that they are sufficiently represented since he writes that they will be valued by the AI insofar as humans care about them. Reading this from Stuart Russell shifted me toward thinking that moral circle expansion probably does matter for the long-term future. It seems quite plausible (likely?) that AGI will follow this kind of value function which does not directly care about animals rather than broadly anti-speciesist values, since AI researchers are not generally anti-speciesists. In this case, moral circle expansion across the general population would be essential.
(Another factor is that Russellâs reward modeling depends on receiving feedback occasionally from humans to learn their preferences, which is much more difficult to do with animals. Thus, under an approach similar to reward modeling, AGI developers probably wonât bother to directly include animal preferences, when that involves all the extra work of figuring out how to get the AI to discern animal preferences. And how many AI researchers want to risk, say, mosquito interests overwhelming human interests?)
In comparison, if an AGI was planned to only care about the interests of people in, say, Western countries, that would instantly be widely decried as racist (at least in todayâs Western societies) and likely not be developed. So while moral circle expansion encompasses caring about people in other countries, Iâm less concerned that large groups of humans will not have their interests represented in the AGIâs values than I am about nonhuman animals.
It may be more cost-effective to have targeted approach of increasing anti-speciesism among AI researchers and doing anti-speciesist AI alignment philosophy/âresearch (e.g., more details on how AI following preference utilitarianism can also intrinsically care about animal preferences, accounting for preferences of digital sentience given the problem that they can easily replicate and dominate preference calculations), but anti-speciesism among the general population still seems to be an important component of reducing risk of having a bad far future.
Thanks for funding this research. Notes:
Ostensibly it seems like much of Sentience Instituteâs (SI) current research is focused on identifying those MCE strategies which historically have turned out to be more effective among the strategies which have been tried. I think SI as an organization is based on the experience of EA as a movement in having significant success with MCE in a relatively short period of time. Successfully spreading the meme of effective giving; increasing concern for the far future in notable ways; and corporate animal welfare campaigns are all dramatic achievements for a young social movement like EA. While these arenât on the scale of shaping MCE over the course of the far future, these achievements makes it seem more possible EA and allied movements can have an outsized impact by pursuing neglected strategies for values-spreading.
On terminology, to say the focus is on non-human animals, or even moral patients which typically come to mind when describing âanimal-likeâ minds, i.e., familiar vertebrates is inaccurate. âSentient beingâ, âmoral patientâ or ânon-human agents/âbeingsâ are terms which are inclusive of non-human animals, and other types of potential moral patients posited. Admittedly these arenât catchy terms.
This is something I also struggle with in understanding the post. it seems like we need:
AI creators can be convinced to expand their moral circle
Despite (1), they do not wish to be convinced to expand their moral circle
The AI follows this second desire to not be convinced to expand their moral circle
I imagine this happening with certain religious things; e.g. I could imagine someone saying âI wish to think the Bible is true even if I could be convinced that the Bible is falseâ.
But it seems relatively implausible with regards to MCE?
Particularly given that AI safety talks a lot about things like CEV, it is unclear to me whether there is really a strong trade-off between MCE and AIA.
(Note: Jacy and I discussed this via email and didnât really come to a consensus, so thereâs a good chance I am just misunderstanding his argument.)
Hm, yeah, I donât think I fully understand you here either, and this seems somewhat different than what we discussed via email.
My concern is with (2) in your list. â[T]hey do not wish to be convinced to expand their moral circleâ is extremely ambiguous to me. Presumably you mean theyâwithout MCE advocacy being doneâwouldnât put in wide-MC* values or values that lead to wide-MC into an aligned AI. But I think itâs being conflated with, âthey actively opposeâ or âthey would answer ânoâ if asked, âDo you think your values are wrong when it comes to which moral beings deserve moral consideration?ââ
I think they donât actively oppose it, they would mostly answer ânoâ to that question, and itâs very uncertain if they will put the wide-MC-leading values into an aligned AI. I donât think CEV or similar reflection processes reliably lead to wide moral circles. I think they can still be heavily influenced by their initial set-up (e.g. what the values of humanity when reflection begins).
This leads me to think that you only need (2) to be true in a very weak sense for MCE to matter. I think itâs quite plausible that this is the case.
*Wide-MC meaning an extremely wide moral circle, e.g. includes insects, small/âweird digital minds.
Why do you think this is the case? Do you think there is an alternative reflection process (either implemented by an AI, by a human society, or combination of both) that could be defined that would reliably lead to wide moral circles? Do you have any thoughts on what would it look like?
If we go through some kind of reflection process to determine our values, I would much rather have a reflection process that wasnât dependent on whether or not MCE occurred before hand, and I think not leading to a wide moral circle should be considered a serious bug in any definition of a reflection process. It seems to me that working on producing this would be a plausible alternative or at least parallel path to directly performing MCE.
I think that thereâs an inevitable tradeoff between wanting a reflection process to have certain properties and worries about this violating goal preservation for at least some people. This blogpost is not about MCE directly, but if you think of âBAAN thought experimentâ as âwe do moral reflection and the outcome is such a wide circle that most people think it is extremely counterintuitiveâ then the reasoning in large parts of the blogpost should apply perfectly to the discussion here.
That is not to say that trying to fine tune reflection processes is pointless: I think itâs very important to think about what our desiderata should be for a CEV-like reflection process. Iâm just saying that there will be tradeoffs between certain commonly mentioned desiderata that people donât realize are there because they think there is such a thing as âgenuinely free and open-ended deliberation.â
Thanks for commenting, Lukas. I think Lukas, Brian Tomasik, and others affiliated with FRI have thought more about this, and I basically defer to their views here, especially because I havenât heard any reasonable people disagree with this particular point. Namely, I agree with Lukas that there seems to be an inevitable tradeoff here.
I tend to think of moral values as being pretty contingent and pretty arbitrary, such that what values you start with makes a big difference to what values you end up with even on reflection. People may âimprintâ on the values they receive from their culture to a greater or lesser degree.
Iâm also skeptical that sophisticated philosophical-type reflection will have significant influence over posthuman values compared with more ordinary political/âeconomic forces. I suppose philosophers have sometimes had big influences on human politics (religions, Marxism, the Enlightenment), though not necessarily in a clean âcarefully consider lots of philosophical arguments and pick the best onesâ kind of way.
Iâd qualify this by adding that the philosophical-type reflection seems to lead in expectation to more moral value (positive or negative, e.g. hedonium or dolorium) than other forces, despite overall having less influence than those other forces.
I thought this piece was good. I agree that MCE work is likely quite high impactâperhaps around the same level as X-risk workâand that it has been generally ignored by EAs. I also agree that it would be good for there to be more MCE work going forward. Hereâs my 2 cents:
You seem to be saying that AIA is a technical problem and MCE is a social problem. While I think there is something to this, I think there are very important technical and social sides to both of these. Much of the work related to AIA so far has been about raising awareness about the problem (eg the book Superintelligence), and this is more a social solution than a technical one. Also, avoiding a technological race for AGI seems important for AIA, and this also is more a social problem than a technical one.
For MCE, the 2 best things I can imagine (that I think are plausible) are both technical in nature. First, I expect clean meat will lead to the moral circle expanding more to animals. I really donât see any vegan social movement succeeding in ending factory farming anywhere near as much as I expect clean meat to. Second, Iâd imagine that a mature science of consciousness would increase MCE significantly. Many people donât think animals are conscious, and almost no one thinks anything besides animals can be conscious. How would we even know if an AI was conscious, and if so, if it was experiencing joy or suffering? The only way would be if we develop theories of consciousness that we have high confidence in. But right now weâre very limited in studying consciousness, because our tools at interfacing with the brain are crude. Advanced neurotechnologies could change thatâthey could allow us to potentially test hypotheses about consciousness. Again, developing these technologies would be a technical problem.
Of course, these are just the first ideas that come into my mind, and there very well may be social solutions that could do more than the technical solutions I mentioned, but I donât think we should rule out the potential role of technical solutions, either.
Thanks for the comment! A few of my thoughts on this:
If one is convinced non-extinction civilization is net positive, this seems true and important. Sorry if I framed the post too much as one or the other for the whole community.
Maybe. My impression from people working on AIA is that they see it as mostly technical, and indeed they think much of the social work has been net negative. Perhaps not Superintelligence, but at least the work thatâs been done to get media coverage and widespread attention without the technical attention to detail of Bostromâs book.
I think the more important social work (from a pro-AIA perspective) is about convincing AI decision-makers to use the technical results of AIA research, but my impression is that AIA proponents still think getting those technical results is probably the more important projects.
Thereâs also social work in coordinating the AIA community.
Sure, though one big issue with technology is that it seems like we can do far less to steer its direction than we can do with social change. Clean meat tech research probably just helps us get clean meat sooner instead of making the tech progress happen when it wouldnât otherwise. The direction of the far future (e.g. whether clean meat is ever adopted, whether the moral circle expands to artificial sentience) probably matters a lot more than the speed at which it arrives.
Of course, this gets very complicated very quickly, as we consider things like value lock-in. Sentience Institute has a bit of basic sketching on the topic on this page.
I disagree that âmany people donât think animals are conscious.â I almost exclusively hear that view in from the rationalist/âLessWrong community. A recent survey suggested that 87.3% of US adults agree with the statement, âFarmed animals have roughly the same ability to feel pain and discomfort as humans,â and presumably even more think they have at least some ability.
Iâm fairly skeptical of this personally, partly because I donât think thereâs a fact of the matter when it comes to whether a being is conscious. I think Brian Tomasik has written eloquently on this. (I know this is an unfortunate view for an animal advocate like me, but it seems to have the best evidence favoring it.)
I would guess that increasing understanding of cognitive science would generally increase peopleâs moral circles if only because people would think more about these kinds of questions. Of course, understanding cognitive science is no guarantee that youâll conclude that animals matter, as we can see from people like Dennett, Yudkowsky, Peter Carruthers, etc.
Agreed.
I think thatâs right. Specifically, I would advocate consciousness research as a foundation for principled moral circle expansion. I.e., if we do consciousness research correctly, the equations themselves will tell us how conscious insects are, whether algorithms can suffer, how much moral weight we should give animals, and so on.
On the other hand, if there is no fact of the matter as to what is conscious, weâre headed toward a very weird, very contentious future of conflicting/âincompatible moral circles, with no âground truthâ or shared principles to arbitrate disputes.
Edit: Iâd also like to thank Jacy for posting this- I find it a notable contribution to the space, and clearly a product of a lot of hard work and deep thought.
Thanks for writing this, I thought it was a good article. And thanks to Greg for funding it.
My pushback would be on the cooperation and coordination point. It seems that a lot of other people, with other moral values, could make a very similar argument: that they need to promote their values now, as the stakes as very high with possible upcoming value lock-in. To people with those values, these arguments should seem roughly as important as the above argument is to you.
Christians could argue that, if the singularity is approaching, it is vitally important that we ensure the universe wonât be filled with sinners who will go to hell.
Egalitarians could argue that, if the singularity is approaching, it is vitally important that we ensure the universe wonât be filled with wider and wider diversities of wealth.
Libertarians could argue that, if the singularity is approaching, it is vitally important that we ensure the universe wonât be filled with property rights violations.
Naturalists could argue that, if the singularity is approaching, it is vitally important that we ensure the beauty of nature wonât be bespoiled all over the universe.
Nationalists could argue that, if the singularity is approaching, it is vitally important that we ensure the universe will be filled with people who respect the flag.
But it seems that it would be very bad if everyone took this advice literally. We would all end up spending a lot of time and effort on propaganda, which would probably be great for advertising companies but not much else, as so much of it is zero sum. Even though it might make sense, by their values, for expanding-moral-circle people and pro-abortion people to have a big propaganda war over whether foetuses deserve moral consideration, it seems plausible weâd be better off if they both decided to spend the money on anti-malaria bednets.
In contrast, preventing the extinction of humanity seems to occupy a privileged positionânot exactly comparable with the above agendas, though I canât exactly cache out why it seems this way to me. Perhaps to devout Confucians a pre-occupation with preventing extinction seems to be just another distraction from the important task of expressing filial piety â though I doubt this.
(Moral Realists, of course, could argue that the situation is not really symmetric, because promoting the true values is distinctly different from promoting any other values.)
Yeah, I think thatâs basically right. I think moral circle expansion (MCE) is closer to your list items than extinction risk reduction (ERR) is because MCE mostly competes in the values space, while ERR mostly competes in the technology space.
However, MCE is competing in a narrower space than just values. Itâs in the MC space, which is just the space of advocacy on what our moral circle should look like. So I think itâs fairly distinct from the list items in that sense, though you could still say theyâre in the same space because all advocacy competes for news coverage, ad buys, recruiting advocacy-oriented people, etc. (Technology projects could also compete for these things, though there are separations, e.g. journalists with a social beat versus journalists with a tech beat.)
I think the comparably narrow space of ERR is ER, which also includes people who donât want extinction risk reduced (or even want it increased), such as some hardcore environmentalists, antinatalists, and negative utilitarians.
I think these are legitimate cooperation/âcoordination perspectives, and itâs not really clear to me how they add up. But in general, I think this matters mostly in situations where you actually can coordinate. For example, in the US general election when Democrats and Republicans come together and agree not to give to their respective campaigns (in exchange for their counterpart also not doing so). Or if there were anti-MCE EAs with whom MCE EAs could coordinate (which I think is basically what youâre saying with âweâd be better off if they both decided to spend the money on anti-malaria bednetsâ).
Fortunately, not everyone does take this advice literally :).
This is very similar to the tragedy of the commons. If everyone acts out of their own self motivated interests, then everyone will be worse off. However, the situation as you described does not fully reflect reality because none of the groups you mentioned are actually trying to influence AI researchers at the moment. Therefore, MCE has a decisive advantage. Of course, this is always subject to change.
I find that it is often the case that people will dismiss any specific moral recommendation for AI except this one. Personally I donât see a reason to think that there are certain universal principles of minimal alignment. You may argue that human extinction is something that almost everyone agrees is badâbut now the principle of minimal alignment has shifted to âhave the AI prevent things that almost everyone agrees is badâ which is another privileged moral judgement that I see no intrinsic reason to hold.
In truth, I see no neutral assumptions to ground AI alignment theory in. I think this is made even more difficult because even relatively small differences in moral theory from the point of view of information theoretic descriptions of moral values can lead to drastically different outcomes. However, I do find hope in moral compromise.
As EA as a movement has grown so far, the community appears to converge upon a rationalization process whereby most of us have realized what is centrally morally important is the experiences of well-being of a relatively wide breadth of moral patients, and the relatively equal moral weight assigned to well-being of each moral patient. The difference between SI and those who focus on AIA is primarily their differing estimates of the expected value of far-future in terms of average or total well-being. Among the examples you provided, it seems some worldviews are more amenable to the rationalization process which lends itself to consequentialism and EA. Many community members were egalitarians and libertarians who find common cause now in trying to figure out if to focus on AIA or MCE. I think your point is important in that ultimately advocating for this type of values spreading could be bad. However what appears to be an extreme amount of diversity could end up looking less fraught in a competition among values as divergent worldviews converge on similar goals.
Since different types of worldviews, like any amenable to aggregate consequentialist frameworks, can collate around a single goal of something like MCE. The relevance of your point, then, would hinge upon how universal MCE really is or can be across worldviews, relative to other types of values, such that it wouldnât clash with many worldviews in a values-spreading contest. This is a matter of debate I havenât thought of. It seems an important way to frame solutions to the challenge to Jacyâs point you raise.
Thanks for this post. Some scattered thoughts:
This doesnât seem like a big consideration to me. Even if unfriendly AI comes sooner by an entire decade, this matters little on a cosmic timescale. An argument I find more compelling: If we plot the expected utility of an AGI as a function of the amount of effort put into aligning it, there might be a âvalley of bad alignmentâ that is worse than no attempt at alignment at all. (A paperclip maximizer will quickly kill us and not generate much long-term suffering, whereas an AI that understands the importance of human survival but doesnât understand any other values will imprison us for all eternity. Something like that.)
Iâd like to know more about why people think that our moral circles have expanded. I suspect activism plays a smaller role than you think. Steven Pinker talks about possible reasons for declining violence in his book The Better Angels of Our Nature. Iâm guessing this is highly related to moral circle expansion.
One theory I havenât seen elsewhere is that self-interest plays a big role in moral circle expansion. Consider the example of slavery. The BBC writes:
Youâll note that the villains of the slave story are the slaversâpeople with an interest in slavery. The heroes seem to have been Britons who would not lose much if slavery was outlawed (though I guess boycotting sugar would go against their self-interest?) Similarly, I think I remember reading that poor northern whites were motivated to fight in the US Civil War because they were worried their labor would be displaced by slave labor.
According to this story, the expanding circle is a side effect of the world growing wealthier. As lower levels of Maslowâs hierarchy are met, people care more about humanitarian issues. (Iâm assuming that genetic relatedness predicts where on the hierarchy another being falls.) Conquest is less common now because itâs more profitable to control a multinational company than control lots of territory. Slavery is less common because unskilled laborers are less of an asset & more of a liability, and itâs hard to coerce skilled labor. Violence has declined because sub-replacement fertility means weâre no longer in a zero-sum competition for resources. (Note that the bloodiest war in recent memory happened in the Democratic Republic of Congo, a country where women average six children eachâsource. Congo has a lot of mineral wealth, which seems to incentivize conflict. Probably this wealth doesnât diminish as much in the presence of conflict as much as e.g. manufacturing wealth would.)
I suppose a quick test for the Maslowâs hierarchy story is to check whether wealthy people are more likely to be vegan (controlling for meat calories being as expensive as non-meat calories).
I donât think everyone is completely self-interested all the time, but I think people are self-interested enough that it makes sense for activists to apply leverage strategically.
Re: a computer program used to mine asteroids, Iâd expect certain AI alignment work to be useful here. If we understand AI algorithms more deeply, an asteroid miner can be simpler and less likely sentient. Contrast with the scenario where AI progress is slow, brain emulations come before AGI, and the asteroid miner is piloted using an emulation of someoneâs brain.
Iâm not comfortable relying on innate human goodness to deal with moral dilemmas. Iâd rather eliminate incentives for immoral behavior. In the presence of bad incentives, I worry about activism backfiring as people come up with rationalizations for their immoral behavior. See e.g. biblical justifications for slavery in the antebellum south. Instead of seeing the EA movement as something that will sweep the globe and make everyone altruistic, Iâm more inclined to see it as a team of special forces working to adjust the incentives that everyone else operates under in order to create good outcomes as a side effect of everyone else working towards their incentives.
Another moral circle expansion story involves improved hygiene. See also.
Singer and Pinker talk a lot about the importance of reason and empathy to the expanding moral circle. This might be achieved through better online discussion platforms, widespread adoption of meditation, etc.
Anyway, I think that if we take a broad view of moral circle expansion, the best way to achieve it might be some unexpected thing: improving the happiness of voters who control nuclear weapons, helping workers deal with technological job displacement, and so on. IMO, more EAs should work on world peace.
I come back to this post quite frequently when considering whether to prioritize MCE (via animal advocacy) or AI safety. It seems that these two cause areas often attract quite different people with quite different objectives, so this post is unique in its attempt to compare the two based on the same long-term considerations.
I especially like the discussion of bias. Although some might find the whole discussion a bit ad hominem, I think people in EA should take seriously the worry that certain features common in the EA community (e.g., an attraction towards abstract puzzles) might bias us towards particular cause areas.
I recommend this post for anyone interested in thinking more broadly about longtermism.
So Iâm curious for your thoughts. I see this concern about âincidental suffering of worker-agentsâ stated frequently, which may be likely in many future scenarios. However, it doesnât seem to be a crucial consideration, specifically because I care about small/âweird minds with non-complex experiences (your first consideration).
Caring about small minds seems to imply that âOpportunity Cost/âLost Risksâ are the dominate considerationâif small minds have moral value comparable to large minds, then the largest-EV risk is not optimizing for small minds and wasting resources thrown at large minds with complex/âexpensive experiences (or thrown at something even less efficient, like biological beings, any non-total-consequentialist view, etc). This would you lose you many orders of magnitude of optimized happiness, and this loss would be worse than the other scenariosâ aggregate incidental suffering. Even if this inefficient moral position merely reduced optimized happiness by 10% - far less than an order of magnitudeâthis would dominate incidental suffering, even if the incidental suffering scenarios were significantly more probable. And even if you very heavily weight suffering compared to happiness, my math still suggests this conclusion survives by a significant margin).
Also note that Moral Circle Expansion is relevant conditional on solving the alignment problem, so weâre in the set of worlds where the alignment problem was actually solved in some way (humanityâs values are somewhat intact). So, the risk is that whatever-weâre-optimizing-the-future-for is far less efficient than ideal hedonium could have been, because weâre wasting it on complex minds, experiences that require lots of material input, or other not-efficiently-value-creating things. âOh, what might have beenâ, etc. Note this still says values spreading might be very important, but I think this version has a slightly different flavor that implies somewhat different actions. Thoughts?
Next to the counterpoints mentioned by Gregory Lewis, I think there is an additional reason why MCE seems less effective than more targeted interventions to improve the quality of the long-term future: Gains from trade between humans with different values become easier to implement as the reach of technology increases. As long as a non-trivial fraction of humans end up caring about animal wellbeing or digital minds, it seems likely it would be cheap for other coalitions to offer trades. So whether 10% of future people end up with an expanded moral circle or 100% may not make much of a difference to the outcome: It will be reasonably good either way if people reap the gains from trade.
One might object that it is unlikely that humans would be able to cooperate efficiently, given that we donât see this type of cooperation happening today. However, I think itâs reasonable to assume that staying in control of technological progress beyond the AGI transition requires a degree of wisdom and foresight that is very far away from where most societal groups are at today. And if humans do stay in control, then finding a good solution for value disagreements may be the easier problem, or at worst similarly hard. So it feels to me that most likely, we either we get a future that goes badly for reasons related to lack of coordination and sophistication in the pre-AGI stage, or we get a future where humans set things up wisely enough to actually design an outcome that is nice (or at least not amongst the 10% of worst outcomes) by the lights of nearly everyone.
Brian Tomasik made the point that conditional on human values staying in control, we may be very unlikely to get something like broad moral reflection. Instead, values could be determined by a very small group of individuals who happened to be in power by the time AGI arrives (as opposed to individuals ending up there because they were unusually foresighted and also morally motivated). This feels possible too, but it seems to not be the likely default to me because I suspect that youâd need to necessarily increase your philosophical sophistication in order to stay in control of AGI, and that probably gives you more pleasant outcomes (correlational claim). Iterated amplification for instance, as an approach to AI alignment, has several uses for humans: Humans are not only where the resulting values come from, but theyâre also in charge of keeping the bootstrapping process on track and corrigible. And as this post on factored cognition illustrates, this requires sophistication to set up. So if thatâs the bar that AGI creators need to pass before they can determine how âhuman valuesâ are to be extrapolated, maybe we shouldnât be too pessimistic about the outcome. It seems kind of unlikely that someone would go through all of that only to be like âIâm going to implement my personal best guess about what matters to me, with little further reflection, and no other humans get a say here.â Similarly, it also feels unlikely that people would go through with all that and not find a way to make subparts of the population reasonably content about how sentient subroutines are going to be used.
Now, I feel a bit confused about the feasibility of AI alignment if you were to do it somewhat sloppily and with lower standards. I think that thereâs a spectrum from âit just wouldnât work at all and not be competitiveâ (and then people would have to try some other approach) to âit would produce a capable AGI but it would be vulnerable to failure modes like adversarial exploits or optimization daemons, and so it would end up with not human valuesâ. These failure modes, to the very small degree I currently understand them, sound like they would not be sensitive to whether the human whose approval you tried to approximate had an expanded moral circle or not. I might be wrong about that. If people mostly want sophisticated alignment procedures because they care about preserving the option for philosophical reflection, rather than because they also think that you simply run into large failure modes otherwise, then it seems like (conditional on some kind of value alignment) whether we get an outcome with broad moral reflection is not so clear. If itâs technically easier to build value-aligned AI with very parochial values, then MCE could make a relevant difference to these non-reflection outcomes.
But all in all my argument is that itâs somewhat strange to assume that a group of people could succeed at building an AGI optimized for its creatorsâ values, without having to put in so much thinking about how to get this outcome right that theyâd almost canât help but become reasonably philosophically sophisticated in the process. And sure, philosophically sophisticated people can still have fairly strange values by your own lights, but it seems like thereâs more convergence. Plus Iâd at least be optimistic about their propensity to strive towards positive-sum outcomes, given how little scarcity youâd have if the transition does go well.
Of course, maybe value-alignment is going to work very differently from what people currently think. The main way Iâd criticize my above points is that theyâre based on heavy-handed inside-view thinking about how difficult I (and others Iâm updating towards) expect the AGI transition to be. If AGI will be more like the Industrial Revolution rather than something that is even more difficult to stay remotely in control of, or if some other technology proves to be more consequential than AGI, then my argument has less force. I mainly see this as yet another reason to caveat that the ex ante plausible-seeming position that MCE can have a strong impact on AGI outcomes starts to feel more and more conjunctive the more you zoom in and try to identify concrete pathways.
Interesting points. :) I think there could be substantial differences in policy between 10% support and 100% support for MCE depending on the costs of appeasing this faction and how passionate it is. Or between 1% and 10% support for MCE applied to more fringe entities.
Iâm not sure if sophistication increases convergence. :) If anything, people who think more about philosophy tend to diverge more and more from commonsense moral assumptions.
Yudkowsky and I seem to share the same metaphysics of consciousness and have both thought about the topic in depth, yet we occupy almost antipodal positions on the question of how many entities we consider moral patients. I tend to assume that oneâs starting points matter a lot for what views one ends up with.
I agree with this. It seems like the world where Moral Circle Expansion is useful is the world where:
The creators of AI are philosophically sophisticated (or persuadable) enough to expand their moral circle if they are exposed to the right arguments or work is put into persuading them.
They are not philosophically sophisticated enough to realize the arguments for expanding the moral circle on their own (seems plausible).
They are not philosophically sophisticated enough to realize that they might want to consider a distribution of arguments that they could have faced and could have persuaded them about what is morally right, and design AI with this in mind (ie CEV), or with the goal of achieving a period of reflection where they can sort out the sort of arguments that they would want to consider.
I think Iâd prefer pushing on point 3, as it also encompasses a bunch of other potential philosophical mistakes that AI creators could make.
This post is extremely valuableâthank you! You have caused me to reexamine my views about the expected value of the far future.
What do you think are the best levers for expanding the moral circle, besides donating to SI? Is there anything else outside of conventional EAA?
Thanks! Thatâs very kind of you.
Iâm pretty uncertain about the best levers, and I think research can help a lot with that. Tentatively, I do think that MCE ends up aligning fairly well with conventional EAA (perhaps it should be unsurprising that the most important levers to push on for near-term values are also most important for long-term values, though it depends on how narrowly youâre drawing the lines).
A few exceptions to that:
Digital sentience probably matters the most in the long run. There are good reasons to be skeptical we should be advocating for this now (e.g. itâs quite outside of the mainstream so it might be hard to actually get attention and change minds; itâd probably be hard to get funding for this sort of advocacy (indeed thatâs one big reason SI started with farmed animal advocacy)), but Iâm pretty compelled by the general claim, âIf you think X value is what matters most in the long-term, your default approach should be working on X directly.â Advocating for digital sentience is of course neglected territory, but Sentience Institute, the Nonhuman Rights Project, and Animal Ethics have all worked on it. People for the Ethical Treatment of Reinforcement Learners has been the only dedicated organization AFAIK, and Iâm not sure what their status is or if theyâve ever paid full-time or part-time staff.
I think views on value lock-in matter a lot because of how they affect food tech (e.g. supporting The Good Food Institute). I place significant weight on this and a few other things (see this section of an SI page) that make me think GFI is actually a pretty good bet, despite my concern that technology progresses monotonically.
Because what might matter most is societyâs general concern for weird/âsmall minds, we should be more sympathetic to indirect antispeciesism work like that done by Animal Ethics and the fundamental rights work of the Nonhuman Rights Project. From a near-term perspective, I donât think these look very good because I donât think weâll see fundamental rights be a big reducer of factory farm suffering.
This is a less-refined view of mine, but Iâm less focused than I used to be on wild animal suffering. It just seems to cost a lot of weirdness points, and naturogenic suffering doesnât seem nearly as important as anthropogenic suffering in the far future. Factory farm suffering seems a lot more similar to far future dystopias than does wild animal suffering, despite WAS dominating utility calculations for the next, say, 50 years.
I could talk more about this if youâd like, especially if youâre facing specific decisions like where exactly to donate in 2018 or what sort of job youâre looking for with your skillset.
I thought this was very interesting, thanks for writing up. Two comments
It was useful to have a list of reasons why you think the EV of the future could be around zero, but it still found it quite vague/âhard to imagineâwhy exactly would more powerful minds be mistreating less powerful minds? etc. - so Iâd would have liked to see that sketched in slightly more depth.
Itâs not obvious to me itâs correct/âcharitable to draw the neglectedness of MCE so narrowly. Canât we conceive of a huge ammount of moral philosophy, and well as social activism, both new and old, as MCE? Isnât all EA outreach an indirect form of MCE?
Iâm sympathetic to both of those points personally.
1) I considered that, and in addition to time constraints, I know others havenât written on this because thereâs a big concern of talking about it making it more likely to happen. I err more towards sharing it despite this concern, but Iâm pretty uncertain. Even the detail of this post was more than several people wanted me to include.
But mostly, Iâm just limited on time.
2) Thatâs reasonable. I think all of these boundaries are fairly arbitrary; we just need to try to use the same standards across cause areas, e.g. considering only work with this as its explicit focus. Theoretically, since Neglectedness is basically just a heuristic to estimate how much low-hanging fruit there is, weâre aiming at âThe space of work that might take such low-hanging fruit away.â In this sense, Neglectedness could vary widely. E.g. thereâs limited room for advocating (e.g. passing out leaflets, giving lectures) directly to AI researchers, but this isnât affected much by advocacy towards the general population.
I do think moral philosophy that leads to expanding moral circles (e.g. writing papers supportive of utiltiarianism), moral-circle-focused social activism (e.g. anti-racism, not as much something like campaigning for increased arts funding that seems fairly orthogonal to MCE), and EA outreach (in the sense that the A of EA means a wide moral circle) are MCE in the broadest somewhat-useful definition.
Casparâs blog post is a pretty good read on the nuances of defining/âutilizing Neglectedness.
A very interesting and engaging article indeed.
I agree that people often underestimate the value of strategic value spreading. Oftentimes, proposed moral models that AI agents will follow have some lingering narrowness to them, even when they attempt to apply the broadest of moral principles. For instance, in Chapter 14 of Superintelligence, Bostrom highlights his common good principle:
Clearly, even something as broad as that can be controversial. Specifically, it doesnât speak at all about any non-human interests except insofar as humans express widely held beliefs to protect them.
I think one thing to add is that AIA researchers who hold more traditional moral beliefs (as opposed to wide moral circles and transhumanist beliefs) are probably less likely to believe that moral value spreading is worth much. The reason for this is obvious: if everyone around you holds, more or less, the same values that you do, then why change anyoneâs mind? This may explain why many people dismiss the activity you proposed.
Historically it doesnât seem to be true. As AIA becomes more mainstream, itâll be attracting a wider diversity of people, which may induce a form of common grounding and normalization of the values in the community. We should be looking for opportunities to collect data on this in the future to see how attitudes within AIA change. Of course this could lead to attempts to directly influence the proportionate representation of different values within EA. Thatâd be prone to all the hazards of an internal tug of war pointed out in other comments on this post. Because the vast majority of the EA movement focused on the impact of advanced AI on the far future are relatively coordinated and with sufficiently similar goals there isnât much risk of internal fraction in the near future. I think organizations from MIRI to FRI are also averse to growing AIA in ways which drive the trajectory of the field away from what EA currently values.
This is the main reason I think the far future is high EV. I think we should be focusing on p(Hedonium) and p(Delorium) more than anything else. Iâm skeptical that, from a hedonistic utilitarian perspective, byproducts of civilization could come close to matching the expected value from deliberately tiling the universe (potentially multiverse) with consciousness optimized for pleasure or pain. If p(H)>p(D), the future of humanity is very likely positive EV.
You say
Could you elaborate more on why this is the case? I would tend to think that a prior would be that theyâre equal, and then you update on the fact that they seem to be asymmetrical, and try to work out why that is the case, and whether those factors will apply in future. They could be fundamentally asymmetrical, or evolutionary pressures may tend to create minds with these asymmetries. The arguments Iâve heard for why are:
The worst thing that can happen to an animal, in terms of genetic success, is much worse than the best thing.
This isnât entirely clear to me: I can imagine a large genetic win such as securing a large harem could be comparable to the genetic loss of dying, and many animals will in fact risk death for this. This seems particularly true considering that dying leaving no offspring doesnât make your contribution to the gene pool zero, just that itâs only via your relatives.
There is selection against strong positive experiences in a way that there isnât against strong negative experiences.
The argument here is, I think, that strong positive experiences will likely result in the animal sticking in the blissful state, and neglecting to feed, sleep, etc, whereas strong negative experiences will just result in the animal avoiding a particular state, which is less maladaptive. This argument seems stronger to me but still not entirely satisfyingâit seems to be quite sensitive to how you define states.
My current position is that the amount of pleasure/âsuffering that conscious entities will experience in a far-future technological civilization will not be well-defined. Some arguments for this:
Generally utility functions or reward functions are invariant under affine transformations (with suitable rescaling for the learning rate for reward functions). Therefore they cannot be compared between different intelligent agents as a measure of pleasure.
The clean separation of our civilization into many different individuals is an artifact of how evolution operates. I donât expect far future civilization to have a similar division of its internal processes into agents. Therefore the method of counting conscious entities with different levels of pleasure is inapplicable.
Theoretical computer science gives many ways to embed one computational process within another so that it is unclear whether or how many times the inner process âoccursâ, such as running identical copies of the same program, using a quantum computer to run the same program with many inputs in superposition, and homomorphic encryption. Similar methods we donât know about will likely be discovered in the future.
Our notions of pleasure and suffering are mostly defined extensionally with examples from the present and the past. I see no reason that such an extensionally-derived concept to have a natural definition that applies to extremely different situations. Uncharitably, it seems like the main reason people assume this is a sort of wishful thinking due to their normal moral reasoning breaking down if they allow pleasure/âsuffering to be undefined.
Iâm currently uncertain about how to make decisions relating to the far future in light of the above arguments. My current favorite position is to try to understand the far future well enough until I find something I have strong moral intuitions about.
Possibly the biggest unknown in ethics is whether bits matter, or whether atoms matter.
If you assume bits matter, then I think this naturally leads into a concept cluster where speaking about utility functions, preference satisfaction, complexity of value, etc, makes sense. You also get a lot of weird unresolved thought-experiments like homomorphic encryption.
If you assume atoms matter, I think this subtly but unavoidably leads to a very different concept clusterâqualia turns out to be a natural kind instead of a leaky reification, for instance. Talking about the âunity of value thesisâ makes more sense than talking about the âcomplexity of value thesisâ.
TL;DR: I think youâre right that if we assume computationalism/âfunctionalism is true, then pleasure and suffering are inherently ill-defined, not crisp. They do seem well-definable if we assume physicalism is true, though.
Thanks for reminding me that I was implicitly assuming computationalism. Nonetheless, I donât think physicalism substantially affects the situation. My arguments #2 and #4 stand unaffected; you have not backed up your claim that qualia is a natural kind under physicalism. While itâs true that physicalism gives clear answers for the value of two identical systems or a system simulated with homomorphic encryption, it may still be possible to have quantum computations involving physically instantiated conscious beings, by isolating the physical environment of this being and running the CPT reversal of this physical system after an output has been extracted to maintain coherence. Finally, physicalism adds its own questions, namely, given a bunch of physical systems that all appear to have behavior that appears to be conscious, which ones are actually conscious and which are not. If I understood you correctly, physicalism as a statement about consciousness is primary a negative statement, âthe computational behavior of a system is not sufficient to determine what sort of conscious activity occurs thereâ, which doesnât by itself tell you what sort of conscious activity occurs.
It seems to me your #2 and #4 still imply computationalism and/âor are speaking about a straw man version of physicalism. Different physical theories will address your CPT reversal objection differently, but it seems pretty trivial to me.
I would generally agree, but would personally phrase this differently; rather, as noted here, there is no objective fact-of-the-matter as to what the âcomputational behaviorâ of a system is. I.e., no way to objectively derive what computations a physical system is performing. In terms of a positive statement about physicalism & qualia, Iâm assuming something on the order of dual-aspect monism /â neutral monism. And yes insofar as a formal theory of consciousness which has broad predictive power would depart from folk intuition, Iâd definitely go with the formal theory.
Thanks for the link. I didnât think to look at what other posts you have published and now I understand your position better.
As I now see it, there two critical questions for distinguishing the different positions on the table:
Does our intuitive notion of pleasure/âsuffering have objective precisely defined fundamental concept underlying it?
In practice, is it a useful approach to look for computational structures exhibiting pleasure/âsuffering in the distant future as a means to judge possible outcomes?
Brian Tomasik answers these questions âNo/âYesâ, and a supporter of the Sentience Institute would probably answer âYesâ to the second question. Your answers are âYes/âNoâ, and so you prefer to work on finding the underlying theory for pleasure/âsuffering. My answers are âNo/âNoâ, and am at a loss.
I see two reasons why a person might think that pleasure/âpain of conscious entities is a solid enough concept to answer âYesâ to either of these questions (not counting conservative opinions over what futures are possible for question 2). The first is a confusion caused by subtle implicit assumptions in the way we talk about consciousness, which makes a sort of conscious experience from which includes in it pleasure and pain seem more ontologically basic than it really is. I wonât elaborate on this in this comment, but for now you can round me as an eliminativist.
The second is what I was calling âa sort wishful thinkingâ in argument #4: These people have moral intuitions that tell them to care about othersâ pleasure and pain, which implies not fooling themselves about how much pleasure and pain others experience. On the other hand, there are many situations where their intuition does not give them a clear answer, but also tells them that picking an answer arbitrarily is like fooling themselves. They resolve this tension by telling themselves, âthere is a âcorrect answerâ to this dilemma, but I donât know what it is. I should act to best approximate this âcorrect answerâ with the information I have.â People then treat these âcorrect answersâ like other things they are ignorant about, and in particular imagine that a scientific theory might be able to answer these questions in the same way science answered other things we used to be ignorant about.
However, this expectation infers something external, the existence of a certain kind of scientific theory, from evidence that is internal, their own cognitive tensions. This seems fallacious to me.
Thanks, this is helpful. My general position on your two questions is indeed âYes/âNoâ.
The question of âwhat are realityâs natural kinds?â is admittedly complex and thereâs always room for skepticism. That said, Iâd suggest the following alternatives to your framing:
Whether the existence of qualia itself is âcrispâ seems prior to whether pain/âpleasure are. I call this the âreal problemâ of consciousness.
Iâm generally a little uneasy with discussing pain/âpleasure in technically precise contexts- I prefer âemotional valenceâ.
Another reframe to consider is to disregard talk about pain/âpleasure, and instead focus on whether value is well-defined on physical systems (i.e. the subject of Tegmarkâs worry here). Conflation of emotional valence & moral value can then be split off as a subargument.
Generally speaking, I think if one accepts that itâs possible in principle to talk about qualia in a way that âcarves reality at the jointsâ, itâs not much of a stretch to assume that emotional valence is one such natural kind (arguably the âc. elegans of qualiaâ). I donât think weâre logically forced to assume this, but I think itâs prima facie plausible, and paired with some of our other work it gives us a handhold for approaching qualia in a scientific/âpredictive/âfalsifiable way.
Essentially, QRI has used this approach to bootstrap the worldâs first method for quantifying emotional valence in humans from first principles, based on fMRI scans. (It also should work for most non-human animals; itâs just harder to validate in that case.) We havenât yet done the legwork on connecting future empirical results here back to the computationalism vs physicalism debate, but itâs on our list.
TL;DR: If consciousness is a âcrispâ thing with discoverable structure, we should be able to build/âpredict useful things with this that cannot be built/âpredicted otherwise, similar to how discovering the structure of electromagnetism let us build/âpredict useful things we could not have otherwise. This is probably the best route to solve these metaphysical disagreements.
It wasnât clear to me from your comment, but based on your link I am presuming that by âcrispâ you mean âamenable to generalizable scientific theoriesâ (rather than âontologically basicâ). I was using âpleasure/âpainâ as a catch-all term and would not mind substituting âemotional valenceâ.
Itâs worth emphasizing that just because a particular feature is crisp does not imply that it generalizes to any particular domain in any particular way. For example, a single ice crystalline has a set of directions in which the molecular bonds are oriented which is the same throughout the crystal, and this surely qualifies as a âcrispâ feature. Nonetheless, when the ice melts, this feature becomes undefinedâno direction is distinguished from any other direction in water. When figuring out whether a concept from one domain extends to a new domain, to posit that thereâs a crisp theory describing the concept does not answer this question without any information on what that theory looks like.
So even if there existed a theory describing qualia and emotional valence as it exists on Earth, it need not extend to being able to describe every physically possible arrangement of matter, and I see no reason to expect it to. Since a far future civilization will be likely to approach the physical limits of matter in many ways, we should not assume that it is not one such arrangement of matter where the notion of qualia is inapplicable.
This is an important point and seems to hinge on the notion of reference, or the question of how language works in different contexts. The following may or may not be new to you, but trying to be explicit here helps me think through the argument.
Mostly, words gain meaning from contextual embedding- i.e. theyâre meaningful as nodes in a larger network. Wittgenstein observed that often, philosophical confusion stems from taking a perfectly good word and trying to use it outside its natural remit. His famous example is the question, âwhat time is it on the sun?â. As you note, maybe notions about emotional valence are similar- trying to âuniversalizeâ valence may be like trying to universalize time-zones, an improper move.
But thereâs another notable theory of meaning, where parts of language gain meaning through deep structural correspondence with reality. Much of physics fits this description, for instance, and itâs not a type error to universalize the notion of the electromagnetic force (or electroweak force, or whatever the fundamental unification turns out to be). I am essentially asserting that qualia is like this- that we can find universal principles for qualia that are equally and exactly true in humans, dogs, dinosaurs, aliens, conscious AIs, etc. When I note Iâm a physicalist, I intend to inherit many of the semantic properties of physics, how meaning in physics âworksâ.
I suspect all conscious experiences have an emotional valence, in much the same way all particles have a charge or spin. I.e. itâs well-defined across all physical possibilities.
Do you think we should move the conversation to private messages? I donât want to clutter a discussion thread thatâs mostly on a different topic, and Iâm not sure whether the average reader of the comments benefits or is distracted by long conversations on a narrow subtopic.
Your comment appears to be just reframing the point I just made in your own words, and then affirming that you believe that the notion of qualia generalizes to all possible arrangements of matter. This doesnât answer the question, why do you believe this?
By the way, although there is no evidence for this, it is commonly speculated by physicists that the laws of physics allow multiple metastable vacuum states, and the observable universe only occupies one such vacuum, and near different vacua there different fields and forces. If this is true then the electromagnetic field and other parts of the Standard Model are not much different from my earlier example of the alignment of an ice crystal. One reason this view is considered plausible is simply the fact that itâs possible: Itâs not considered so unusual for a quantum field theory to have multiple vacuum states, and if the entire observable universe is close to one vacuum then none of our experiments give us any evidence on what other vacuum states are like or whether they exist.
This example is meant to illustrate a broader point: I think that making a binary distinction between contextual concepts and universal concepts is oversimplified. Rather, hereâs how I would put it: Many phenomena generalize beyond the context in which they were originally observed. Taking advantage of this, physicists deliberate seek out the phenomena that generalize as far as possible, and over history broadened their grasp very far. Nonetheless, they avoid thinking about any concept as âuniversalâ, and often when they do think a concept generalizes they have a specific explanation for why it should, while if thereâs a clear alternative to the concept generalizing they keep an open mind.
So again: Why do you think that qualia and emotional valence generalize to all possible arrangements of matter?
EA forum threads auto-hide so Iâm not too worried about clutter.
I donât think youâre fully accounting for the difference in my two models of meaning. And, I think the objections you raise to consciousness being well-defined would also apply to physics being well-defined, so your arguments seem to prove too much.
To attempt to address your specific question, I find the hypothesis that âqualia (and emotional valence) are well-defined across all arrangements of matterâ convincing because (1) it seems to me the alternative is not coherent (as I noted in the piece on computationalism I linked for you) and (2) it seems generative and to lead to novel and plausible predictions I think will be proven true (as noted in the linked piece on quantifying bliss and also in Principia Qualia).
All the details and sub arguments can be found in those links.
Will be traveling until Tuesday; probably with spotty internet access until then.
I havenât responded to you for so long firstly because I felt like we got to the point in the discussion where itâs difficult to get across anything new and I wanted to be attentive to what I say, and then because after a while without writing anything I became disinclined from continuing. The conversation may close soon.
Some quick points:
My whole point in my previous comment is that the conceptual structure of physics is not what you make it out to be, and so your analogy to physics is invalid. If you want to say that my arguments against consciousness apply equally well to physics you will need to explain the analogy.
My views on consciousness that I mentioned earlier but did not elaborate on are becoming more relevant. It would be a good idea for me to explain them in more detail.
I read your linked piece on quantifying bliss and I am unimpressed. I concur with the last paragraph of this comment.
Thanks very much for writing this, and thanks to Greg for funding it! I think this is a really important discussion. Some slightly rambling thoughts below.
We can think about 3 ways of improving the EV of the far future:
1: Changing incentive structures experienced by powerful agents in the future (e.g. avoiding arms races, power struggles, selection pressures)
2: a) Changing the moral compass of powerful agents in the future in specific directions (e.g. MCE).
b) Indirect ways to improve the moral compass of powerful agents in the future (e.g. philosophy research, education, intelligence/âempathy enhancement)
All of these are influenced both by strategies such as activism, improving institutions, and improving education, as well as by AIA. I am inclined to think of AIA as a particularly high-leverage point at which we can have influence on these.
However, these are issues are widely encountered. Consider 2b: we have to decide how to educate the next generation of humans, and they may well end up with ethical beliefs that are different from ours, so we must judge how much to try and influence or constrain them, and how much to accept that the changes are actually progress. This is similar to the problem of defining CEV: we have some vague idea of the direction in which better values lie (more empathy, more wisdom, more knowledge), but we canât say exactly what the values should be. For this intervention, working on AIA may be more important than activism because it has more leverageâit is likely to be more tractable and have greater influence on the future than the more diffuse ways in we can push on education and intergenerational moral progress.
This framework also suggests that MCE is just one example of a collection of similar interventions. MCE involves pushing for a fairly specific belief and behaviour change on a principle thatâs fairly uncontroversial. You could also imagine similar interventionsâfor instance, helping people reduce unwanted aggressive or sadistic behaviour. We could call this something like âuncontroversial moral progressâ: helping individuals and civilisation to live by their values more. (on a side note: sometimes I think of this as the minimal core of EA: trying to live according to your best guess of whatâs right)
The choice between working on 2a and 2b depends, among other things, on your level of moral uncertainty.
I am inclined to think that AIA is the best way to work on 1 and 2b, as it is a particularly high-leverage intervention point to shape the power structures and moral beliefs that exist in the future. It gives us more of a clean slate to design a good system, rather than having to work within a faulty system.
I would really like to see more work on MCE and other examples of âuncontroversial moral progressâ. Historical case studies of value changes seem like a good starting point, as well as actually testing the tractability of changing peopleâs behaviour.
I also really appreciated your perspective on different transformative AI scenarios, as Iâm worried Iâm thinking about it in an overly narrow way.
Thank you for providing an abstract for your article. I found it very helpful.
(and I wish more authors here would do so as well)
Random thought: (factory farm) animal welfare issues will likely eventually be solved by cultured (lab grown) meat when it becomes cheaper than growing actual animals. This may take a few decades, but social change might take even longer. The article even suggests technical issues may be easier to solve, so why not focus more on that (rather than on MCE)?
I just took it as an assumption in this post that weâre focusing on the far future, since I think basically all the theoretical arguments for/âagainst that have been made elsewhere. Hereâs a good article on it. I personally mostly focus on the far future, though not overwhelmingly so. Iâm at something like 80% far future, 20% near-term considerations for my cause prioritization decisions.
To clarify, the post isnât talking about ending factory farming. And I donât think anyone in the EA community thinks we should try to end factory farming without technology as an important component. Though I think there are good reasons for EAs to focus on the social change component, e.g. there is less for-profit interest in that component (most of the tech money is from for-profit companies, so itâs less neglected in this sense).
Thank you for this piece. I enjoyed reading it and Iâm glad that weâre seeing more people being explicit about their cause-prioritization decisions and opening up discussion on this crucially important issue.
I know that itâs a weak consideration, but I hadnât, before I read this, considered the argument for the scale of values spreading being larger than the scale of AI alignment (perhaps because, as you pointed out, the numbers involved in both are huge) so thanks for bringing that up.
Iâm in agreement with Michael_S that hedonium and delorium should be the most important considerations when weâre estimating the value of the far-future, and from my perspective the higher probability of hedonium likely does make the far-future robustly positive, despite the valid points you bring up. This doesnât necessarily mean that we should focus on AIA over MCE (I donât), but it does make it more likely that we should.
Another useful contribution, though others may disagree, was the biases section: the biases that could potentially favour AIA did resonate with me, and they are useful to keep in mind.
That makes sense. If I were convinced hedonium/âdolorium dominated to a very large degree, and that hedonium was as good as dolorium is bad, I would probably think the far future was at least moderately +EV.
Isnât hedonium inherently as good as dolorium is bad? If itâs not, canât we just normalize and then treat them as the same? I donât understand the point of saying there will be more hedonium than dolorium in the future, but the dolorium will matter more. Theyâre vague and made-up quantities, so canât we just set it so that âmore hedonium than doloriumâ implies âmore good than badâ?
He defines hedonium/âdolorium as the maximum positive/ânegative utility you can generate with a certain amount of energy:
âFor example, I think a given amount of dolorium/âdystopia (say, the amount that can be created with 100 joules of energy) is far larger in absolute moral expected value than hedonium/âutopia made with the same resources.â
Exactly. Let me know if this doesnât resolve things, zdgroff.
While the central thesis to expand oneâs moral circles can be well-enjoyed by the community, this post is not selling it well. This is exemplified by the âOne might be biased towards AIA ifâŠâ section, which makes assumptions about individuals who focus on AI alignment. Further, while the post includes a section on cooperation, it discourages it. [Edit: Prima facie,] the post does not invite critical discussion. Thus,
I would not recommend this post to any readers interested in moral circles expansion, AI alignment, or cooperation.Thus, I would recommend this post to readers interested in moral circles expansion, AI alignment, and cooperation, as long as they are interested in a vibrant discourse.Do you think thereâs a better way to discuss biases that might push people to one cause or another? Or that we shouldnât talk about such potential biases at all?
What do you mean by this post discouraging cooperation?
What do you expect an invitation for critical discussion to look like? I usually take that to be basically implicit when something is posted to the EA Forum, unless the author states otherwise.
Hm, 1) how do you define a bias? What is your reference for evaluation whether something is biased? The objective should be to make the best decisions with available information at any given time while supporting innovation and keeping âopenminded.â This âbiasâ assessment should be conducted to identify harmful actions that individuals deem positive due to their biases and inform overall prioritization decisionmaking rather than seeking to change oneâs perspectives on causes they prefer. This can contribute to systemic change and optimal specialization development by individuals. This is a better way to approach biases.
The section on cooperation discourages collaboration because it understands cooperation as asserting oneâs perspectives where these are not welcome rather than advancing ventures. The part also states: âinsofar as MCE is uncooperative, I think a large number of other EA interventions, including AIA, are similarly uncooperative.â These authorâs assumptions, if not critically examined against evidence, can discourage persons who could be seeking encouragement to cooperate with others in this article from doing so, because one may wish to avoid sharing perspectives where they are not welcome.
An invitation for critical discussion can include an argument for the writingâs relevance to the development of answers to open-ended questions. But I can agree with your point that this can be superfluous, so would add (added) prima facie and edited the conclusion.
Impressive articleâI especially liked the biases section. I would recommend doing a quantitative model of cost effectiveness comparing to AIA, as I have done for global agricultural catastrophes, especially because neglectedness is hard to define in your case.
Thanks for writing it.
Here are my reasons for the belief wild animal/âsmall minds/â⊠suffering agenda is based mostly on errors and uncertainties. Some of the uncertainties should warrant research effort, but I do not believe the current state of knowledge justifies prioritization ofany kind of advocacy or value spreading.
1] The endeavour seems to be based on extrapolating intuitive models far outside the scope for which we have data. The whole suffering calculus is based on extrapolating the concept of suffering far away from the domain for which we have data from human experience.
2] Big part of it seems arbitrary. When expanding the moral circle toward small computational processes and simple systems, why not expand it toward large computational processes and complex systems? E.g. we can think about the DNA based evolution as about large computational/âoptimization processâsuddenly âwild animal sufferingâ has a purpose and traditional environmnet and biodiversity protection efforts make sense.
(Similarly we could argue much âhuman utilityâ is in the larger system structure above individual humans)
3] We do not know how to measure and aggregate utility of mind states. Like, really donât know. E.g. it sems to me completely plausible the utility of 10 people reaching some highly joyful mindstates is the dominanat contribution over all human and animal minds.
4] Part of the reasoning usually seems contradictory. If the human cognitive processes are in the priviledged position of creating meaning in this universe ⊠well, then they are in the priviledged postion, and there _is_ a categorical difference between humans and other minds. If they are not in the priviledged positions, how it comes humans should impose their ideas about meaning on other agents?
5] MCE efforts directed toward AI researchers with the intent of influencing values of some powerful AI may increase x-risk. E.g. if the AI is not âspeciistâ and gives the same weight to satysfing prefrences of all humans and all chicken, the chicken would outnumber humans.
You raise some good points. (The following reply doesnât necessarily reflect Jacyâs views.)
I think the answers to a lot of these issues are somewhat arbitrary matters of moral intuition. (As you said, âBig part of it seems arbitrary.â) However, in a sense, this makes MCE more important rather than less, because it means expanded moral circles are not an inevitable result of better understanding consciousness/âetc. For example, Yudkowskyâs stance on consciousness is a reasonable one that is not based on a mistaken understanding of present-day neuroscience (as far as I know), yet some feel that Yudkowskyâs view about moral patienthood isnât wide enough for their moral tastes.
Another possible reply (that would sound better in a political speech than the previous reply) could be that MCE aims to spark discussion about these hard questions of what kinds of minds matter, without claiming to have all the answers. I personally maintain significant moral uncertainty regarding how much I care about what kinds of minds, and Iâm happy to learn about other peopleâs moral intuitions on these things because my own intuitions arenât settled.
Or if we take a suffering-focused approach to these large systems, then this could provide a further argument against environmentalism. :)
I selfishly consider my moral viewpoint to be âprivilegedâ (in the sense that I prefer it to other peopleâs moral viewpoints), but this viewpoint can have in its content the desire to give substantial moral weight to non-human (and human-but-not-me) minds.
@Matthew_Barnett As a senior electrical engineering student, proficient in a variety of programming languages, I do think and believe that AI is important to think about and discuss. The theoretical threat of a malevolent strong AI would be immense. But that does not mean one has cause or a valid reason to support CS grad students financially.
A large, significant, asteroid collision with Earth would also be quite devastating. Yet, to fund and support aerospace grads does not follow. Perhaps I really mean this: AI safety is an Earning to Give non sequitur.
Lastly, again, there is no evidence or results. Effective Altruism is about being beneficent instead of merely benevolent (meaning well). In other words, making decisions off well researched initiatives (e.g., bed nets). Since strong AI does not exist, it does not make sense to support though E2G. (Iâm not saying it will never exist; that is unknown.) Of course, there are medium-term (systematic change) with results that more or less rely on historical-type empiricismâbut thatâs still some type of evidence. For poverty we have RCTs and developmental economics. For AI safety [something?]. For animal suffering we have proof that less miserable conditions can become a reality.
I donât think anyone here is suggesting supporting random CS grads financially. Although, they might endorse something like that indirectly by funding AI alignment research, which tends to attract CS grads.
I agree that simply because an asteroid collision would be devastating, it does not follow that we should necessarily focus on that work in particular. However, there are variables which I think you might be overlooking.
The reason why people are concerned with AI alignment is not necessarily because of the scope of the issue, but also the urgency and tractability of the problem. The urgency of the problem comes from the idea that advanced AI will probably be developed this century. The tractability of the problem comes from the idea that there exists a set of goals that we could in theory put into an AI goals that are congruent with oursâyou might want to read up on the Orthogonality Thesis.
Furthermore, it is dangerous to assume that we should judge the effectiveness of certain activities merely based on prior evidence or results. There are some activities which are just infeasible to give post hoc judgements aboutâand this issue is one of them. The inherent nature of the problem is that we will probably only get about one chance to develop superintelligenceâbecause if we fail, then we will all probably die or otherwise be permanently unable to alter its goals.
To give you an analogy, few would agree that because climate change is an unprecedented threat, it therefore follows that we should wait until after the damage has been done to assess the best ways of mitigating it. Unfortunately for issues that have global scope, it doesnât look like we get a redo if things start going badly.
If you want to learn more about the research, I recommend reading Superintelligence by Nick Bostrom. The vast majority of AI alignment researchers are not worried about malevolent AI despite your statement. I mean this is in the kindest way possible, but if you really want to be sure that youâre on the right side of a debate, itâs worth understanding the best arguments against your position, not the worst.
Please, what AIA organizations? MIRI? And do not worry about offending me. I do not intend to offend. If I do/âdid though my tone or however, I am sorry.
That being said, I wish you wouldâve examined the actual claims I presented. I did not claim AI researchers are worried about a malevolent AI. I am not against researchers; research in robotics, industrial PLCs, nanotech, whateverâare fields in their own right. It is donating my income, as an individual that I take offense. People can fund whatever they want: A new planetary wing at a museum, research in robotics, research in CS, research in CS philosophy.
Although, Earning to Give does not follow. Thinking about and discussing the risks of strong AI does make sense, and we both seem to agree it is important. The CS grad students being supported, however, what makes them different from a random CS grad? Just because they claim to be researching AIA? Following the money, there is not a clear answer on which CS grad students are receiving it. Low or zero transparency. MIRI or no? Am I missing some public information?
Second, what do you define as advanced AI? Before, I said strong AI. Is that what you mean? Is there some sort of AI in between? Iâm not aware. This is crucially where I split with AI safety. The theory is an idea of a belief about the far future. To claim that weâre close to developing strong AI is unfounded to me. What in this century is so close to strong AI? Neural networks do not seem to be (from my light research).
I do not believe climate change is as simple to define a âbeforeâ and âafter.â Perhaps a large rogue solar flair or the Yellowstone supervolcano. Or perhaps even a time travel analogy would suffice ~ time travel safety research. There is no tractability/âsolvability. [Blank] cannot be defined because it doesnât exist; unfounded and unknown phenomena cannot be solved. Climate change exists. It is a very real reality. It has solvability. A belief in an idea about the future is a poor reason for claiming some sort of tractability for funding. Strong AI safety (singularity safety) has âsolvabilityâ for thinking about and discussingâbut, again, it does not follow that one should give monetarily. I feel like Iâm beating a dead horse with this point.
For the book recommendation, I looked into it. Iâd rather read about morality/âethics directly or further delve into better learning Java, Python, Logix5000, LabVIEW, etc.
SE for
SE against
Yes, MIRI is one. FHI is another.
You did, however, say âThe theoretical threat of a malevolent strong AI would be immense. But that does not mean one has cause or a valid reason to support CS grad students financially.â I assumed you meant that you believed someone was giving an argument along the lines of âsince malevolent AI is possible, then we should support CS grads.â If that is not what you meant, then I donât see the relevance of mentioning malevolent AI.
Since you also stated that you had an issue with me not being charitable, I would reciprocate likewise. I agree that we should be charitable to each otherâs opinions.
Having truthful views is not about winning debate. Itâs about making sure that you hold good beliefs for good reasons, end of story. I encourage you to imagine this conversation not as a way to convince me that Iâm wrongâbut more of a case study about what the current arguments are, and whether they are valid. In the end, you donât get points for winning an argument. You get points for actually holding correct views.
Therefore, itâs good to make sure that your beliefs actually hold weight under scrutiny. Not in a, âyou canât find the flaw after 10 minutes of self-sabotaged thinkingâ sort of way, but in a very deep understanding sort of way.
I agree people can fund whatever they want. Itâs important to make a distinction between normative questions and factual ones. Itâs true that people can fund whatever project they like; however, itâs also true that some projects have a high value from an impersonal utilitarian perspective. It is this latter category that I care about, which is why I want to find projects with particular high value. I believe that existential risk mitigation and AI alignment is among these projects, although I fully admit that I may be mistaken.
If you agree that thinking about something is valuable, why not also agree that funding that thing is valuable. It seems you think that the field should just get a certain threshold of funding that allows certain people to think about the problem just enoughâbut not too much. I donât a reason to believe that the field of AI alignment has reached that critical threshold. On the contrary, I believe the field is far from it at the moment.
I suppose when you make a donation to MIRI, itâs true that you canât be certain about how they spend that money (although I might be wrong about this, I havenât actually donated to MIRI). Generally though, funding an organization is about whether you think that their mission is neglected, and whether you think that further money would make a marginal impact in their cause area. This is no different than any other charity that EA aligned people endorse.
It might be confusing that there are all these terms for AI. To taboo the words âadvanced AIâ, âstrong AIâ, âAGIâ or othersâwhat I am worried about is an information processing system that can achieve broad success in cognitive tasks in a way that rivals or surpasses humans. I hope that makes it clear.
Iâm not quite clear what you mean here. If you mean we are worried about AI in the far future, fine. But then in the next sentence you say that weâre worried about being close to strong AI. How can we simultaneously believe both. If AI is near then I care about the near-term future. If AI is not near, then I care about the long-term future. I do not claim either, however. I think it is important consideration even if itâs a long way off.
This is what Iâm referring to when I talk about how important it is to really, truly understand something before developing an informed opinion about it. If you admit that you have only done light research, how can you be confident that you are right. Doing a bit of research might give you an edge for debate purposes, but we are talking about the future of life on Earth here. We really need to know the answers to these questions.
Lumping all existential risks in a single category and then asserting that thereâs no tractability is a simplified approach. First what we need is the probability of any given existential risk occurring. For instance, if scientists discovered that the Yellowstone supervolcano was probably about to erupt sometime in the next few centuries, Iâd definitely agree we should do research in that area, and we should fund that research as well. In fact, some research is being done in that area and Iâm happy that itâs being done.
Iâd agree with you if it was an idea asserted without evidence or reason. But thereâs a whole load of arguments about why it is a tractable field, and how we can do things nowâyes right nowâabout making the future safer. Ignorance of these arguments does not mean they do not exist.
Remember, ask yourself first what is true. Then form your opinion. Do not go the other way.
I am not trying to âwinâ anything. I am stating why MIRI is not transparent, and does not deal in scalable issues. As an individual, Earning to Give, it does not follow to fund such things under the guise of Effective Altruism. Existential risk is important to think about and discuss as individuals. However, funding CS grad students does not make sense in the light of Effective Altruism.
Funding does not increase âthinking.â The whole point of EA is to not give blindly. For example, giving food aid, although meaning well, can have a very negative effect (i.e., the crowding out effect on the local market). Nonmaleficence should be oneâs initial position in regards to funding.
Lastly, no I rarely accept something as true first. I do not first accept the null hypothesis. âBut thereâs a whole load of arguments about why it is a tractable fieldââWhat are they? Again, none of the actual arguments were examined: How is MIRI going about tractable/âsolvable issues? Who of MIRI is getting the funds? How is time travel safety not as relevant as AI safety?
Thanks for this discussion, which I find quite interesting. I think the effectiveness and efficiency of funding research projects concerning risks of AI is a largely neglected topic. Iâve posted some concerns on this below an older thread on MIRI: http://ââeffective-altruism.com/ââea/ââ14c/ââwhy_im_donating_to_miri_this_year/ââdce
the primary problem being the lack of transparency on the side of Open Phil. concerning the evaluative criteria used in their decision to award MIRI with an extremely huge grant.
The lists were interesting on how they allude to the different psychology and motivations each EA has between the two camps. I hope someday I can have a civil discussion with someone not directly benefiting from AIA (such as being involved in the research). Aside, I have a friend whoâs crazy about futurism, 2045 Initiative/âpropaganda, and in love with everything Musk says on Twitter.
âThe idea of âhelping the worst offâ is appealing.â Why wouldnât it be? Copenhagen Consensus.
âTheir reaction when they look about extinction risk or AI safety is nonsensicalâ, imaginary and completely unknownâzero tractability. No evidence to go off of since such technology does not exist. Why give to CS grad students? Itâs like trying to fund a mission to Mars, not priority. Itâs like funding time travel safety research, non sequitur.
âThey are generally an unhappy person.â I just had to laugh and compare how one interested in AI safety matched up. A neo-Freudian Jung/âMBTI type of deal. Almost like zodiac signs. Although, the Minnesota Multiphasic Personality Inventory (MMPI) is rigorousâso who am I to judge this informal inventory.
Anyway, I simply do not see that individual action or donation to AIA research has measurable outcomes. Weâre talking about Strong AI hereâit doesnât even exist! Not that it couldnât though. In the future, even the medium-term future, general standards of living could be significantly improved. Synthetic meat on a production scale is a much more realistic research area (or even anti-malaria mosquitoes) instead of making a fuss about imaginary-theoretical events. Weâre at a unique sliver in time where it is extremely practical to help lessen the suffering of humans and animals in the near and medium-term future. (I.e., we have rapid transportation and instant information transfer.)
Just because an event is theoretical doesnât mean that it wonât occur. An asteroid hitting the Earth is theoretical, but something I think you might realize is quite real when it impacts.
Some say that superintelligence doesnât have precedence, but I think thatâs overlooking a key fact. The rise of homo sapiens has radically altered the worldâand all signs point toward intelligence as the cause. We think at the moment that intelligence is just a matter of information processing, and therefore, there should be a way that it could be done by our own computers some day, if only we figured out the right algorithms to implement.
If we learn that superintelligence is impossible, that means our current most descriptive scientific theories are wrong, and we will have learned something new. Thatâs because that would indicate that humans are somehow cosmically special, or at least have hit the ceiling for general intelligence. On the flipside, if we create superintelligence, none of our current theories of how the world operates must be wrong.
Thatâs why itâs important to take seriously. Because the best evidence we have available tells us that itâs possible, not that itâs impossible.