I’m quite uncertain, but in general I don’t think it’s been the case that “if X technology goes well for humans, it’ll go well for animals”. I think in some key cases, it’s been the exact opposite, actually—e.g., industrialization leading to the rise of factory farming and killing/causing suffering to many more animals.
However, I also don’t think that AGI is going to be quite different from most technologies, at least in some ways (and definitely as it goes past AGI to ASI), and so I’m quite uncertain about how “going well for humans” might positively impact “going well for animals” in this specific case.
But I still see AGI as mostly being a technology developed by humans for human purposes, so it will be guided as such. And humans still predominantly use other animals as resources (for food, testing, raw materials, etc.). So, I think the default trajectory would probably be negative unless there is significant effort invested in helping AGI go well for nonhumans specifically.
I’m quite uncertain, but in general I don’t think it’s been the case that “if X technology goes well for humans, it’ll go well for animals”. I think in some key cases, it’s been the exact opposite, actually—e.g., industrialization leading to the rise of factory farming and killing/causing suffering to many more animals.
However, I also don’t think that AGI is going to be quite different from most technologies, at least in some ways (and definitely as it goes past AGI to ASI), and so I’m quite uncertain about how “going well for humans” might positively impact “going well for animals” in this specific case.
But I still see AGI as mostly being a technology developed by humans for human purposes, so it will be guided as such. And humans still predominantly use other animals as resources (for food, testing, raw materials, etc.). So, I think the default trajectory would probably be negative unless there is significant effort invested in helping AGI go well for nonhumans specifically.