I spent all day crying about this. An arms race is about the least safe way to approach. And we contributed to this. Many important people read Leopold’s report. He promoted it quite hard. But the background work predates Leopold’s involvement.
We were totally careless and self aggrandizing. I hope other people don’t pay for our sins.
What makes you say this? I agree that it is likely that Aschenbrenner’s report was influential here, but did we make Aschenbrenner write chapter IIId of Situational Awareness the way he did?
But the background work predates Leopold’s involvement.
Is there some background EA/aligned work that argues for an arms race? Because the consensus seems to be against starting a great power war.
I think a non-trivial fraction of Aschenbrenner’s influence as well as intellectual growth is due to us and the core EA/AI-Safety ideas, yeah. I doubt he would have written it if the extended community didn’t exist, and if he wasn’t mentored by Holden, etc.
I don’t disagree with this at all. But does this mean that blame can be attributed to the entire EA community? I think not.
Re mentorship/funding: I doubt that his mentors were hoping that he would accelerate the chances of an arms race conflict. As a corollary, I am sure nukes wouldn’t have been developed if the physics community in the 1930s didn’t exist or mentored different people or adopted better ethical norms. Even if they did the latter, it is unclear if that would have prevented the creation of the bomb.
(I found your comments under Ben West’s posts insightful; if true, it highlights a divergence between the beliefs of the broader EA community and certain influential EAs in DC and AI policy circles.)
Currently, it is just a report, and I hope it stays that way.
I spent all day crying about this. An arms race is about the least safe way to approach. And we contributed to this. Many important people read Leopold’s report. He promoted it quite hard. But the background work predates Leopold’s involvement.
We were totally careless and self aggrandizing. I hope other people don’t pay for our sins.
This sounds very much like the missile gap/bomber gap narrative, and yeah this is quite bad news if they actually adopt the commitments pushed here.
The evidence that China is racing to AGI is quite frankly very little, and I see a very dangerous arms race that could come:
https://forum.effectivealtruism.org/posts/cXBznkfoPJAjacFoT/are-you-really-in-a-race-the-cautionary-tales-of-szilard-and
I feel that. This report saddens me and I think its recommendations are very bad.
I don’t feel that I contributed to this. Perhaps I could have done more to prevent it, although it’s not obvious to me what I could have done.
What makes you say this? I agree that it is likely that Aschenbrenner’s report was influential here, but did we make Aschenbrenner write chapter IIId of Situational Awareness the way he did?
Is there some background EA/aligned work that argues for an arms race? Because the consensus seems to be against starting a great power war.
I think a non-trivial fraction of Aschenbrenner’s influence as well as intellectual growth is due to us and the core EA/AI-Safety ideas, yeah. I doubt he would have written it if the extended community didn’t exist, and if he wasn’t mentored by Holden, etc.
I don’t disagree with this at all. But does this mean that blame can be attributed to the entire EA community? I think not.
Re mentorship/funding: I doubt that his mentors were hoping that he would accelerate the chances of an arms race conflict. As a corollary, I am sure nukes wouldn’t have been developed if the physics community in the 1930s didn’t exist or mentored different people or adopted better ethical norms. Even if they did the latter, it is unclear if that would have prevented the creation of the bomb.
(I found your comments under Ben West’s posts insightful; if true, it highlights a divergence between the beliefs of the broader EA community and certain influential EAs in DC and AI policy circles.)
Currently, it is just a report, and I hope it stays that way.