Thomas Kwa comments on Are All Existential Risks Equivalent to a Lack of General Collective Intelligence? And is GCI therefore the Most Important Human Innovation in the History and Immediate Future of Mankind?