I have never heard of the ideological Turing Tests that Claire referenced in their post. Those seem interesting. I have felt skeptical about the Turing Tests. That they tell us more about ourselves than they do about AI seems to reflect the nature of my skepticism.
I think that the question of/the definition of what intelligence is will be an important piece of AI. It seems that this question/definition is still vague and/or not agreed upon yet. Sometimes, I have thought that we probably haven’t delved enough into what our own intelligence is, what makes it tick, etc. to start conferring intelligence to other entities. So shifting the focus of Turing Tests from AIs to ourselves seems like a good idea to me. I can foresee ideological Turing Tests enhancing our empathy of others and revealing biases we had about others.
I have never heard of the ideological Turing Tests that Claire referenced in their post. Those seem interesting. I have felt skeptical about the Turing Tests. That they tell us more about ourselves than they do about AI seems to reflect the nature of my skepticism.
I think that the question of/the definition of what intelligence is will be an important piece of AI. It seems that this question/definition is still vague and/or not agreed upon yet. Sometimes, I have thought that we probably haven’t delved enough into what our own intelligence is, what makes it tick, etc. to start conferring intelligence to other entities. So shifting the focus of Turing Tests from AIs to ourselves seems like a good idea to me. I can foresee ideological Turing Tests enhancing our empathy of others and revealing biases we had about others.
I think the idea is from Bryan Caplan originally: https://www.econlib.org/archives/2011/06/the_ideological.html