I co-presented the above poster at the PPE Society Sixth Annual Meeting 2022 (henceforth ‘PPE Meeting’) in New Orleans on the 4th of November. Most of the 380 attendees were academics doing research on areas of philosophy that interact with politics or economics. The poster session, which was held at the end of a day of talks, lasted 1h30. There were around 6 posters being presented. In addition to providing attendees with a preview of the poster prior to the session, I gave them the following description:
In our poster session, we want to give an overview of the ‘alignment problem’ of artificial general intelligence (AGI). This is the problem of how to get AGI to do what we want. So far, it seems surprisingly and worryingly difficult. As things stand, AGI will likely be misaligned, resulting in catastrophic consequences. In addition to arguing why AGI is likely to be misaligned, we will also try to defend the assumption that AGI will be developed this century.
Goals
Practice talking with other academics about AI safety
Increase academics’ exposure to AI safety
Highlight AI safety as a special problem amongst the many problems within machine ethics and ethics of AI more generally
Increase my own understanding of AI safety and, in particular, get to grips with the core arguments for why we should be concerned
Gain a better understanding of how other academics conceive of AI safety, which arguments they perceive as unpersuasive and which they perceive as persuasive
Convince at least a couple of academics to take AI safety seriously and increase the likelihood that they connect their research to AI safety
Results
I believe I by and large achieved these goals. However, I was disappointed by the low turnout at the poster session. Out of (supposedly) 380 attendees, I estimate that around 60 were aware of my presentation topic, 25 were able to at least glance at the poster and 8 came to read the poster or interact with us.
Reactions
Which research question(s) am I working on?
Humans do all these things that I’ve claimed make AGI dangerous by default; are humans therefore just as dangerous?
Children do some of these things and they don’t immediately share our values but through praise and blame we mould them to alignment; can we not do something similar with AGI? In particular, could we train them to optimise for praise minus blame (or some similar function of the two) and then deploy them and incrementally increase their intelligence and capabilities, as is the case for human children?
Okay, AI compute will increase but why think that this will lead to human-like intelligence and the ability to self-improve?
Good calls
I divided the poster’s content into 3 sections, with two images in the centre. I think the images helped attract people’s attention and the sectioning made the structure of the arguments clearer and made it easy for people to stand on one side and read that side of the poster. With many people I began by talking about the third (right-hand-side) section, and only addressed the left-hand-side when they were sceptical of us ever developing AGI.
I used bullet points instead of full-sentence paragraphs. I think this made the poster more interesting (in any case, less boring) to read. It also meant that, as a result of less text, I could increase the text size.
Bad calls
I expected people who were confused by things like ‘AGI’ to come ask me out of curiosity but I think most preferred not to, perhaps due to unease in revealing their ignorance. For this reason, I should have made the poster more readable on its own.
I should have maxed out the size of the poster so that people could read it even from far away.
I should have displayed a list of promising research questions relevant to PPE academics. I had hoped to come up with a list of such questions but could only think of a couple and decided not to include any for the sake of space and because I’d have been unable to share many further insights in conversation.
Uncertainties
Should I have focused more on AI governance given the audience? I think that whilst AI governance might have been more relevant to the attendees, I would not have done a very good job of presenting on it.[1] Additionally, I think that AI governance is difficult to motivate strongly unless the audience is already aware of the alignment problem.
Should I have name-dropped more academics who acknowledge the risks of misaligned AGI? As a philosopher, I’ve been trained to become suspicious around appeals to authority. However, it seems that many philosophy academics are heavily reliant on what those they respect say.[2] I was under the impression that Nick Bostrom and Toby Ord are not so well respected amongst philosophy academics but Stuart Russell is. However, the one time I did name-drop Stuart Russell, my interlocutor hadn’t heard of him but was already in agreement with my conclusion due to Peter Railton supposedly also being in agreement. It seems that appeals to authority seem to work for some people. So, maybe I should have name-dropped more academics who are well-respected amongst philosophers. Relatedly, much of my content, including the two images, are taken from the Youtuber Robert Miles and the alarmist and unaccredited Eliezer Yudkowsky. I chose not to cite them in part because ad hominems are just as psychologically powerful as appeals to authority.
Credits
Thanks to Marius Hobbhahn for sharing his poster and for telling me about his experience presenting on AI safety. Thanks to Andrew Gewecke for co-presenting this poster with me. Thanks to Nick Cohen for answering some of my questions preceding and succeeding the presentation. Thanks also to Robert Miles and Eliezer Yudkowsky for the content and apologies for not citing either of you in the poster itself.
Contact
Feedback is most welcome. Either post in the comments section or reach out to me directly. My contact information is listed on my profile. If you find my poster useful as a template for your own presentation, feel free to steal it as I did from others. Just make sure you share your own writeup and include a link to mine.
I suppose this is normal given that many arguments are complex and we don’t have enough time to figure out for ourselves which are sound and which are not, and what a well-reasoned thinker says probably is a good enough guide to truth in many circumstances.
Poster Session on AI Safety
Context
I co-presented the above poster at the PPE Society Sixth Annual Meeting 2022 (henceforth ‘PPE Meeting’) in New Orleans on the 4th of November. Most of the 380 attendees were academics doing research on areas of philosophy that interact with politics or economics. The poster session, which was held at the end of a day of talks, lasted 1h30. There were around 6 posters being presented. In addition to providing attendees with a preview of the poster prior to the session, I gave them the following description:
In our poster session, we want to give an overview of the ‘alignment problem’ of artificial general intelligence (AGI). This is the problem of how to get AGI to do what we want. So far, it seems surprisingly and worryingly difficult. As things stand, AGI will likely be misaligned, resulting in catastrophic consequences. In addition to arguing why AGI is likely to be misaligned, we will also try to defend the assumption that AGI will be developed this century.
Goals
Practice talking with other academics about AI safety
Increase academics’ exposure to AI safety
Highlight AI safety as a special problem amongst the many problems within machine ethics and ethics of AI more generally
Increase my own understanding of AI safety and, in particular, get to grips with the core arguments for why we should be concerned
Gain a better understanding of how other academics conceive of AI safety, which arguments they perceive as unpersuasive and which they perceive as persuasive
Convince at least a couple of academics to take AI safety seriously and increase the likelihood that they connect their research to AI safety
Results
I believe I by and large achieved these goals. However, I was disappointed by the low turnout at the poster session. Out of (supposedly) 380 attendees, I estimate that around 60 were aware of my presentation topic, 25 were able to at least glance at the poster and 8 came to read the poster or interact with us.
Reactions
Which research question(s) am I working on?
Humans do all these things that I’ve claimed make AGI dangerous by default; are humans therefore just as dangerous?
Children do some of these things and they don’t immediately share our values but through praise and blame we mould them to alignment; can we not do something similar with AGI? In particular, could we train them to optimise for praise minus blame (or some similar function of the two) and then deploy them and incrementally increase their intelligence and capabilities, as is the case for human children?
Okay, AI compute will increase but why think that this will lead to human-like intelligence and the ability to self-improve?
Good calls
I divided the poster’s content into 3 sections, with two images in the centre. I think the images helped attract people’s attention and the sectioning made the structure of the arguments clearer and made it easy for people to stand on one side and read that side of the poster. With many people I began by talking about the third (right-hand-side) section, and only addressed the left-hand-side when they were sceptical of us ever developing AGI.
I used bullet points instead of full-sentence paragraphs. I think this made the poster more interesting (in any case, less boring) to read. It also meant that, as a result of less text, I could increase the text size.
Bad calls
I expected people who were confused by things like ‘AGI’ to come ask me out of curiosity but I think most preferred not to, perhaps due to unease in revealing their ignorance. For this reason, I should have made the poster more readable on its own.
I should have maxed out the size of the poster so that people could read it even from far away.
I should have displayed a list of promising research questions relevant to PPE academics. I had hoped to come up with a list of such questions but could only think of a couple and decided not to include any for the sake of space and because I’d have been unable to share many further insights in conversation.
Uncertainties
Should I have focused more on AI governance given the audience? I think that whilst AI governance might have been more relevant to the attendees, I would not have done a very good job of presenting on it.[1] Additionally, I think that AI governance is difficult to motivate strongly unless the audience is already aware of the alignment problem.
Should I have name-dropped more academics who acknowledge the risks of misaligned AGI? As a philosopher, I’ve been trained to become suspicious around appeals to authority. However, it seems that many philosophy academics are heavily reliant on what those they respect say.[2] I was under the impression that Nick Bostrom and Toby Ord are not so well respected amongst philosophy academics but Stuart Russell is. However, the one time I did name-drop Stuart Russell, my interlocutor hadn’t heard of him but was already in agreement with my conclusion due to Peter Railton supposedly also being in agreement. It seems that appeals to authority seem to work for some people. So, maybe I should have name-dropped more academics who are well-respected amongst philosophers. Relatedly, much of my content, including the two images, are taken from the Youtuber Robert Miles and the alarmist and unaccredited Eliezer Yudkowsky. I chose not to cite them in part because ad hominems are just as psychologically powerful as appeals to authority.
Credits
Thanks to Marius Hobbhahn for sharing his poster and for telling me about his experience presenting on AI safety. Thanks to Andrew Gewecke for co-presenting this poster with me. Thanks to Nick Cohen for answering some of my questions preceding and succeeding the presentation. Thanks also to Robert Miles and Eliezer Yudkowsky for the content and apologies for not citing either of you in the poster itself.
Contact
Feedback is most welcome. Either post in the comments section or reach out to me directly. My contact information is listed on my profile. If you find my poster useful as a template for your own presentation, feel free to steal it as I did from others. Just make sure you share your own writeup and include a link to mine.
Relatedly, I feel that we need better online resources concerning AI governance. The topic doesn’t even have it’s own Wikipedia page yet!
I suppose this is normal given that many arguments are complex and we don’t have enough time to figure out for ourselves which are sound and which are not, and what a well-reasoned thinker says probably is a good enough guide to truth in many circumstances.