Studying behaviour and interactions of boundedly rational agents, AI alignment and complex systems.
Research fellow at Future of Humanity Institute, Oxford. Other projects: European Summer Program on Rationality. Human-aligned AI Summer School. Epistea Lab.
Just wanted to flag the group is heavily selected for belief alignment with something like “EA/Constellation/Trajan House” views, and “AI enabled human takeovers” was promoted as agenda to prioritize in multiple widely read memos by high statues people in the community (which the organisers prioritized in the reading list).
I dislike the “echo chambre” effect where the steps are:
- invite people partially based on alignment with the idea cluster
- tell them to read memos advocating something written by some of the most central people in the cluster
- poll attendees
- results are framed as “leaders and key thinkers in the x-risk and AI safety communities agree”
It is in some sense useful, but in my view the cluster of people invited represents maybe ~30% of thinking about x-risk and AI safety, and its mostly an amplification of existing voices.
”The slight lean against misaligned AI takeover resources is perhaps the most surprising result for this audience, and merits closer examination.”
This is unsurprising given the marginal and somewhat confusing nature of the question. My wild guess is—
some attendees voted for everything; it is unclear what does it mean on the margin, probably to grow everything, and prioritize more neglected topics?
- some attendees understood the marginal question as “assuming fixed pie, how to change the allocation”—with this understanding you need to assign something negative weight for consistency