EAGxNordics Unofficial Review Thread
I am usually really curious to get a taste of the overall atmosphere and insights gained from EAGs or EAGx events that I don’t attend. These gatherings, which host hundreds or even thousands of Effective Altruists, serve as valuable opportunities to exchange knowledge and potentially offer a snapshot of the most pressing EA themes and current projects. I attended EAGxNordics and, as a student, I will share my observations in a bullet-point format. Thank you Adash H-Moller for great comments and suggestions. Other attendees are very welcome to add their experiences / challenge these perspectives in the comments:
The majority of the participants seemed to have come from (obviously) Sweden, Norway, Finland, Estonia, Denmark and The Netherlands. Small countries with relatively tight EA communities.
I was particularly impressed with the line-up of speakers from non-EA-labeled think tanks and institutes—I think it provides a strong benefit, especially to those EAs who are quite familiar with EA but would not find out about these adjacent initiatives. It also serves to reduce the extent to which we’re in our own bubble.
I talked to numerous participants of the Future Academy—who all learned about EA through that program. They shared great experiences in policy, entrepreneurship and education (from before they knew about EA) and I think they are a great addition to the community.
Attendees can be more ambitious: Both in their conference experience and in their approach to EA. I spoke to too many students who had <5 1on1s planned, whereas these are regarded as one of the best ways to operate during a conference. Also, in terms of the career plans and EA projects I asked about, I would have loved to see bigger goals then the ones I heard.
I attended talks by employees of (the) GFI, Charity Entrepreneurship and The Simon Institute. The things they had in common:
They work on problems that are highly neglected (One speaker cited from a podcast: “No one is coming, it is up to us”)
They do their homework thoroughly
A key factor for their impact is their cooperation with local NGOs, governments and intergovernmental organizations.
(Suggested by Adash) The talk by an employee of Nähtamatud Loomad about ‘invisible animals’ was great and provided useful insight into what corporate lobbying actually looks like on the ground—I think specific, object-level content is great for keeping us grounded.
There could be more focus on analyzing EA as a community and considering what EA needs more of / needs to do differently, I asked a few people exactly those questions.
David Nash mentioned a range of critiques:
For the rest our fellow vikings and visitors tended to be less vocal in direct criticism
A lot of people talked about AI Safety
I felt there was a large group of students who were excited about contributing to this field
Participants with other backgrounds mentioned this as well and multiple participants voiced the preference for (a) more balanced content/narrative around topics like global development, animal welfare, etc.
(Suggested by Adash: N is small so take with a pinch of salt) I found the conversations I had with some early-career AI safety enthusiasts to show a lack of understanding of paths to x-risk and criticisms of key assumptions. I’m wondering if the early-stage AI field-building funnel might cause an echo chamber of unexamined AI panic that undermines general epistemics and cause-neutral principles.