I’ve been in the animal advocacy space since ~2012 and EA since ~2016.
I have a background in mathematics, philosophy, data, and programming. Thinking about how to do the most good with my life led me to animal advocacy, and more specifically working on ending animal farming.
I worked at Mercy For Animals for two years doing data analysis and research, and in 2022 I founded the organization Connect For Animals after conducting an analysis of the gaps in the animal advocacy space where I could have the most impact.
My favorite social media app is Goodreads.
This is very cool work and I would love to see it get more attention, and potentially get it or something like it incorporated into future AI model safety reports.
A couple things:
Does your approach roughly mirror the assessment approaches that companies currently take for assessing potential harm to humans? I imagine that it will be easier to make the case for a benchmark like this if it’s as similar as possible to existing best-in-class benchmarks. (You may have done this.)
I would recommend improving the graphs significantly. When I look at model performance, safety assessments, etc., 90% of the “rapid takeaway” value is in the graphs and having them be extremely easy to understand and very well designed. I cannot, at a glance determine what the graphs above are about other than “here are some models with some scores higher than others”. (Is score good? Bad?) I’d recommend that the graph should communicate 95% of what you want people to know so it can be rapidly shared.
Thanks for your work! Very important stuff.