I think its incredibly hard to determine, but probably negative EV, as has been the history (IMO) of many (perhaps even most?) EA AI safety interventions. Well intended but for-the-worse-in-the-end. As a side note, I think the history of EA intervention in AI is a great example of just how hard it is to intentionally positively influence even the mid-term future.
It would only be positive if it actually contributed to some kind of slowdown or policy change, which seems pretty unlikely to me. I don’t think an eval reslut will ever sound an alarm bell, that would only come from some kind of obvious capability shift or warning shot that affected the general public.
More likely is that it slightly speeds up AI development basically like @MichaelDickens said.
1. Adding to the race dynamic with the clear targets and competition to be at the ’top′ 2. Giving AI companies better tools to measure progress 3. Incentivising investment
I think its incredibly hard to determine, but probably negative EV, as has been the history (IMO) of many (perhaps even most?) EA AI safety interventions. Well intended but for-the-worse-in-the-end. As a side note, I think the history of EA intervention in AI is a great example of just how hard it is to intentionally positively influence even the mid-term future.
It would only be positive if it actually contributed to some kind of slowdown or policy change, which seems pretty unlikely to me. I don’t think an eval reslut will ever sound an alarm bell, that would only come from some kind of obvious capability shift or warning shot that affected the general public.
More likely is that it slightly speeds up AI development basically like @MichaelDickens said.
1. Adding to the race dynamic with the clear targets and competition to be at the ’top′
2. Giving AI companies better tools to measure progress
3. Incentivising investment