Can you describe exactly how much you think the average person, or average AI researcher, is willing to sacrifice on a personal level for a small chance at saving humanity? Are they willing to halve their income for the next ten years? Reduce by 90%?
I think in a world where there was a top down societal effort to try to reduce alignment risk, you might see different behavior. In the current world, I think the “personal choice” framework really is how it works because (for better or worse) there is not (yet) strong moral or social values attached to capability vs safety work.
Can you describe exactly how much you think the average person, or average AI researcher, is willing to sacrifice on a personal level for a small chance at saving humanity? Are they willing to halve their income for the next ten years? Reduce by 90%?
I think in a world where there was a top down societal effort to try to reduce alignment risk, you might see different behavior. In the current world, I think the “personal choice” framework really is how it works because (for better or worse) there is not (yet) strong moral or social values attached to capability vs safety work.