Over all I think this is a good post. However this part surprised me.
However, I am personally worried about people skill-building for a couple of years and then not switching to doing the most valuable alignment work they can, because it can be easy to justify that your work is helping when it isn’t. This can happen even at labs that claim to have a safety focus! Working at any of Anthropic, DeepMind, Redwood Research, or OpenAI seems like a safe bet though.
I agree with the first bit. I’m also worried that people motivated to help with alignment end up contributing more to capability than alignment. But I’m very surprised by your list of what orgs to trust on this. Redwood Research seems fine, but the other ones?
Do you mean the specifically the safety teams of Anthropic, DeepMind and OpenAI or the full orgs? If you meant just the safety teams, I would appreciate if you made this explicit in the post, to prevent any misunderstanding. Especially since this seems to have become the go to post to give to people looking of AI Safety career guidance.
If you mean the full orgs, not just the safety team, I would like to know your reasoning.
Telling people to trust the three leading capabilities orgs seems very bad to me.
Over all I think this is a good post. However this part surprised me.
I agree with the first bit. I’m also worried that people motivated to help with alignment end up contributing more to capability than alignment. But I’m very surprised by your list of what orgs to trust on this. Redwood Research seems fine, but the other ones?
Do you mean the specifically the safety teams of Anthropic, DeepMind and OpenAI or the full orgs? If you meant just the safety teams, I would appreciate if you made this explicit in the post, to prevent any misunderstanding. Especially since this seems to have become the go to post to give to people looking of AI Safety career guidance.
If you mean the full orgs, not just the safety team, I would like to know your reasoning.
Telling people to trust the three leading capabilities orgs seems very bad to me.