I think the longer blog posts by Anthropic and OpenAI on their approaches to alignment are very important, under-appreciated and sometimes (I think falsely) dismissed as disingenuous.
Commentary from skeptical researchers about these plans could be interesting to include as well.
I think the longer blog posts by Anthropic and OpenAI on their approaches to alignment are very important, under-appreciated and sometimes (I think falsely) dismissed as disingenuous.
Commentary from skeptical researchers about these plans could be interesting to include as well.