The “Modeling Transformative AI Risk” project which I assisted with has the intent of explaining this, and we have a fairly extensive but not fully comprehensive report on the conceptual models that we think are critical, online here. (A less edited and polished version is on the alignment forum here.)
I think that you could probably read through the report itself in a week, going slowly and thinking through the issues—but doing so requires background in many of the questions discussed in Miles’ Youtube Channel, and the collection of AGI safety fundamentals resources which others recommended. Assuming a fairly basic understanding of machine learning and optimization, which probably requires the equivalent an undergraduate degree in a related field, the linked material on AI safety questions that you’d need to study to understand the issues, plus that report should get you to a fairly good gears-level understanding. I’d expect that 3 months of research and reading by someone with a fairly strong undergraduate background, or closer to a year for someone starting from scratch, would be sufficient to have an overall gears level model of the different aspects of the risk.
Given that, I will note that contributing to solving the problems requires quite a bit more investment in skill building—and depending on what you are planning on doing to address the risks, this could be equivalent to an advanced degree in mathematics, machine learning, policy, or international relations.
The “Modeling Transformative AI Risk” project which I assisted with has the intent of explaining this, and we have a fairly extensive but not fully comprehensive report on the conceptual models that we think are critical, online here. (A less edited and polished version is on the alignment forum here.)
I think that you could probably read through the report itself in a week, going slowly and thinking through the issues—but doing so requires background in many of the questions discussed in Miles’ Youtube Channel, and the collection of AGI safety fundamentals resources which others recommended. Assuming a fairly basic understanding of machine learning and optimization, which probably requires the equivalent an undergraduate degree in a related field, the linked material on AI safety questions that you’d need to study to understand the issues, plus that report should get you to a fairly good gears-level understanding. I’d expect that 3 months of research and reading by someone with a fairly strong undergraduate background, or closer to a year for someone starting from scratch, would be sufficient to have an overall gears level model of the different aspects of the risk.
Given that, I will note that contributing to solving the problems requires quite a bit more investment in skill building—and depending on what you are planning on doing to address the risks, this could be equivalent to an advanced degree in mathematics, machine learning, policy, or international relations.
Thank you, this is very much what I was looking for
Here’s the most up-to-date version of the AGI Safety Fundamentals curriculum. Be sure to check out Richard Ngo’s “AGI safety from first principles” report. There’s also a “Further resources” section at the bottom linking to pages like “Lots of links” from AI Safety Support.