Is there a concrete intervention that Credo AI might deploy in order to prevent existential risk from misaligned AI? If not, in which ways could RAI get us broadly closer to this goal?
See above for a more general response about existential risk.
To a “concrete intervention”—the current state of AI assessment is relatively poor. Many many models are deployed with the barest of adequacy assessment. Building a comprehensive assessment suite and making it easy to deploy on all productionizing ML systems is hugely important. Will it guard against issues related to existential risk? I don’t know honestly. But if someone comes up with good assessments that will probe such an ambiguous risk, we will incorporate it into the product!
Is there a concrete intervention that Credo AI might deploy in order to prevent existential risk from misaligned AI? If not, in which ways could RAI get us broadly closer to this goal?
See above for a more general response about existential risk.
To a “concrete intervention”—the current state of AI assessment is relatively poor. Many many models are deployed with the barest of adequacy assessment. Building a comprehensive assessment suite and making it easy to deploy on all productionizing ML systems is hugely important. Will it guard against issues related to existential risk? I don’t know honestly. But if someone comes up with good assessments that will probe such an ambiguous risk, we will incorporate it into the product!