No, I’m talking somewhat narrowly about intent alignment, i.e. ensuring that our AI system is “trying” to do what we want. We are a relatively focused technical team, and a minority of the organization’s investment in safety and preparedness.
The policy team works on identifying misuses and developing countermeasures, and the applied team thinks about those issues as they arise today.
Hi, quick question, not sure this is the best place for it but curious:
Does work to “align GTP-3” include work to identify the most egregious uses for GTP-3 and develop countermeasures?
Cheers
No, I’m talking somewhat narrowly about intent alignment, i.e. ensuring that our AI system is “trying” to do what we want. We are a relatively focused technical team, and a minority of the organization’s investment in safety and preparedness.
The policy team works on identifying misuses and developing countermeasures, and the applied team thinks about those issues as they arise today.