Details on how an IAEA-style AI regulator would function?
Is anyone aware of work going into detail on how an international regulator for AI would function, how compliance might be monitored etc?
Is anyone aware of work going into detail on how an international regulator for AI would function, how compliance might be monitored etc?
Not really, or it depends on what kinds of rules the IAIA would set.
For monitoring large training runs and verifying compliance, see Verifying Rules on Large-Scale NN Training via Compute Monitoring (Shavit 2023).
Some more sketching of auditing with model evals is in Model evaluation for extreme risks (DeepMind 2023).
For completeness, here’s what OpenAI says in its “Governance of superintelligence” post:
It’s interesting how OpenAI basically concedes that it’s a fruitless effort further down in the very same post:
It’s not hard to imagine compute eventually becoming cheap and fast enough to train GPT4+ models on high-end consumer computers. How does one limit homebrewed training runs without limiting capabilities that are also used for non-training purposes?
This doesn’t point to detailed work in the space, but in “Nearcast-based ‘deployment problem’ analysis”, Karnofsky writes:
And here’s that footnote:
I don’t have a link to the report itself but Jason Hausenloy started some work on this a few months ago. https://youtu.be/1QY1L61TKx0