This is really useful and makes sense—thanks for sharing your findings!
In my experience talking about an existing example of a problem like recommendation systems prioritizing unintended data points + example of meaningful AI capability usually gets people interested. Those two combined would probably be bad if we’re not careful. Jumping to the strongest/worst scenarios usually makes people recoil because it’s bad and unexpected and doesn’t make sense why you’re jumping to such an extreme outcome.
Do you have any examples of resources you were unaware of before? That could be useful to include as a section both for the actual resources and thinking about how to find such sources in the future.
The resources I was unaware of were usually highly specific technical papers (e.g. on some aspect of interpretability), so nothing helpful for a general audience.
This is really useful and makes sense—thanks for sharing your findings!
In my experience talking about an existing example of a problem like recommendation systems prioritizing unintended data points + example of meaningful AI capability usually gets people interested. Those two combined would probably be bad if we’re not careful. Jumping to the strongest/worst scenarios usually makes people recoil because it’s bad and unexpected and doesn’t make sense why you’re jumping to such an extreme outcome.
Do you have any examples of resources you were unaware of before? That could be useful to include as a section both for the actual resources and thinking about how to find such sources in the future.
Reflects my experience!
The resources I was unaware of were usually highly specific technical papers (e.g. on some aspect of interpretability), so nothing helpful for a general audience.