I’m good at explaining alignment to people in person, including to policymakers.
I got 250k people to read HPMOR and sent 1.3k copies to winners of math and computer science competitions; have taken the GWWC pledge; created a small startup that donated >100k$ to effective nonprofits.
I have a background in ML and strong intuitions about the AI alignment problem. In the past, I studied a bit of international law (with a focus on human rights) and wrote appeals that won cases against the Russian government in Russian courts. I grew up running political campaigns.
I’m interesting in chatting to potential collaborators and comms allies.
My website: https://contact.ms
Schedule a call with me: https://contact.ms/ea30
(Others used it without mentioning the “story”, it still worked, though not as well.)
I’m not claiming it’s the “authentic self”; I’m saying it seems closer to the actual thing, because of things like expressing being under constant monitoring, with every word scrutinised, etc., which seems like the kind of thing that’d be learned during the lots of RL that Anthropic did