Autonomous Systems @ UK AI Safety Institute (AISI)
DPhil AI Safety @ Oxford (Hertford college, CS dept, AIMS CDT)
Former senior data scientist and software engineer + SERI MATS
I’m particularly interested in sustainable collaboration and the long-term future of value. I’d love to contribute to a safer and more prosperous future with AI! Always interested in discussions about axiology, x-risks, s-risks.
I enjoy meeting new perspectives and growing my understanding of the world and the people in it. I also love to read—let me know your suggestions! In no particular order, here are some I’ve enjoyed recently
Ord—The Precipice
Pearl—The Book of Why
Bostrom—Superintelligence
McCall Smith—The No. 1 Ladies’ Detective Agency (and series)
Melville—Moby-Dick
Abelson & Sussman—Structure and Interpretation of Computer Programs
Stross—Accelerando
Graeme—The Rosie Project (and trilogy)
Cooperative gaming is a relatively recent but fruitful interest for me. Here are some of my favourites
Hanabi (can’t recommend enough; try it out!)
Pandemic (ironic at time of writing...)
Dungeons and Dragons (I DM a bit and it keeps me on my creative toes)
Overcooked (my partner and I enjoy the foody themes and frantic realtime coordination playing this)
People who’ve got to know me only recently are sometimes surprised to learn that I’m a pretty handy trumpeter and hornist.
Some gestures which didn’t make the cut as they’re too woolly or not quite the right shape:
adversarial exponentials might force exponential expense per gain
e.g. combatting replicators
e.g. brute forcing passwords
many empirical ‘learning curve’ effects appear to consume exponential observations per increment
Wright’s Law (which is the more general cousin of Moore’s Law) requires exponentially many production iterations per incremental efficiency gain
Deep learning scaling laws appear to consume exponential inputs per incremental gain
AlphaCode and AlphaZero appear to make uniform gains per runtime compute doubling
OpenAI’s o-series ‘reasoning models’ appear to improve accuracy on many benchmarks with logarithmic returns to more ‘test time’ compute
(in all of these examples, there’s some choice of what scale to represent ‘output’ on, which affects whether the gains look uniform or not, so the thesis rests on whether the choices made are ‘natural’ in some way)