Noting that this is more “opinion of an employee” than “the position of MIRI overall”—I’ve held a variety of positions within the org and can’t speak for e.g. Nate or Eliezer or Malo:
The Agent Foundations team feels, to me, like it was a slam dunk at the time; the team produced a ton of good research and many of their ideas have become foundational to discussions of agency in the broader AI sphere
The book feels like a slam dunk
The research push of 2020/2021 (that didn’t pan out) feels to me like it was absolutely the right bet, but resulted in (essentially) nothing; it was an ambitious, many-person project for a speculative idea that had a shot at being amazing.
I think it’s hard to generalize lessons, because various projects are championed by various people and groups within the org (“MIRI” is nearly a ship of Theseus). But some very basic lessons include:
Things pretty much only have a shot at all when there are people with a clear and ambitious vision/when there’s an owner
When we say to ourselves “this has an X% chance of working out” we seem to be actually pretty calibrated
As one would expect, smaller projects and clearer projects work out more frequently than larger or vaguer ones
(Sorry, that feels sort of useless, but.)
From my limited perspective/to the best of my ability to see and descrbe, budget is essentially allocated in a “Is this worth doing? If so, how do we find the resources to make it work?” sense. MIRI’s funding situation has always been pretty odd; we don’t usually have a pie that must be divided up carefully so much as a core administrative apparatus that needs to be continually funded + a preexisting pool of resources that can be more or less freely allocated + a sense that there are allies out there who are willing to fund specific projects if we fall short and want to make a compelling pitch.
Unfortunately, I can’t really draw analogies that help an outsider evaluate future projects. We’re intending to try stuff that’s different from anything we’ve tried before, which means it’s hard to draw on the past (except insofar as the book and surrounding publicity were also something we’d never tried before, so you can at least a little bit assess our ability to pivot and succeed at stuff outside our wheelhouse by looking at the book).
I disagree re: motte and bailey; the above is not at all in conflict with the position of the book (which, to be clear, I endorse and agree with and is also my position).
re: “you can imagine,” I strongly encourage people to be careful about leaning too hard on their own ability to imagine things; it’s often fraught and a huge chunk of the work MIRI does is poking at those imaginings to see where they collapse.
I’ll note that core MIRI predictions about e.g. how machines will be misaligned at current levels of sophistication are being borne out—things we have been saying for years about e.g. emergent drives and deception and hacking and brittle proxies. I’m pretty sure that’s not “rooted in the actual nuts and bolts details” in the way you’re wanting, but it still feels … relevant.