‘Force multipliers’ for EA research

Introduction

TLDR: If solutions to the world’s most pressing problems require people with scarce, expensive skills then we should seek to discover and implement cost effective ‘force multipliers’ that maximise the productivity of these people.

With experience developing entrepreneurial technology products, I’m in a position to do just that. I describe some potential opportunities and ask for feedback and introductions.

Hello, I’m Craig from Melbourne, Australia. I’ve been leading the development of technology-based products for 14 years. I’ve co-created multiple products alongside academic staff and have spent several years in early startups, including as a solo founder for a data analytics company.

Most recently I led the development of a distributed data infrastructure and application development platform for climate science. The platform aims to make developing, operationalising and distributing climate solutions radically faster and easier.

I discovered EA back in 2010 and have quietly practised effective giving since. I’m moving on from my current work and want to discover a new opportunity to apply my skills to.

In this post I lay out an opportunity space with the aim of eliciting feedback and finding the right people to speak to next. I currently have only a nascent understanding of EA research areas[1], so please forgive (and gently correct) any naive errors.

The opportunity

Of the ten priority career reviews written by 80,000 hours, six emphasise gaining a technical/​quantitative PhD from a top university:

(To be clear, 80,000 hours does not argue that getting a quantitative PhD from an elite university is the only way to contribute to AI safety or other pressing problems.)

Developing important skills for addressing the world’s most pressing problems can be slow and difficult:

“Most of [our priority paths] are difficult to enter — you may need to start by investing in building skills for several years, and there may be relatively few positions available.”[2]

Other than encouraging early career EAs to pursue these paths, how else could we relieve specialist skills gaps and accelerate progress in these areas?

1. Increase the productivity of existing researchers

Increasing the productivity of longtermist researchers may be significantly easier, cheaper and faster than increasing the number of researchers, while providing equivalent value.

Collecting, cleaning and organising data can take up a significant proportion of data scientists’ time[3]. Project-based funding models, fragmented data licensing and contract negotiation inhibits collaboration and slows down progress. There’s often a significant gap between promising research and operational solutions.

Researchers have spent years developing a deep specialisation. But in my experience, they often approach their work as generalists; spending much of their time on tasks for which they have no comparative advantage.

Longtermist research is still relatively new and niche, so I’d expect the ecosystem of supporting data, tooling and complementary roles to be limited. Given the expense and difficulty of developing new researchers, investing in their supporting ecosystem may be a cost effective way to increase overall productivity—or at least not be subject to currently central constraints.

2. Increase supply of skilled work by reducing barriers to entry

People with doctoral degrees from elite universities make up just a fraction of a percent of the global population. Expanding the potential pool of top contributors beyond this demographic could help resolve critical skills gaps and increase the rate of progress. If it remains, say, 30 times easier (and much more lucrative) to become a commercial machine learning engineer than to work on AI safety, it is hard to see how safety can keep pace with capabilities.

Improving the ecosystem of infrastructure, tooling and data available to longtermist researchers and engineers could make working in these areas more approachable and desirable. This may increase the pool of motivated and talented people wanting to work on these issues.

Potential approaches

Benchmark datasets

Several benchmark datasets have been cited more than 100,000 times in machine learning research, and have also been used in operational solutions. Examples include datasets containing handwriting (e.g. MNIST), images of objects (e.g. ImageNet), audio clips (e.g. AudioSet) and product reviews (e.g. IMDB). Having pre-prepared datasets makes novel research significantly easier and faster. They also enable comparable results between different approaches and techniques.

Data collection efforts can be long running and extensive, worthy of an independent effort. The GroupLens project has collected 25 million movie ratings by operating an IMDB-like website since 1995. Common Crawl (with just two staff) creates and maintains petabytes of open web crawl data, which were used to train GPT-3[4].

Some examples relevant to longtermist research exist already (e.g. DeepMind’s AI Safety Gridworlds). There will be many more valuable opportunities for datasets and environments. What if we had quality, fine-grained data on millions of people’s espoused values, their perceptions of consequences, or retrospective evaluations of their past decisions?

Versions of this opportunity have been described by these EA Future Fund project ideas winners and by 80,000 hours here.

Libraries, Infrastructure & Platforms

Tooling (including libraries, infrastructure and platforms) has the potential to accelerate research by speeding up, eliminating, or improving the quality of the work undertaken.

Global cloud providers and research infrastructure provide cheap and scalable generic inputs such as compute and storage. Specialised platforms build on top of these to provide domain specific capabilities (see EcoCommons, an upcoming ecological research platform). Other opportunities could lie in coordination, crowdsourcing and resource allocation (e.g. “Kaggle for AI safety”).

Technical Standards

Since 1995, the Coupled Model Intercomparison Project (CMIP) has provided a framework and standards for running climate modelling experiments in a way that enables contributors to independently produce comparable results[5]. The latest iteration, CMIP6, contains contributions from 49 different modelling groups and more than 1000 individual scientists.

Standards (including practices, ontologies and definitions) provide an effective way for many parties to contribute to a greater effort without the overhead of close coordination.

Standards could potentially increase collective productivity of longtermist research, and also be a method for implementing interventions (e.g. AI safety and biosecurity standards).

The ask

Entrepreneurial product development seems to be a relatively neglected skillset in the EA community (outside earning to give). I believe promising high-impact opportunities exist that would be a good fit for my experience, and I’d like to work on them. I’m open to either joining an existing organisation or starting a new one.

Here’s how you can help.

I need to talk to people! Especially people who:

  • are doing high-leverage technical/​data heavy EA work, and would be happy to describe their work and its challenges

  • are interested in working on this topic—discovering, creating and operating an impactful technology ‘force multiplier’

  • would be interested in providing seed funding for an opportunity in this space

If you are one of these people, or know someone who is, please leave me a comment or direct message.

I am also seeking feedback. What already exists, what’s been tried before, which approaches seem more/​less promising to you and why?

  1. ^

    I have used both “EA research” and “longtermist research” in this post. The applicability to longtermist areas (particularly software and data heavy AI safety) is more obvious to me. If you can see an opportunity anywhere within or adjacent to EA I’m keen to hear about it.

  2. ^
  3. ^
  4. ^
  5. ^