How to improve your productivity: a systematic approach to sustainably increasing work output

Introduction

My estimated working life is long. I likely have around 40 productive years left. Given how valuable it is to achieve peak sustainable productivity, Joey and I have decided to implement an experiment. Theoretically, if you spent an entire year on productivity experiments and were able to increase output by 3% for the rest of your career, that would be worth doing, all things being equal. This is especially true because the experiment does not take away from high impact projects, but is run alongside them.

To do a strong trial you need three things 1) an idea worth testing, 2) a method for measuring productivity, and 3) a methodology.

An idea worth testing

There are millions of possible things that could affect your performance, and you cannot test them all at the same time. Sadly, there is little strong evidence on the specifics of how much a given technique increases output. Most of the research is very soft and focused on whether something helps at all without giving the size of the effect. Even if you had this data, due to huge interpersonal variability, it would be very hard to know how much things that work on others will work on you specifically.

Thankfully, this data is enough to generate a list of techniques that might help in significant ways. With a bit of research and a brainstorming session you should be able to list several dozen. Try to think divergently about your options. One angle might be to consider ways to save time, but you might also consider ways to improve focus or increase your energy levels.

A large list is great, but you will never have time to test everything that might work, so the next step is grouping and ranking your options.

Grouping and ranking

If you knew what techniques were best, you would not need to run the experiments in the first place, so really your rankings will be tentative guesses. Try not to get too attached to them.

You will be able to get a basic estimate of which ones to start with if you rank them by:

  1. How effective others seem to find them

  2. How effective you think it will be on you you relative to the normal population,

  3. How long/ā€‹hard it would be to test,

A good one to start with might be something particularly easy because this will help you get into the habit of testing and give you an easy win to motivate yourself.

A method for measuring productivity

Measuring productivity is a challenge, particularly for more diverse jobs. For jobs where there is a clear metric for success (e.g. a factory) some metrics might be both fairly clear, easy to measure, and correlated pretty strongly with your endline goal and thus ā€œtrue productivityā€. For jobs with a variety of tasks, or where the tasks are less measurable (e.g. being a good manager), such evaluations are harder to make. Unfortunately, our job and most EA jobs fall much more in the latter category. Thankfully, we do not need a perfect metric of productivity, just some measurable proxies we would expect to correlate with it. When these are combined into an index you can get a decent sense of whether an experiment is working and how significant its effect is on you relative to the other experiments you are running.

If the index is weighted before you start gathering data, itā€™s harder to game the results. However, you might find that as the process goes on you will have to change the weightings or the metrics themselves to better reflect your more intuitive sense. For example, one score might correlate very strongly with another and thus will not be worth tracking. I also recommend doing a sanity-test by putting an absurd experiment (e.g. working 1 hour a week) and see how it scores in your system. If it scores well, you need to make some adjustments.

The index of variables I used was a combination of:

  • My daily subjective productivity score

  • Number of hours worked (using time tracking software)

  • Daily independent assessment of my productivity by colleagues

  • The number of unintended breaks/ā€‹distractions

  • Number of creative insights generated

  • Number of missed deadlines

  • Attentiveness during management Skype calls

  • The number of changes suggested by the reviewers of my research reports

  • My personal happiness and subjective enthusiasm for work.

We also flag in the notes section confounding variables that might introduce noise in the data, such as sickness. To analyze the results, we take the values and the rating on each criteria and create a normalized distribution for it so that the average is zero and each value is converted to a number of standard deviations above or below the mean. I expect each of these numbers to show an incomplete picture, but together I would put pretty high confidence that a week where we score well on all these metrics relative to baseline would be a more productive week. The numbers, as well as a notes section of relevant events, are recorded daily.

A methodology for testing

The final thing you need is a consistent methodology for testing. This makes it clearer when something works and allows you to build in more robustness by reducing the chance of noise influencing your results.

We plan on testing each method for one week. We will then compare the results to the previous week and the average of the last three months. Doing one experiment each week over three months will result in around twelve experiments. However, one week is not long enough. Novelty could easily cause all or most of the effect, and many unrelated things in life can happen that could affect things. A week where your productivity increases relative to average is suggestive but not enough. A re-testing of the most promising three experiments will allow the effect to be more easily discerned from the noise. This second test could be carried out for a longer time, say one month, with fewer experiments. During this time minor variations could be done. For example, if working at a coworking space yielded big improvements in the first week, during the month-long test you might try out different desks within the space. After six months, three of the experiments will have been double-tested, and nine more will have been tested once.

Afterwards, we will insert the data into a spreadsheet along with other factors, like cost, difficulty of implementation, and the value of an extra productive hour. This way, one can calculate which of the productivity experiments are worth integrating in the long run. If three or more members of the Charity Entrepreneurship team conduct experiments, that could start to inform others on what would be best to test or try first for themselves personally or at an organizational level.

Conclusion

We expect that the results of systematic testing like this could lead to large gains in productivity. It will also create generalizable data for others to learn from. Others using the same methodology could also contribute and collectively build a knowledge base about how to best improve our output as a community.